text
stringlengths
10
951k
source
stringlengths
39
44
Polaris Sales Agreement The Polaris Sales Agreement was a treaty between the United States and the United Kingdom which began the UK Polaris programme. The agreement was signed on 6 April 1963. It formally arranged the terms and conditions under which the Polaris missile system was provided to the United Kingdom. The United Kingdom had been planning to buy the air-launched Skybolt missile to extend the operational life of the British V bombers, but the United States decided to cancel the Skybolt program in 1962 as it no longer needed the missile. The crisis created by the cancellation prompted an emergency meeting between the President of the United States, John F. Kennedy, and the Prime Minister of the United Kingdom, Harold Macmillan, which resulted in the Nassau Agreement, under which the United States agreed to provide Polaris missiles to the United Kingdom instead. The Polaris Sales Agreement provided for the implementation of the Nassau Agreement. The United States would supply the United Kingdom with Polaris missiles, launch tubes, and the fire control system. The United Kingdom would manufacture the warheads and submarines. In return, the US was given certain assurances by the United Kingdom regarding the use of the missile, but not a veto on the use of British nuclear weapons. The British Polaris ballistic missile submarines were built on time and under budget, and came to be seen as a credible deterrent that enhanced Britain's international status. Along with the 1958 US–UK Mutual Defence Agreement, the Polaris Sales Agreement became a pillar of the nuclear Special Relationship between Britain and the United States. The agreement was amended in 1982 to provide for the sale of the Trident missile system. During the early part of the Second World War, Britain had a nuclear weapons project, codenamed Tube Alloys. In August 1943, the Prime Minister of the United Kingdom, Winston Churchill and the President of the United States, Franklin Roosevelt, signed the Quebec Agreement, which merged Tube Alloys with the American Manhattan Project. The British government trusted that the United States would continue to share nuclear technology, which it regarded as a joint discovery, but the 1946 McMahon Act ended cooperation. Fearing a resurgence of United States isolationism, and Britain losing its great power status, the British government restarted its own development effort, now codenamed High Explosive Research. The first British atomic bomb was tested in Operation Hurricane on 3 October 1952. The subsequent British development of the hydrogen bomb, and a favourable international relations climate created by the Sputnik crisis, led to the McMahon Act being amended in 1958, and the restoration of the nuclear Special Relationship in the form of the 1958 US–UK Mutual Defence Agreement (MDA), which allowed Britain to acquire nuclear weapons systems from the United States. Britain's nuclear weapons armament was initially based on free-fall bombs delivered by the V bombers of the Royal Air Force (RAF), but the possibility of the manned bomber becoming obsolete by the late 1960s due to improvements in anti-aircraft defences was foreseen. In 1953, work began on a medium-range ballistic missile (MRBM) called Blue Streak, but by 1958, there were concerns about its vulnerability to a pre-emptive nuclear strike. To extend the effectiveness and operational life of the V bombers, an air-launched, rocket-propelled standoff missile called Blue Steel was developed, but it was anticipated that the air defences of the Soviet Union would improve to the extent that V bombers might still find it difficult to attack their targets. A solution appeared to be the American Skybolt missile, which combined the range of Blue Streak with the mobile basing of the Blue Steel, and was small enough that two could be carried on an Avro Vulcan bomber. An institutional challenge to Skybolt came from the United States Navy, which was developing a submarine-launched ballistic missile (SLBM), the UGM-27 Polaris. The US Chief of Naval Operations, Admiral Arleigh Burke, kept the First Sea Lord, Lord Mountbatten, apprised of its development. By moving the deterrent out to sea, Polaris offered the prospect of a deterrent that was invulnerable to a first strike, and reduced the risk of a nuclear strike on the British Isles. The British Nuclear Deterrent Study Group (BNDSG) produced a study that argued that SLBM technology was as yet unproven, that Polaris would be expensive, and that given the time it would take to build the boats, it could not be deployed before the early 1970s. The Cabinet Defence Committee therefore approved the acquisition of Skybolt in February 1960. The Prime Minister, Harold Macmillan, met with the President, Dwight D. Eisenhower, in March 1960, and secured permission to buy Skybolt. In return, the Americans could base the US Navy's Polaris ballistic missile submarines in the Holy Loch in Scotland. The financial arrangement was particularly favourable to Britain, as the US was charging only the unit cost of Skybolt, absorbing all the research and development costs. With this agreement in hand, the cancellation of Blue Streak was announced in the House of Commons on 13 April 1960. The subsequent American decision to cancel Skybolt created a political crisis in the UK, and an emergency meeting between Macmillan and President John F. Kennedy was called in Nassau, Bahamas. Macmillan rejected the US offers of paying half the cost of developing Skybolt, and of supplying the AGM-28 Hound Dog missile instead. This brought options down to Polaris, but the Americans would only supply it on condition that it be used as part of a proposed Multilateral Force (MLF). Kennedy ultimately relented, and agreed to supply Britain with Polaris missiles, while "the Prime Minister made it clear that except where Her Majesty's Government may decide that supreme national interests are at stake, these British forces will be used for the purposes of international defence of the Western Alliance in all circumstances." A joint statement to this effect, the Nassau Agreement, was issued on 21 December 1962. With the Nassau Agreement in hand, it remained to work out the details. Vice Admiral Michael Le Fanu had a meeting with the United States Secretary of Defense, Robert S. McNamara, on 21 December 1962, the final day of the Nassau conference. He found McNamara eager to help, and enthusiastic about the idea of Polaris costing as little as possible. The first issue identified was how many Polaris boats should be built. While the Vulcans to carry Skybolt were already in service, the submarines to carry Polaris were not, and there was no provision in the defence budget for them. Some naval officers feared that their construction would adversely impact the hunter-killer submarine programme. The First Sea Lord, Admiral of the Fleet Sir Caspar John, denounced the "millstone of Polaris hung around our necks" as "potential wreckers of the real navy". The number of missiles required was based on substituting for Skybolt. To achieve the same capability, the BNDSG calculated that this would require eight Polaris submarines, each of which would have 16 missiles, for a total of 128 missiles, with 128 one-megaton warheads. It was subsequently decided to halve this, based on the decision that the ability to destroy twenty Soviet cities would have nearly as great a deterrent effect as the ability to destroy forty. The Admiralty considered the possibility of hybrid submarines that could operate as hunter-killers while carrying eight Polaris missiles, but McNamara noted that this would be inefficient, as twice as many submarines would need to be on station to maintain the deterrent, and cautioned that the effect of tinkering with the US Navy's 16-missile layout was unpredictable. The Treasury costed a four-boat Polaris fleet at £314 million by 1972/73. A Cabinet Defence Committee meeting on 23 January 1963 approved the plan for four boats, with Thorneycroft noting that four boats would be cheaper and faster to build. A mission led by Sir Solly Zuckerman, the Chief Scientific Adviser to the Ministry of Defence, left for the United States to discuss Polaris on 8 January 1963. It included the Vice Chief of the Naval Staff, Vice Admiral Sir Varyl Begg; the Deputy Secretary of the Admiralty, James Mackay; Rear Admiral Hugh Mackenzie; and physicist Sir Robert Cockburn and F. J. Doggett from the Ministry of Aviation. That the involvement of the Ministry of Aviation might be a complicating factor was foreseen, but it had experience with nuclear weapons development. Mackenzie had been the Flag Officer Submarines until 31 December 1962, when Le Fanu had appointed him the Chief Polaris Executive (CPE). As such, he was directly answerable to Le Fanu as Controller of the Navy. His CPE staff was divided between London and Foxhill, near Bath, Somerset, where Royal Navy had its ship design, logistics and weapons groups. It was intended as a counterpart to the United States Navy Special Projects Office (SPO), with whom it would have to deal. The principal finding of the Zuckerman mission was that the Americans had developed a new version of the Polaris missile, the A3. With a range extended of , it had a new weapons bay housing three re-entry vehicles (REBs or Re-Entry Bodies in US Navy parlance) and a new W58 warhead to penetrate improved Soviet anti-missile defences expected to become available around 1970. A decision was therefore required on whether to purchase the old A2 missile or the new A3. The Zuckerman mission came out in favour of the new A3 missile, although it was still under development and not expected to enter service until August 1964, as the deterrent would remain credible for much longer. The decision was endorsed by the First Lord of the Admiralty, Lord Carrington, in May 1963, and was officially made by Thorneycroft on 10 June 1963. The choice of the A3 created a problem for the Atomic Weapons Research Establishment (AWRE) at Aldermaston, for the Skybolt warhead that had recently been tested in the Tendrac nuclear test at the Nevada Test Site in the United States would require a redesigned Re-Entry System (RES) in order to be fitted to a Polaris missile, at an estimated cost of between £30 million and £40 million. The alternative was to make a British copy of the W58. While the AWRE was familiar with the W47 warhead used in the A2, it knew nothing of the W58. A presidential determination was required to release information on the W58 under the MDA, but with this in hand, a mission led by John Challens, the Chief of Warhead Development at the AWRE, visited the Lawrence Livermore Laboratory from 22 to 24 January 1963, and was shown details of the W58. The Zuckerman mission found the SPO helpful and forthcoming, but there was one major shock. The British were expected to contribute to the research and development costs of the A3, backdated to 1 January 1963. These were expected to top $700 million by 1968. Skybolt had been offered to the UK at unit cost, with the US absorbing the research and development costs, but no such agreement had been reached at Nassau for Polaris. Thorneycroft baulked at the prospect of paying research and development costs, but McNamara pointed out that the United States Congress would not stand for an agreement that placed all the burden on the United States. Macmillan instructed the British Ambassador to the United States, Sir David Ormsby-Gore, to inform Kennedy that Britain was not willing to commit to an open-ended sharing of research and development costs, but, as a compromise, would pay an additional five per cent for each missile. He asked that Kennedy be informed that a breakdown of the Nassau Agreement would likely cause the fall of his government. Ormsby-Gore met with Kennedy that very day, and while Kennedy noted that the five per cent offer "was not the most generous offer he had ever heard of", he accepted it. McNamara, certain that the United States was being ripped off, calculated the five percent on top of not just the missiles, but their fire control and navigation systems as well, adding around £2 million to the bill. On Ormsby-Gore's advice, this formulation was accepted. An American mission now visited the United Kingdom. This was led by Paul H. Nitze, the Assistant Secretary of Defense for International Security Affairs, and included Walt W. Rostow, the Director of Policy Planning at the State Department, and Admiral Ignatius J. Galantin, the head of the SPO. The Americans had ideas about how the programme should be organised. They foresaw the UK Polaris programme having project officers from both countries, with a Joint Steering Task Group that met regularly to provide advice. This was accepted, and would become part of the final agreement. However, a follow-up British mission under Leslie Williams, the Director General Atomic Weapons at the Ministry of Aviation, whose members included Challens and Rear Admiral Frederick Dossor, was given a letter by the SPO with a list of subjects that were off limits. These included penetration aids, which were held to be outside the scope of the Nassau Agreement. One remaining obstacle in the path of the programme was how it would be integrated with the MLF. The British response to the MLF concept "ranged from unenthusiastic to hostile throughout the military establishment and in the two principal political parties". Apart from anything else, it was estimated to cost as much as £100 million over ten years. Nonetheless, the Foreign Office argued that Britain must support the MLF. The Nassau Agreement had invigorated the MLF effort in the United States. Kennedy appointed Livingston T. Merchant to negotiate the MLF with the European governments, which he did in February and March 1963. While reaffirming support for those parts of the Nassau Agreement concerning the MLF, the British were successful in getting them omitted from the Polaris Sales Agreement. The British team completed drafting the agreement in March 1963, and copies were circulated for discussion. The contracts for their construction were announced that month. The Polaris boats would be the largest submarines built in Britain up to that time, and would be built by Vickers Armstrong Shipbuilders in Barrow-in-Furness and Cammell Laird in Birkenhead. For similar reasons to the US Navy, the Royal Navy decided to base the boats at Faslane, on the Gareloch, not far from the US Navy's base on the Holy Loch. The drawback of the site was that it isolated the Polaris boats from the rest of the navy. The Polaris Sales Agreement was signed in Washington, DC, on 6 April 1963 by Ormsby-Gore and Dean Rusk, the United States Secretary of State. The two liaison officers were appointed in April; Captain Peter la Niece became the Royal Navy project officer in Washington, DC, while Captain Phil Rollings became the US Navy project officer in London. The Joint Steering Task Group held its first meeting in Washington on 26 June 1963. The shipbuilding programme would prove to be a remarkable achievement, with the four submarines built on time and within the budget. The first boat, was launched in September 1966, and commenced its first deterrent patrol in June 1968. The annual running costs of the Polaris boats came to around two per cent of the defence budget, and they came to be seen as a credible deterrent that enhanced Britain's international status. Along with the more celebrated 1958 US–UK Mutual Defence Agreement, the Polaris Sales Agreement became a pillar of the nuclear Special Relationship between Britain and the United States. The Polaris Sales Agreement provided an established framework for negotiations over missiles and re-entry systems. The legal agreement took the form of amending the Polaris Sales Agreement through an exchange of notes between the two governments so that "Polaris" in the original now also covered the purchase of Trident. There were also some amendments to the classified annexes of the Polaris Sales Agreement to delete the exclusion of penetrating aids. Under the Polaris Sales Agreement, the United Kingdom paid a five per cent levy on the cost of equipment supplied in recognition of US research and development costs already incurred. For Trident, a payment of $116 million was substituted. The United Kingdom procured the Trident system from America and fitted them to their own submarines, which had only 16 missile tubes like Polaris rather than the 24 in the American . The first , , entered operational service in December 1994, by which time the Cold War had ended.
https://en.wikipedia.org/wiki?curid=24815
Proto-Indo-Europeans The Proto-Indo-Europeans were a hypothetical prehistoric ethnolinguistic group of Eurasia who spoke Proto-Indo-European (PIE), the ancestor of the Indo-European languages according to linguistic reconstruction. Knowledge of them comes chiefly from that linguistic reconstruction, along with material evidence from archaeology and archaeogenetics. The Proto-Indo-Europeans likely lived during the late Neolithic, or roughly the 4th millennium BC. Mainstream scholarship places them in the Pontic–Caspian steppe zone in Eastern Europe (present day Ukraine and southern Russia). Some archaeologists would extend the time depth of PIE to the middle Neolithic (5500 to 4500 BC) or even the early Neolithic (7500 to 5500 BC), and suggest alternative location hypotheses. By the early second millennium BC, descendants of the Proto-Indo-Europeans had reached far and wide across Eurasia, including Anatolia (Hittites), the Aegean (the ancestors of Mycenaean Greece), the north of Europe (Corded Ware culture), the edges of Central Asia (Yamnaya culture), and southern Siberia (Afanasievo culture). Using linguistic reconstruction from old Indo-European languages such as Latin and Sanskrit, hypothetical features of the Proto-Indo-European language are deduced. Assuming that these linguistic features reflect culture and environment of the Proto-Indo-Europeans, the following cultural and environmental traits are widely proposed: Researchers have made many attempts to identify particular prehistoric cultures with the Proto-Indo-European-speaking peoples, but all such theories remain speculative. The scholars of the 19th century who first tackled the question of the Indo-Europeans' original homeland (also called "Urheimat", from German), had essentially only linguistic evidence. They attempted a rough localization by reconstructing the names of plants and animals (importantly the beech and the salmon) as well as the culture and technology (a Bronze Age culture centered on animal husbandry and having domesticated the horse). The scholarly opinions became basically divided between a European hypothesis, positing migration from Europe to Asia, and an Asian hypothesis, holding that the migration took place in the opposite direction. In the early 20th century, the question became associated with the expansion of a supposed "Aryan race", a fallacy promoted during the expansion of European empires and the rise of "scientific racism". The question remains contentious within some flavours of ethnic nationalism (see also Indigenous Aryans). A series of major advances occurred in the 1970s due to the convergence of several factors. First, the radiocarbon dating method (invented in 1949) had become sufficiently inexpensive to be applied on a mass scale. Through dendrochronology (tree-ring dating), pre-historians could calibrate radiocarbon dates to a much higher degree of accuracy. And finally, before the 1970s, parts of Eastern Europe and Central Asia had been off limits to Western scholars, while non-Western archaeologists did not have access to publication in Western peer-reviewed journals. The pioneering work of Marija Gimbutas, assisted by Colin Renfrew, at least partly addressed this problem by organizing expeditions and arranging for more academic collaboration between Western and non-Western scholars. The Kurgan hypothesis, the most widely held theory, depends on linguistic and archaeological evidence, but is not universally accepted. It suggests PIE origin in the Pontic-Caspian steppe during the Chalcolithic. A minority of scholars prefer the Anatolian hypothesis, suggesting an origin in Anatolia during the Neolithic. Other theories (Armenian hypothesis, Out of India theory, Paleolithic Continuity Theory, Balkan hypothesis) have only marginal scholarly support. In regard to terminology, in the 19th and early 20th centuries, the term "Aryan" was used to refer to the Proto-Indo-Europeans and their descendants. However, "Aryan" more properly applies to the Indo-Iranians, the Indo-European branch that settled parts of the Middle East and South Asia, as only Indic and Iranian languages explicitly affirm the term as a self-designation referring to the entirety of their people, whereas the same Proto-Indo-European root (*aryo-) is the basis for Greek and Germanic word forms which seem only to denote the ruling elite of Proto-Indo-European (PIE) society. In fact, the most accessible evidence available confirms only the existence of a common, but vague, socio-cultural designation of "nobility" associated with PIE society, such that Greek socio-cultural lexicon and Germanic proper names derived from this root remain insufficient to determine whether the concept was limited to the designation of an exclusive, socio-political elite, or whether it could possibly have been applied in the most inclusive sense to an inherent and ancestral "noble" quality which allegedly characterized all ethnic members of PIE society. Only the latter could have served as a true and universal self-designation for the Proto-Indo-European people. By the early twentieth century this term had come to be widely used in a racist context referring to a hypothesized white, blonde and blue eyed "master race", culminating with the pogroms of the Nazis in Europe. Subsequently, the term "Aryan" as a general term for Indo-Europeans has been largely abandoned by scholars (though the term "Indo-Aryan" is still used to refer to the branch that settled in Southern Asia). According to some archaeologists, PIE speakers cannot be assumed to have been a single, identifiable people or tribe, but were a group of loosely related populations ancestral to the later, still partially prehistoric, Bronze Age Indo-Europeans. This view is held especially by those archaeologists who posit an original homeland of vast extent and immense time depth. However, this view is not shared by linguists, as proto-languages, like all languages before modern transport and communication, occupied small geographical areas over a limited time span, and were spoken by a set of close-knit communities—a tribe in the broad sense. Researchers have put forward a great variety of proposed locations for the first speakers of Proto-Indo-European. Few of these hypotheses have survived scrutiny by academic specialists in Indo-European studies sufficiently well to be included in modern academic debate. In 1956 Marija Gimbutas (1921–1994) first proposed the Kurgan hypothesis. The name originates from the "kurgans" (burial mounds) of the Eurasian steppes. The hypothesis suggests that the Indo-Europeans, a nomadic culture of the Pontic-Caspian steppe (now part of Eastern Ukraine and Southern Russia), expanded in several waves during the 3rd millennium BC. Their expansion coincided with the taming of the horse. Leaving archaeological signs of their presence (see battle-axe people), they subjugated the peaceful European neolithic farmers of Gimbutas' Old Europe. As Gimbutas' beliefs evolved, she put increasing emphasis on the patriarchal, patrilinear nature of the invading culture, sharply contrasting it with the supposedly egalitarian, if not matrilinear culture of the invaded, to a point of formulating essentially feminist archaeology. A modified form of this theory by JP Mallory (1945- ), dating the migrations earlier (to around 3500 BC) and putting less insistence on their violent or quasi-military nature, remains the most widely accepted view of the Proto-Indo-European expansion. The Armenian hypothesis, based on the glottalic theory, suggests that the Proto-Indo-European language was spoken during the 4th millennium BC in the Armenian Highland. This Indo-Hittite model does not include the Anatolian languages in its scenario. The phonological peculiarities of PIE proposed in the glottalic theory would be best preserved in the Armenian language and the Germanic languages, the former assuming the role of the dialect which remained "in situ", implied to be particularly archaic in spite of its late attestation. Proto-Greek would be practically equivalent to Mycenean Greek and would date to the 17th century BC, closely associating Greek migration to Greece with the Indo-Aryan migration to India at about the same time (viz., Indo-European expansion at the transition to the Late Bronze Age, including the possibility of Indo-European Kassites). The Armenian hypothesis argues for the latest possible date of Proto-Indo-European ("sans" Anatolian), a full millennium later than the mainstream Kurgan hypothesis. In this, it figures as an opposite to the Anatolian hypothesis, in spite of the geographical proximity of the respective "Urheimaten" suggested, diverging from the time-frame suggested there by a full three millennia. The Anatolian hypothesis proposes that the Indo-European languages spread peacefully into Europe from Asia Minor from around 7000 BC with the advance of farming ("wave of advance"). The leading propagator of the theory is Colin Renfrew. The culture of the Indo-Europeans as inferred by linguistic reconstruction raises difficulties for this theory, since early neolithic cultures had neither the horse, nor the wheel, nor metal, terms for all of which are securely reconstructed for Proto-Indo-European. Renfrew dismisses this argument, comparing such reconstructions to a theory that the presence of the word "café" in all modern Romance languages implies that the ancient Romans had cafés too. The linguistic counter-argument to this might state that whereas there can be no clear Proto-Romance reconstruction of the word "café" according to historical linguistic methodology, words such as "wheel" in the Indo-European languages clearly point to an archaic form of the protolanguage. Another argument against Renfrew is the fact that ancient Anatolia is known to have been inhabited by non-Indo-European Caucasian-speaking peoples, namely the Hattians, the Chalybes, and the Hurrians. Following the publication of several studies on ancient DNA in 2015, Colin Renfrew has accepted the reality of migrations of populations speaking one or several Indo-European languages from the Pontic steppe towards Northwestern Europe. The rise of archaeogenetic evidence which uses genetic analysis to trace migration patterns also added new elements to the origins puzzle. The Kurgan hypothesis or steppe theory is the most widely accepted proposal to identify the Proto-Indo-European homeland from which the Indo-European languages spread out throughout Europe, Eurasia and parts of Asia. It postulates that the people of a Kurgan culture in the Pontic steppe north of the Black Sea were the most likely speakers of the Proto-Indo-European language (PIE). The term is derived from the Russian kurgan (курга́н), meaning tumulus or burial mound. According to three autosomal DNA studies, haplogroups R1b and R1a, now the most common in Europe (R1a is also very common in South Asia) would have expanded from the Russian steppes, along with the Indo European languages; they also detected an autosomal component present in modern Europeans which was not present in Neolithic Europeans, which would have been introduced with paternal lineages R1b and R1a, as well as Indo European Languages. Studies which analysed ancient human remains in Ireland and Portugal suggest that R1b was introduced in these places along with autosomal DNA from the Eastern European steppes. The subclade R1a1a (R-M17 or R-M198) is most commonly associated with Indo-European speakers, although the subclade R1b1a (P-297) has also been linked to the Centum branch of Indo-European. Data so far collected indicate that there are two widely separated areas of high frequency, one in Eastern Europe, around Poland and the Russian core, and the other in South Asia, around Indo-Gangetic Plain. The historical and prehistoric possible reasons for this are the subject of on-going discussion and attention amongst population geneticists and genetic genealogists, and are considered to be of potential interest to linguists and archaeologists also. A large, 2014 study by Underhill et al., using 16,244 individuals from over 126 populations from across Eurasia, concluded there was compelling evidence, that R1a-M420 originated in the vicinity of Iran. The mutations that characterize haplogroup R1a occurred ~10,000 years BP. Its defining mutation (M17) occurred about 10,000 to 14,000 years ago. Pamjav et al. (2012) believe that R1a originated and initially diversified either within the Eurasian Steppes or the Middle East and Caucasus region. Ornella Semino et al. propose a postglacial (Holocene) spread of the R1a1 haplogroup from north of the Black Sea during the time of the Late Glacial Maximum, which was subsequently magnified by the expansion of the Kurgan culture into Europe and eastward. According to Jones et al. (2015) and Haak et al. (2015), Yamnaya culture was exclusively R1b, autosomic tests indicate that the Yamnaya-people were the result of admixture between "Eastern Hunter-Gatherers" from eastern Europe (EHG) and "Caucasus hunter-gatherers" (CHG). Each of those two populations contributed about half the Yamnaya DNA. According to co-author Dr. Andrea Manica of the University of Cambridge: An analysis by David W. Anthony (2019) also suggests a genetic origin of proto-Indo-Europeans (the Yamnaya people) in the Eastern European steppe north of the Caucasus, derived from a mixture of Eastern European hunter-gatherers and hunter-gatherers from the Caucasus. Anthony also suggests that the proto-Indo-European language formed mainly from a base of languages spoken by Eastern European hunter-gathers with influences from languages of northern Caucasus hunter-gatherers, in addition to a possible later influence from the language of the Maikop culture to the south (which is hypothesized to have belonged to the North Caucasian family) in the later neolithic or bronze age involving little genetic impact. According to Haak et al. (2015), "Eastern European hunter-gatherers" who inhabited Russia were a distinctive population of hunter-gatherers with high affinity to a ~24,000-year-old Siberian from Mal'ta-Buret' culture, or other, closely related Ancient North Eurasian (ANE) people from Siberia and to the Western Hunter Gatherers (WHG). Remains of the "Eastern European hunter-gatherers" have been found in Mesolithic or early Neolithic sites in Karelia and Samara Oblast, Russia, and put under analysis. Three such hunter-gathering individuals of the male sex have had their DNA results published. Each was found to belong to a different Y-DNA haplogroup: R1a, R1b, and J. R1b is also the most common Y-DNA haplogroup found among both the Yamnaya and modern-day Western Europeans. The Near East population were most likely hunter-gatherers from the Caucasus (CHG) c.q. Iran Chalcolithic related people with a major CHG-component. Jones et al. (2015) analyzed genomes from males from western Georgia, in the Caucasus, from the Late Upper Palaeolithic (13,300 years old) and the Mesolithic (9,700 years old). These two males carried Y-DNA haplogroup: J* and J2a. The researchers found that these Caucasus hunters were probably the source of the farmer-like DNA in the Yamnaya, as the Caucasians were distantly related to the Middle Eastern people who introduced farming in Europe. Their genomes showed that a continued mixture of the Caucasians with Middle Eastern took place up to 25,000 years ago, when the coldest period in the last Ice Age started. According to Lazaridis et al. (2016), "a population related to the people of the Iran Chalcolithic contributed ~43% of the ancestry of early Bronze Age populations of the steppe." According to Lazaridis et al. (2016), these Iranian Chalcolithic people were a mixture of "the Neolithic people of western Iran, the Levant, and Caucasus Hunter Gatherers." Lazaridis et al. (2016) also note that farming spread at two places in the Near East, namely the Levant and Iran, from where it spread, Iranian people spreading to the steppe and south Asia. Haak et al. (2015) studied DNA from 94 skeletons from Europe and Russia aged between 3,000 and 8,000 years old. They concluded that about 4,500 years ago there was a major influx into Europe of Yamnaya culture people originating from the Pontic-Caspian steppe north of the Black Sea and that the DNA of copper-age Europeans matched that of the Yamnaya. A 2017 archaeogenetics study of Mycenaean and Minoan remains published in the journal "Nature" concluded that the Mycenaean Greeks were genetically closely related with the Minoans but unlike the Minoans also had a 13-18% genetic contribution from Bronze age steppe populations. Luigi Luca Cavalli-Sforza and Alberto Piazza argue that Renfrew and Gimbutas reinforce rather than contradict each other. states that "It is clear that, genetically speaking, peoples of the Kurgan steppe descended at least in part from people of the Middle Eastern Neolithic who immigrated there from Turkey." Piazza and Cavalli-Sforza (2006) state that: Spencer Wells suggests in a 2001 study that the origin, distribution and age of the R1a1 haplotype points to an ancient migration, possibly corresponding to the spread by the Kurgan people in their expansion across the Eurasian steppe around 3000 BC. About his old teacher Cavalli-Sforza's proposal, states that "there is nothing to contradict this model, although the genetic patterns do not provide clear support either", and instead argues that the evidence is much stronger for Gimbutas' model: David Reich (2018), noting the presence of some Indo-European languages (such as Hittite) in parts of ancient Anatolia, argues that "the most likely location of the population that first spoke an Indo-European language was south of the Caucasus Mountains, perhaps in present-day Iran or Armenia, because ancient DNA from people who lived there matches what we would expect for a source population both for the Yamnaya and for ancient Anatolians." Yet, Reich also notes that "...the evidence here is circumstantial as no ancient DNA from the Hittites themselves has yet been published." Kristian Kristiansen, in an interview with "Der Spiegel" in May 2018, stated that the Yamnaya culture may have had a predecessor at the Caucasus, where "proto-proto-Indo-European" was spoken. Recent DNA-research has led to renewed suggestions of a Caucasian homeland for the 'proto-Indo-Europeans'. According to Kroonen et al. (2018), Damgaard et al. (2018) ancient Anatolia "show no indication of a large-scale intrusion of a steppe population." They further note that this lends support to the Indo-Hittite hypothesis, according to which both proto-Anatolian and proto-Indo-European split-off from a common mother language "no later than the 4th millennium BCE." Haak et al. (2015) states that "the Armenian plateau hypothesis gains in plausibility" since the Yamnaya partly descended from a Near Eastern population, which resembles present-day Armenians." Wang et al. (2018) note that the Caucasus served as a corridor for gene flow between the steppe and cultures south of the Caucasus during the Eneolithic and the Bronze Age, stating that this "opens up the possibility of a homeland of PIE south of the Caucasus." However, Wang et al. also comment that the most recent genetic evidence supports an expansion of proto-Indo-Europeans through the steppe, noting: "but the latest ancient DNA results from South Asia also lend weight to a spread of Indo-European languages "via the steppe belt. The spread of some or all of the proto-Indo-European branches would have been possible via the North Caucasus and Pontic region and from there, along with pastoralist expansions, to the heart of Europe. This scenario finds support from the well attested and now widely documented 'steppe ancestry' in European populations, the postulate of increasingly patrilinear societies in the wake of these expansions (exemplified by R1a/R1b), as attested in the latest study on the Bell Beaker phenomenon." However, David W. Anthony in a 2019 analysis, criticizes the "southern" or "Armenian" hypothesis (addressing Reich, Kristiansen, and Wang). Among his reasons being: that the Yamnaya lack evidence of genetic influence from the bronze age or late neolithic Caucasus (deriving instead from an earlier mixture of Eastern European hunter-gatherers and Caucasus hunter-gatherers) and have paternal lineages that seem to derive from the hunter-gatherers of the Eastern European Steppe rather than the Caucasus, as well as a scarcity in the Yamnaya of the Anatolian Farmer admixture that had become common and substantial in the Caucasus around 5,000 BC. Anthony instead suggests a genetic and linguistic origin of proto-Indo-Europeans (the Yamnaya) in the Eastern European steppe north of the Caucasus, from a mixture of these two groups (EHG and CHG). He suggests that the roots of Proto-Indo-European ("archaic" or proto-proto-Indo-European) were in the steppe rather than the south and that PIE formed mainly from a base of languages spoken by Eastern European hunter-gathers with some influences from languages of Caucasus hunter-gatherers. The genetic basis of a number of physical features of the proto-Indo-European people were ascertained by the ancient DNA studies conducted by Haak "et al." (2015), Wilde "et al." (2014) and Mathieson "et al." (2015): they were genetically tall (phenotypic height is determined by both genetics and environmental factors), overwhelmingly dark-eyed (brown), dark-haired and had a skin colour that was moderately light, though somewhat darker than that of the average modern European.
https://en.wikipedia.org/wiki?curid=24818
Peter Mark Roget Peter Mark Roget ( ; 18 January 1779 – 12 September 1869) was a British physician, natural theologian, lexicographer and founding Secretary of The Portico Library. He is best known for publishing, in 1852, the "Thesaurus of English Words and Phrases", a classified collection of related words. He also submitted a paper to the Royal Society documenting the phi phenomenon in 1824. Peter Mark Roget was born in London, the son of Jean (John) Roget (1751–1783), a Genevan cleric, and his wife Catherine Romilly, sister of Samuel Romilly. After his father's death, the family moved to Edinburgh, in 1793, and he shortly began to study medicine at the University of Edinburgh, graduating in 1798. Samuel Romilly, who had supported his education, also introduced Roget into Whig social circles. Roget then attended lectures at London medical schools. Living in Clifton, Bristol from 1798 to 1799, he knew Thomas Beddoes and Humphry Davy, and frequented the Pneumatic Institute. Not making a quick start to a medical career, Roget in 1802 took a position as a tutor to the sons of John Leigh Philips, with whom he began a Grand Tour, during the Peace of Amiens, travelling with a friend Lovell Edgeworth, son of Richard Lovell Edgeworth. When the Peace abruptly ended, he was detained as a prisoner in Geneva. He was able to bring his pupils back to England, in late 1803, but Edgeworth was held in captivity until Napoleon's fall. Roget became, with help from Samuel Romilly, private physician to William Petty, 1st Marquess of Lansdowne, who died in 1805. He then succeeded Thomas Percival at Manchester Infirmary, and began to lecture on physiology. He moved to London in 1808, and in 1809 became a licentiate of the Royal College of Physicians. After an extended period of dispensary work and lecturing, in particular at the Russell Institution and Royal Institution, he was taken onto the staff of the Queen Charlotte Hospital in 1817. He also lectured at the London Institution and the Windmill Street School. Sir Samuel Romilly committed suicide, dying in Roget's presence, in 1818. Roget had been called in as adviser by the family, following the death of Lady Romilly. In 1823 Roget and Peter Mere Latham were brought in to investigate disease at Millbank Penitentiary. In 1828 Roget, with William Thomas Brande and Thomas Telford, submitted a report on London's water supply. In 1834 he became the first Fullerian Professor of Physiology at the Royal Institution. One of those who helped found the University of London in 1837, he was an examiner in physiology there. He gave up medical practice in 1840. Roget in later life became deaf, and was cared for by his daughter Kate. He died while on holiday in West Malvern, Worcestershire, aged 90, and is buried there in the cemetery of St James's Church. Roget retired from professional life in 1840, and by 1846 was working on the book that perpetuates his memory today. It has been claimed that Roget struggled with depression for most of his life, and that the thesaurus arose partly from an effort to battle it. A biographer stated that his obsession with list-making as a coping mechanism was well established by the time Roget was eight years old. In 1805, he began to maintain a notebook classification scheme for words, organized by meaning. During this period he also moved to Manchester, where he became the first secretary of the Portico Library. The catalogue of words was first printed in 1852, titled "Thesaurus of English Words and Phrases Classified and Arranged so as to Facilitate the Expression of Ideas and Assist in Literary Composition". During Roget's lifetime the work had twenty-eight printings. After his death it was revised and expanded by his son, John Lewis Roget (1828–1908), and later by John's son, the engineer Samuel Romilly Roget (1875–1953). Roget was elected a Fellow of the Royal Society in 1815, in recognition of a paper on a slide rule with a loglog scale. He was a secretary of the Society from 1827 to 1848. On 9 December 1824, Roget presented a paper on a peculiar optical illusion to the "Philosophical Transactions", which was published in 1825, as "Explanation of an optical deception in the appearance of the spokes of a wheel when seen through vertical apertures." The paper was noted by Michael Faraday and by Joseph Plateau, who both mentioned it in their articles that presented new illusions with apparent motion. It has often been heralded as the basis for the persistence of vision theory, which has for a long time been falsely regarded as the principle causing the perception of motion in animation and film. In 1834, Roget claimed to have invented "the Phantasmascope or Phenakisticope" in the spring 1831, a few years before Plateau introduced this first device that demonstrated stroboscopic animation. One of the promoters of the Medical and Chirurgical Society of London, which later became the Royal Society of Medicine, Roget was also a founder of the Society for the Diffusion of Useful Knowledge, writing a series of popular manuals for it. He wrote numerous papers on physiology and health, among them the fifth "Bridgewater Treatise", "Animal and Vegetable Physiology considered with reference to Natural Theology" (1834), and articles for the "Encyclopædia Britannica". He was hostile to phrenology, writing against it in a "Britannica" supplement in 1818, and devoting a two-volume work to it (1838). A chess player, in an article in the "London and Edinburgh Philosophical Magazine" Roget solved the general open knight's tour problem. He composed chess problems, and designed an inexpensive pocket chessboard. Canadian writer Keath Fraser published a story, "Roget's Thesaurus," in 1982, which is narrated in Roget's voice. He has Roget speak on his wife's death, from cancer. Roget also appears in Shelagh Stephenson's "An Experiment with an Air Pump", set in 1799, as the only historical character. The play takes place in the fictional household of Joseph Fenwick, and Roget is one of Fenwick's assistants. A picture book biography of Roget entitled "The Right Word: Roget and His Thesaurus" was published by Eerdmans Books in 2014. It was named a Caldecott Honor book for excellence in illustration and won the Sibert Medal for excellence in children's nonfiction. In 1824 Roget married Mary Taylor (1795–1833), daughter of Jonathan Hobson. They had a son John Lewis (1828–1908), and a daughter Kate.
https://en.wikipedia.org/wiki?curid=24820
Pterodactylus Pterodactylus (from Greek πτεροδάκτυλος, "pterodaktulos", meaning "winged finger") is an extinct genus of euctenochasmatian pterodactyloid pterosaur, whose members are commonly known as pterodactyls ( ). It is thought to contain only a single species, Pterodactylus antiquus, which is the first pterosaur species to be named and identified as a flying reptile. Fossil remains of "Pterodactylus" have primarily been found in the Solnhofen limestone of Bavaria, Germany, which dates back to the Late Jurassic period (early Tithonian stage), about 150.8 to 148.5 million years ago. More fragmentary remains of "Pterodactylus" have tentatively been identified from elsewhere in Europe, as well as in Africa. Many studies conclude that "Pterodactylus" was a carnivore that probably preyed upon fish as well as other small animals. Like all pterosaurs, "Pterodactylus" had wings formed by a skin and muscle membrane stretching from its elongated fourth finger to its hind limbs. It was supported internally by collagen fibres and externally by keratinous ridges. "Pterodactylus" was a small pterosaur compared to other famous genera such as "Pteranodon" and "Quetzalcoatlus", and it also lived earlier, during the Late Jurassic period, while both "Pteranodon" and "Quetzalcoatlus" lived during the Late Cretaceous. "Pterodactylus" lived alongside other small pterosaurs such as the well-known "Rhamphorhynchus", as well as other genera such as "Scaphognathus", "Anurognathus" and "Ctenochasma". The classification of "Pterodactylus" has since been confusing to paleontologists. Some classify "Pterodactylus" within the clade Euctenochasmatia, while others just consider it as a basal member of the suborder Pterodactyloidea. The type specimen of the animal now known as "Pterodactylus antiquus" was one of the first pterosaur fossils ever to be identified. The first "Pterodactylus" specimen was described by the Italian scientist Cosimo Alessandro Collini in 1784, based on a fossil skeleton that had been unearthed from the Solnhofen limestone of Bavaria. Collini was the curator of the "Naturalienkabinett", or nature cabinet of curiosities (a precursor to the modern concept of the natural history museum), in the palace of Charles Theodore, Elector of Bavaria at Mannheim. The specimen had been given to the collection by Count Friedrich Ferdinand zu Pappenheim around 1780, having been recovered from a lithographic limestone quarry in Eichstätt. The actual date of the specimen's discovery and entry into the collection is unknown however, and it was not mentioned in a catalogue of the collection taken in 1767, so it must have been acquired at some point between that date and its 1784 description by Collini. This makes it potentially the earliest documented pterosaur find; the "Pester Exemplar" of the species "Pterodactylus micronyx" was described in 1779 and possibly discovered earlier than the Mannheim specimen, but it was at first considered to be a fossilized crustacean, and it was not until 1856 that this species was properly described as a pterosaur by German paleontologist Hermann von Meyer. In his first description of the Mannheim specimen, Collini did not conclude that it was a flying animal. In fact, Collini could not fathom what kind of animal it might have been, rejecting affinities with the birds or the bats. He speculated that it may have been a sea creature, not for any anatomical reason, but because he thought the ocean depths were more likely to have housed unknown types of animals. The idea that pterosaurs were aquatic animals persisted among a minority of scientists as late as 1830, when the German zoologist Johann Georg Wagler published a text on "amphibians" which included an illustration of "Pterodactylus" using its wings as flippers. Wagler went so far as to classify "Pterodactylus", along with other aquatic vertebrates (namely plesiosaurs, ichthyosaurs, and monotremes), in the class Gryphi, between birds and mammals. The German/French scientist Johann Hermann was the one who first stated that "Pterodactylus" used its long fourth finger to support a wing membrane. Back in March 1800, Hermann alerted the French scientist Georges Cuvier to the existence of Collini's fossil, believing that it had been captured by the occupying armies of Napoleon and sent to the French collections in Paris (and perhaps to Cuvier himself) as war booty; at the time special French political commissars systematically seized art treasures and objects of scientific interest. Hermann sent Cuvier a letter containing his own interpretation of the specimen (though he had not examined it personally), which he believed to be a mammal, including the first known life restoration of a pterosaur. Hermann restored the animal with wing membranes extending from the long fourth finger to the ankle and a covering of fur (neither wing membranes nor fur had been preserved in the specimen). Hermann also added a membrane between the neck and wrist, as is the condition in bats. Cuvier agreed with this interpretation, and at Hermann's suggestion, Cuvier became the first to publish these ideas in December 1800 in a very short description. However, contrary to Hermann, Cuvier was convinced the animal was a reptile.The specimen had not in fact been seized by the French. Rather, in 1802, following the death of Charles Theodore, it was brought to Munich, where Baron Johann Paul Carl von Moll had obtained a general exemption of confiscation for the Bavarian collections. Cuvier asked von Moll to study the fossil but was informed it could not be found. In 1809 Cuvier published a somewhat longer description, in which he named the animal a "ptero-dactyle" and refuted a hypothesis by Johann Friedrich Blumenbach that it would have been a shore bird. Cuvier remarked: "It is not possible to doubt that the long finger served to support a membrane that, by lengthening the anterior extremity of this animal, formed a good wing." Contrary to von Moll's report, the fossil was not missing; it was being studied by Samuel Thomas von Sömmerring, who gave a public lecture about it on 27 December 1810. In January 1811, von Sömmerring wrote a letter to Cuvier deploring the fact that he had only recently been informed of Cuvier's request for information. His lecture was published in 1812, and in it von Sömmerring named the species "Ornithocephalus antiquus". The animal was described as being both a mammal, a bat, and a form in between mammals and birds, i.e. not intermediate in descent but in "affinity" or archetype. Cuvier disagreed, and the same year in his "Ossemens fossiles" provided a lengthy description in which he restated that the animal was a reptile. It was not until 1817 that a second specimen of "Pterodactylus" came to light, again from Solnhofen. This tiny specimen was that year described by von Sömmerring as "Ornithocephalus brevirostris", named for its short snout, now understood to be a juvenile character (this specimen is now thought to represent a juvenile specimen of a different genus, probably "Ctenochasma"). He provided a restoration of the skeleton, the first one published for any pterosaur. This restoration was very inaccurate, von Sömmerring mistaking the long metacarpals for the bones of the lower arm, the lower arm for the humerus, this upper arm for the breast bone and this sternum again for the shoulder blades. Sömmerring did not change his opinion that these forms were bats and this "bat model" for interpreting pterosaurs would remain influential long after a consensus had been reached around 1860 that they were reptiles. The standard assumptions were that pterosaurs were quadrupedal, clumsy on the ground, furred, warmblooded and had a wing membrane reaching the ankle. Some of these elements have been confirmed, some refuted by modern research, while others remain disputed. "Pterodactylus" is known from over 30 fossil specimens, and though most belong to juveniles, many preserve complete skeletons. "Pterodactylus antiquus" was a relatively small pterosaur, with an estimated adult wingspan of about , based on the only known adult specimen, which is represented by an isolated skull. Other "species" were once thought to have been smaller. However, these smaller specimens have been shown to represent juveniles of "Pterodactylus", as well as its contemporary relatives including "Ctenochasma", "Germanodactylus", "Aurorazhdarcho", "Gnathosaurus", and hypothetically "Aerodactylus" if this genus is truly valid. The skulls of adult "Pterodactylus" were long and thin, with about 90 narrow and conical teeth. The teeth extended back from the tips of both jaws, and became smaller farther away from the jaw tips, this was unlike the ones seen in most relatives, where teeth were absent in the upper jaw tip, and were relatively uniform in size. The teeth of "Pterodactylus" also extended farther back into the jaw compared to close relatives, and some were present below the front of the "nasoantorbital fenestra", which is the largest opening in the skull. Another autapomorphy that "Pterodactylus" has is that the skull and jaws were straight, which are unlike the upwardly curved jaws seen in the related ctenochasmatids. "Pterodactylus", like related pterosaurs, had a crest on its skull composed mainly of soft tissues. In adult "Pterodactylus", this crest extended between the back edge of the antorbital fenestra and the back of the skull. In at least one specimen, the crest had a short bony base, also seen in related pterosaurs like "Germanodactylus". Solid crests have only been found on large, fully adult specimens of "Pterodactylus", indicating that this was a display structure that became larger and more well developed as individuals reached maturity. In 2013, pterosaur researcher S. Christopher Bennett noted that other authors claimed that the soft tissue crest of "Pterodactylus" extended backward behind the skull; Bennett himself, however, didn't find any evidence for the crest extending past the back of the skull. Two specimens of "P. antiquus" (the holotype specimen BSP AS I 739 and the incomplete skull BMMS 7, the largest known skull of "P. antiquus") have a low bony crest on their skulls; in BMMS 7 it is 47.5 mm long (1.87 inches, more or less 24% of the estimated total length of its skull) and has a maximum height of 0.9 mm (0.035 inches) above the orbit. Several specimens previously referred to "P. antiquus" preserved evidence of the soft tissue extensions of these crests, including an "occipital lappet", a flexible, tab-like structure extending from the back of the skull. Most of these specimens have been reclassified in the related species "Aerodactylus scolopaciceps", which may however be nothing more than a junior synonym. Even if "Aerodactylus" were valid, at least one specimen with these features is still considered to belong to "Pterodactylus", BSP 1929 I 18, which has an occipital lappet similar to the proposed "Aerodactylus" definition, and also possesses a small triangular soft tissue crest with the peak of the crest positioned above the eyes. Like other pterosaurs (most notably "Rhamphorhynchus"), "Pterodactylus" specimens can vary considerably based on age or level of maturity. Both the proportions of the limb bones, size and shape of the skull, and size and number of teeth changed as the animals grew. Historically, this has led to various growth stages (including growth stages of related pterosaurs) being mistaken for new species of "Pterodactylus". Several detailed studies using various methods to measure growth curves among known specimens have suggested that there is actually only one valid species of "Pterodactylus", "P. antiquus". The youngest immature specimens of "Pterodactylus antiquus" (alternately interpreted as young specimens of the distinct species "P. kochi") have a small number of teeth, as few as 15 in some, and the teeth have a relatively broad base. The teeth of other "P. antiquus" specimens are both narrower and more numerous (up to 90 teeth are present in several specimens). "Pterodactylus" specimens can be divided into two distinct year classes. In the first year class, the skulls are only in length. The second year class is characterized by skulls of around long, but are still immature however. These first two size groups were once classified as juveniles and adults of the species "P. kochi", until further study showed that even the supposed "adults" were immature, and possibly belong to a distinct genus. A third year class is represented by specimens of the "traditional" "P. antiquus", as well as a few isolated, large specimens once assigned to "P. kochi" that overlap "P. antiquus" in size. However, all specimens in this third year class also show sign of immaturity. Fully mature "Pterodactylus" specimens remain unknown, or may have been mistakenly classified as a different genus. The distinct year classes of "Pterodactylus antiquus" specimens show that this species, like the contemporary "Rhamphorhynchus muensteri", likely bred seasonally and grew consistently during its lifetime. A new generation of 1st year class "P. antiquus" would have been produced seasonally, and reached 2nd-year size by the time the next generation hatched, creating distinct 'clumps' of similarly-sized and aged individuals in the fossil record. The smallest size class probably consisted of individuals that had just begun to fly and were less than one year old. The second year class represents individuals one to two years old, and the rare third year class is composed of specimens over two years old. This growth pattern is similar to modern crocodilians, rather than the rapid growth of modern birds. Comparisons between the scleral rings of "Pterodactylus antiquus" and modern birds and reptiles suggest that it may have been diurnal. This may also indicate niche partitioning with contemporary pterosaurs inferred to be nocturnal, such as "Ctenochasma" and "Rhamphorhynchus". Specimens of "Pterodactylus" have been found mainly in the Solnhofen limestone (geologically known as the Altmühltal Formation) of Bavaria, Germany. The main composition of this formation is fine-grained limestone that originated mainly from the nearby towns Solnhofen and Eichstätt, which is formed by mud silt deposits. The Solnhofen Limestone is a diverse Lagerstätte that contains a wide range of different creatures, including highly detailed fossilized imprints of soft bodied organisms such as jellyfishes. Abundant specimens of pterosaurs similar to "Pterodactylus" were also found within the formation, these include the rhamphorhynchids "Rhamphorhynchus" and "Scaphognathus", several gallodactylids such as "Aerodactylus", "Ardeadactylus", "Aurorazhdarcho" and "Cycnorhamphus", the ctenochasmatids "Ctenochasma" and "Gnathosaurus", the anurognathid "Anurognathus", the germanodactylid "Germanodactylus", as well as the basal euctenochasmatian "Diopecephalus". Fossil remains of the dinosaurs "Archaeopteryx" and "Compsognathus" were also found within the limestone, these specimens were related to early evolution of feathers, since they were some of the only ones that had them during the Jurassic period. Various lizard remains were also found alongside those of "Pterodactylus", with several specimens assigned to "Ardeosaurus", "Bavarisaurus" and "Eichstaettisaurus". Crocodylomorph specimens were widely distributed within the fossil site, most were assigned to the metriorhynchid genera "Cricosaurus", "Dakosaurus", "Geosaurus" and "Rhacheosaurus". These genera are colloquially called as marine or sea crocodiles due to their similar built. The rhynchocephalian genera "Homoeosaurus" and "Pleurosaurus" were also found in the formation alongside several turtles such as "Eurysternum" and "Paleomedusa". Fossils of the ichthyosaur "Aegirosaurus" were also present in the site, as well as fish remains, with many specimens assigned to ray-finned fishes such as the halecomorphs "Lepidotes", "Propterus", "Gyrodus", "Mesturus", "Proscinetes", "Caturus", "Ophiopsis" and "Ophiopsiella", the pachycormids "Asthenocormus", "Hypsocormus" and "Orthocormus", as well as the aspidorhynchid "Aspidorhynchus", and the ichthyodectid "Thrissops". The genus now known as "Pterodactylus" was originally named "Petro-Dactyle" by Cuvier in 1809, this was a typographical error however, and was later corrected by him to "Ptéro-Dactyle". In 1812, Samuel Thomas von Sömmerring named the same specimen "Ornithocephalus antiquus". The genus name was emended to the current "Pterodactylus" by Constantine Samuel Rafinesque in 1815. Unaware of Rafinesque's publication, Cuvier himself in 1819 again emended the genus name, but the specific name he then gave, "longirostris", has to give precedence to von Sömmerring's "antiquus". In 1888, English naturalist Richard Lydekker designated "Pterodactylus antiquus" as the type species, and specimen BSP No. AS.I.739 as the holotype of the genus. Paleontologist Hermann von Meyer in 1830, used the name Pterodactyli to contain "Pterodactylus" and other pterosaurs known at the time. This was emended to the family Pterodactylidae by Prince Charles Lucien Bonaparte in 1838. This group has more recently been given several competing definitions. Beginning in 2014, researchers Steven Vidovic and David Martill constructed an analysis in which several pterosaurs traditionally thought of as archaeopterodactyloids closely related to the ctenochasmatoids may have been more closely related to the more advanced ornithocheiroids, or in some cases, fall outside both groups. Their conclusion was published in 2017, in which they placed "Pterodactylus" as a basal member of the suborder Pterodactyloidea. As illustrated below, the results of a different topology are based on a phylogenetic analysis made by Longrich, Martill, and Andres in 2018. Unlike the previous results, they placed "Pterodactylus" within the clade Euctenochasmatia, resulting in a more derived position. Numerous species have been assigned to "Pterodactylus" in the years since its discovery. In the first half of the 19th century any new pterosaur species would be named "Pterodactylus", which thus became a typical "wastebasket taxon". Even after clearly different forms had later been given their own generic name, new species would be created from the very productive sites, throughout Europe and North America, often based on only slightly different material. The earliest reassignments of pterosaur species to "Pterodactylus" started in 1825, with the description of "Rhamphorhynchus"; fossil collector Georg Graf zu Münster alerted the German paleontologist Samuel Thomas von Sömmerring about several distinct fossil specimens, Sömmerring thought that they belonged to an ancient bird. Further fossil preparations had uncovered teeth, to which Graf zu Münster created a skull cast. He later sent the cast to Professor Georg August Goldfuss, who recognized it as a pterosaur, specifically a species of "Pterodactylus". At the time however, most paleontologists incorrectly consider the genus Ornithocephalus to be the valid name for "Pterodactylus", and therefore the specimen found was named as "Ornithocephalus münsteri", which was first mentioned by Graf zu Münster himself. Another specimen was found and described by Graf zu Münster in 1839, he assigned this specimen to a new separate species called "Ornithocephalus longicaudus"; the specific name means "long tail", in reference to the animal's tail size. German paleontologist Hermann von Meyer in 1845 officially emended that the genus "Pterodactylus" had priority over "Ornithocephalus", so he reassigned the species "O. münsteri" and "O. longicaudus" into "Pterodactylus münsteri" and "Pterodactylus longicaudus". In 1846, von Meyer created the new species "Pterodactylus gemmingi" based on long-tailed remains; the specific name honors the fossil collector Carl Eming von Gemming. Later, in 1847, von Meyer finally erected the generic name "Rhamphorhynchus" (meaning "beak snout") due to the distinctively long tails seen in the specimens found, which are much longer than those seen in "Pterodactylus". He assigned the species "P. longicaudus" as the type species of "Rhamphorhynchus", which resulted in a new combination called "Rhamphorhynchus longicaudus". The species "R. münsteri" was later changed to "Rhamphorhynchus muensteri" by Lydekker in 1888, due to the ICZN rule that prohibits non-standard Latin characters, such as "ü", in scientific names. Beginning in 1846, many pterosaur specimens were found near the village of Burham in Kent, England by British paleontologists James Scott Bowerbank and Sir Richard Owen. Bowerbank had assigned fossil remains to two new species; the first was named in 1846 as "Pterodactylus giganteus"; the specific name means "the gigantic one" in Latin, in reference to the large size of the remains, and the second species was named in 1851 as "Pterodactylus cuvieri", in honor of the French scientist Georges Cuvier. Later in 1851, Owen named and described new pterosaur specimens that have been found yet again in England. He assigned these specimens to a new species called "Pterodactylus compressirostris". In 1914 however, paleontologist Reginald Hooley redescribed "P. compressirostris", to which he erected the genus "Lonchodectes" (meaning "lance biter"), and therefore made "P. compressirostris" the type species, and created the new combination "L. compressirostris". In a 2013 review, "P. giganteus" and "P. cuvieri" were reassigned to new genera; "P. giganteus" was reassigned to a genus called "Lonchodraco" (meaning "lance dragon"), which resulted in a new combination called "L. giganteus", and "P. cuvieri" was reassigned to the new genus "Cimoliopterus" (meaning "chalk wing"), creating "C. cuvieri". Back in 1861, Owen had uncovered multiple distinctively looking fossil remains from the Cambridge Greensand, these were assigned to a new species named "Pterodactylus simus", though the British paleontologist Harry Govier Seeley had created a separate generic name called "Ornithocheirus", and reassigned "P. simus" as the type species, which created the combination "Ornithocheirus simus". In 1874, further specimens were found in England, again by Owen, these ones were assigned to a new species called "Pterodactylus sagittirostris", this species however, was reassigned to the genus "Lonchodectes" in 1914 by Hooley, which resulted in an "L. sagittirostris". This conclusion was revised by Rigal "et al." in 2017, who disagreed with Hooley's reassignment, and therefore created the genus "Serradraco", which afterwards resulted in a new combination called "S. sagittirostris". Assigning new pterosaur species to "Pterodactylus" was not only common in Europe, but also in North America; paleontologists such as Othniel Charles Marsh in 1871 for example, described several toothless pterosaur specimens, which were accompanied by teeth that belonged to the fish "Xiphactinus", which Marsh assumed that these teeth belonged to the pterosaur specimens he found, since all pterosaurs discovered at the time had teeth. He then assigned these specimens "Pterodactylus oweni", but this was changed to "Pterodactylus occidentalis" because "P. oweni" was found to have been preoccupied at the time. In 1872, American paleontologist Edward Drinker Cope also found various pterosaur specimens in North America, he assigned these to two new species known as "Ornithochirus umbrosus" and "Ornithochirus harpyia", Cope attempted to assign the specimens he found to the genus "Ornithocheirus", but misspelled forgetting the 'e'. In 1875 however, Cope reassigned the species "O. umbrosus" and "O. harpyia" into "Pterodactylus umbrosus" and "Pterodactylus harpyia", though these species had been considered "nomina dubia" ever since. Paleontologist Samuel Wendell Williston unearthed the first skull of the pterosaur, and found that the animal was toothless, this made Marsh create the genus "Pteranodon" (meaning "toothless wing"), and therefore reassigned all the American pterosaur species, including the ones that he named, from "Pterodactylus" to "Pteranodon". Later, in the 1980s, subsequent revisions by Peter Wellnhofer had reduced the number of recognized species to about half a dozen. Many species assigned to "Pterodactylus" had been based on juvenile specimens, and subsequently been recognized as immature individuals of other species or genera. By the 1990s it was understood that this was even true for part of the remaining species. "P. elegans", for example, was found by numerous studies to be an immature "Ctenochasma". Another species of "Pterodactylus" originally based on small, immature specimens was "P. micronyx". However, it has been difficult to determine exactly of what genus and species "P. micronyx" might be the juvenile form. Stéphane Jouve, Christopher Bennett and others had once suggested that it probably belonged either to "Gnathosaurus subulatus" or one of the species belonging to "Ctenochasma", though after additional research Bennett assigned it to the genus "Aurorazhdarcho". Another species with a complex history is "P. longicollum", named by von Meyer in 1854, based on a large specimen with a long neck and fewer teeth. Many researchers, including David Unwin, have found "P. longicollum" to be distinct from "P. kochi" and "P. antiquus". Unwin found "P. longicollum" to be closer to "Germanodactylus" and therefore requiring a new genus name. It has sometimes been placed in the genus "Diopecephalus" because Harry Govier Seeley based this genus partly on the "P. longicollum" material. However, it was shown by Bennett that the type specimen later designated for "Diopecephalus" was a fossil belonging to "P. kochi", and no longer thought to be separate from "Pterodactylus". "Diopecephalus" is therefore a synonym of "Pterodactylus", and as such is unavailable for use as a new genus for ""P." longicollum". ""P." longicollum" was eventually made the type species of a separate genus "Ardeadactylus". The only well-known and well-supported species left by the first decades of the 21st century were "P. antiquus" and "P. kochi". However, most studies between 1995 and 2010 found little reason to separate even these two species, and treated them as synonymous. More recent studies of pterosaur relationships have found anurognathids and pterodactyloids to be sister groups, which would limit the more inclusive group Caelidracones to just two clades. In 1996, Bennett suggested that the differences between specimens of "P. kochi" and "P. antiquus" could be explained by differences in age, with "P. kochi" (including specimens alternately classified in the species "P. scolopaciceps") representing an immature growth stage of "P. antiquus". In a 2004 paper, Jouve used a different method of analysis and recovered the same result, showing that the "distinctive" features of "P. kochi" were age-related, and using mathematical comparison to show that the two forms are different growth stages of the same species. An additional review of the specimens published in 2013 demonstrated that some of the supposed differences between "P. kochi" and "P. antiquus" were due to measurement errors, further supporting their synonymy. By the 2010s, a large body of research had been developed based on the idea that "P. kochi" and "P. scolopaciceps" were early growth stages of "P. antiquus". However, in 2014, two scientists began publishing research that challenged this paradigm. Steven Vidovic and David Martill concluded that differences between specimens of "P. kochi", "P. scolopaciceps", and "P. antiquus", such as different lengths of neck vertebrae, thinner or thicker teeth, more rounded skulls, and how far the teeth extended back in the jaws, were significant enough to separate them into three distinct species. Vidovic and Martill also performed a phylogenetic analysis which treated all relevant specimens as distinct units, and found that the "P. kochi" type specimen did not form a natural group with that of "P. antiquus". They concluded that the genus "Diopecephalus" could be returned to use to distinguish ""P". kochi" from "P. antiquus". They named the new genus "Aerodactylus" for "P. scolopaciceps" as well. So, what Bennett considered early growth stages of one species, Vidovic and Martill considered representatives of new species. In 2017, Bennett challenged this hypothesis. He claimed that while Vidovic and Martill had identified real differences between the these three groups of specimens, they had not provided any rationale that the differences were enough to distinguish them as species, rather than just individual variation, growth changes, or simply due to crushing and distortion during the fossilization process. Bennett pointed in particular to the data used to distinguish ""Aerodactylus"", which was so different from the data for related species, it might be due to an unnatural assemblage of specimens. As a result, Bennett continued to consider "Diopecephalus" and "Aerodactylus" simply as year-classes of immature "Pterodactylus antiquus". During its over-200-year history, the various species of "Pterodactylus" have gone through a number of changes in classification, and thus have acquired a large number of synonyms. Additionally, a number of species assigned to "Pterodactylus" are based on poor remains that have proven difficult to assign to one species or another, and are therefore considered "nomina dubia" (meaning "doubtful names"). The following list includes names that were used to identify new pterosaur species that now have been reclassified, or until recently, thought to be pertaining to "Pterodactylus" proper, and names based on other material that has as yet not been assigned to other genera. "Pterodactylus" is regarded as one of the most iconic prehistoric creatures, with multiple appearances in books, movies, as well as television series and several videogames. The informal name "pterodactyl" is sometimes used to refer to any kind of animal belonging to the order Pterosauria, though most of the times to "Pterodactylus", as its the most well-known member of the group. The popular aspect of "Pterodactylus" consists of an elongated head crest, and potentially large wings. Studies of "Pterodactylus" however, conclude that it may even lack a bony cranial crest, though several analysis have proven that "Pterodactylus" may in fact had a crest made up of soft tissue instead of bone. "Pterodactylus" is the star character of the 2005 horror film "Pterodactyl", where it is identified with the informal name "pterodactyl", hence the name of the film. In the film, the "pterodactyls" resemble the aspect of the distantly related genus "Pteranodon" due to the elongated bony cranial crest, and their enormous size. One peculiar feature that "Pterodactylus" had in the film is the possession of teeth, while this is generally accurate for "Pterodactylus", the overall appearance of the creatures in the film is similar to that of "Pteranodon", as well as the large size, this makes them resemble some kind of pterosaur identical to "Pteranodon", but with the possession of teeth. Another appearance of "Pterodactylus"-like creatures is in J. R. R. Tolkien's Middle-earth legendarium. In this novel, the Nazgûl, introduced as the Black Riders, are nine characters who rode flying monsters that looked similar in built to "Pterodactylus". Christopher Tolkien, the son of the author, described the flying monsters as "Nazgûl-birds"; his father described the appearance of the steeds as somewhat "pterodactylic", and acknowledged that these were obviously "new mythology".
https://en.wikipedia.org/wiki?curid=24823
Pterosaur Pterosaurs (; from Greek "pteron" and "sauros", meaning "wing lizard") were flying reptiles of the extinct clade or order Pterosauria. They existed during most of the Mesozoic: from the late Triassic to the end of the Cretaceous (228 to 66 million years ago). Pterosaurs are the earliest vertebrates known to have evolved powered flight. Their wings were formed by a membrane of skin, muscle, and other tissues stretching from the ankles to a dramatically lengthened fourth finger. There were two major types of pterosaurs. Basal pterosaurs (also called 'non-pterodactyloid pterosaurs' or ‘rhamphorhynchoids’) were smaller animals with fully toothed jaws and, typically, long tails. Their wide wing membranes probably included and connected the hind legs. On the ground, they would have had an awkward sprawling posture, but their joint anatomy and strong claws would have made them effective climbers, and they may have lived in trees. Basal pterosaurs were insectivores or predators on small vertebrates. Later pterosaurs (pterodactyloids) evolved many sizes, shapes, and lifestyles. Pterodactlyoids had narrower wings with free hind limbs, highly reduced tails, and long necks with large heads. On the ground, pterodactyloids walked well on all four limbs with an upright posture, standing plantigrade on the hind feet and folding the wing finger upward to walk on the three-fingered "hand." They could take off from the ground, and fossil trackways show at least some species were able to run and wade or swim. Their jaws had horny beaks, and some groups lacked teeth. Some groups developed elaborate head crests with sexual dimorphism. Pterosaurs sported coats of hair-like filaments known as pycnofibers, which covered their bodies and parts of their wings. Pycnofibers grew in several forms, from simple filaments to branching down feathers. These are homologous to the down feathers found on both avian and some non-avian dinosaurs, suggesting that early feathers evolved in the common ancestor of pterosaurs and dinosaurs, possibly as insulation. In life, pterosaurs would have had smooth or fluffy coats that did not resemble bird feathers. They were warm-blooded (endothermic) active animals. The respiratory system had efficient unidirectional "flow-through" breathing using air sacs, which hollowed out their bones to an extreme extent. Pterosaurs spanned a wide range of adult sizes, from the very small anurognathids to the largest known flying creatures of all time, including "Quetzalcoatlus" and "Hatzegopteryx", which reached wingspans of at least nine metres. The combination of endothermy, a good oxygen supply and strong muscles allowed pterosaurs to be powerful and capable flyers. Pterosaurs are often referred to by popular media or the general public as "flying dinosaurs", but dinosaurs are defined as the descendants of the last common ancestor of the Saurischia and Ornithischia, which excludes the pterosaurs. Pterosaurs are nonetheless more closely related to birds and other dinosaurs than to crocodiles or any other living reptile, though they are not bird ancestors. Pterosaurs are also colloquially referred to as "pterodactyls", particularly in fiction and by journalists. However, technically, "pterodactyl" only refers to members of the genus "Pterodactylus", and more broadly to members of the suborder Pterodactyloidea of the pterosaurs. Pterosaurs had a variety of lifestyles. Traditionally seen as fish-eaters, the group is now understood to have included hunters of land animals, insectivores, fruit eaters and even predators of other pterosaurs. They reproduced by means of eggs, some fossils of which have been discovered. The anatomy of pterosaurs was highly modified from their reptilian ancestors by the adaptation to flight. Pterosaur bones were hollow and air-filled, like those of birds. This provided a higher muscle attachment surface for a given skeletal weight. The bone walls were often paper-thin. They had a large and keeled breastbone for flight muscles and an enlarged brain able to coordinate complex flying behaviour. Pterosaur skeletons often show considerable fusion. In the skull, the sutures between elements disappeared. In some later pterosaurs, the backbone over the shoulders fused into a structure known as a notarium, which served to stiffen the torso during flight, and provide a stable support for the shoulder blade. Likewise, the sacral vertebrae could form a single synsacrum while the pelvic bones fused also. Basal pterosaurs include the clades Dimorphodontidae ("Dimorphodon"), Campylognathididae ("Eudimorphodon", "Campyognathoides"), and Rhamphorhynchidae ("Rhamphorhynchus", "Scaphognathus"). Pterodactyloids include the clades Ornithocheiroidea ("Istiodactylus", "Ornithocheirus", "Pteranodon"), Ctenochasmatoidea ("Ctenochasma", "Pterodactylus"), Dsungaripteroidea ("Germanodactylus", "Dsungaripterus"), and Azhdarchoidea ("Tapejara", "Tupuxuara", "Quetzalcoatlus"). The two groups overlapped in time, but the earliest pterosaurs in the fossil record are basal pterosaurs, and the latest pterosaurs are pterodactyloids. The position of the clade Anurognathidae ("Anurognathus, Jeholopterus, Vesperopterylus") is debated. Anurognathids (frog-headed pterosaurs) were highly specialized. Small flyers with shortened jaws and a wide gape, some had large eyes suggesting nocturnal or crepescular habits, mouth bristles, and feet adapted for clinging. Parallel adaptations are seen in birds and bats that prey on insects in flight. Pterosaurs had a wide range of sizes. Generally they were rather large. Even the smallest species had a wingspan no less than . The most sizeable forms represent the largest known animals ever to fly, with wingspans of up to . Standing, such giants could reach the height of a modern giraffe. Traditionally, it was assumed that pterosaurs were extremely light relative to their size. Later, it was understood that this would imply unrealistically low densities of their soft tissues. Some modern estimates therefore extrapolate a weight of up to for the largest species. Compared to the other vertebrate flying groups, the birds and bats, pterosaur skulls were typically quite large. Most pterosaur skulls had elongated jaws. Their skull bones tend to be fused in adult individuals. Early pterosaurs often had heterodont teeth, varying in build, and some still had teeth in the palate. In later groups the teeth mostly became conical. Front teeth were often longer, forming a "prey grab" in transversely expanded jaw tips, but size and position were very variable among species. With the derived Pterodactyloidea, the skulls became even more elongated, sometimes surpassing the combined neck and torso in length. This was caused by a stretching and fusion of the front snout bone, the premaxilla, with the upper jaw bone, the maxilla. Unlike most archosaurs, the nasal and antorbital openings of pterodactyloid pterosaurs merged into a single large opening, called the "nasoantorbital fenestra". This feature likely evolved to lighten the skull for flight. In contrast, the bones behind the eye socket contracted and rotated, strongly inclining the rear skull and bringing the jaw joint forward. The braincase was relatively large for reptiles. In some cases, fossilized keratinous beak tissue has been preserved, though in toothed forms, the beak is small and restricted to the jaw tips and does not involve the teeth. Some advanced beaked forms were toothless, such as the Pteranodontidae and Azhdarchidae, and had larger, more extensive, and more bird-like beaks. Some groups had specialised tooth forms. The Istiodactylidae had recurved teeth for eating meat. Ctenochasmatidae used combs of numerous needle-like teeth for filter feeding; "Pterodaustro" could have over a thousand bristle-like teeth. Dsungaripteridae covered their teeth with jawbone tissue for a crushing function. If teeth were present, they were placed in separate tooth sockets. Replacement teeth were generated behind, not below, the older teeth. The public image of pterosaurs is defined by their elaborate head crests. This was influenced by the distinctive backward-pointing crest of the well-known "Pteranodon". The main positions of such crests are the front of the snout, as an outgrowth of the premaxillae, or the rear of the skull as an extension of the parietal bones in which case it is called a "supraoccipital crest". Front and rear crests can be present simultaneously and might be fused into a single larger structure, the most expansive of which is shown by the Tapejaridae. "Nyctosaurus" sported a bizarre antler-like crest. The crests were only a few millimetres thin transversely. The bony crest base would typically be extended by keratinous or other soft tissue. Since the 1990s, new discoveries and a more thorough study of old specimens have shown that crests are far more widespread among pterosaurs than previously assumed. That they were extended by or composed completely of keratin, which does not fossilize easily, had misled earlier research. For "Pterorhynchus" and "Pterodactylus", the true extent of these crests has only been uncovered using ultraviolet photography. While fossil crests used to be restricted to the more advanced Pterodactyloidea, "Pterorhynchus" and "Austriadactylus" show that even some early pterosaurs possessed them. Like the upper jaws, the paired lower jaws of pterosaurs were very elongated. In advanced forms, they tended to be shorter than the upper cranium because the jaw joint was in a more forward position. The front lower jaw bones, the dentaries or "ossa dentalia", were at the tip tightly fused into a central symphysis. This made the lower jaws function as a single connected whole, the mandible. The symphysis was often very thin tranversely and long, accounting for a considerable part of the jaw length, up to 60%. If a crest was present on the snout, the symphysis could feature a matching mandible crest, jutting out to below. Toothed species also bore teeth in their dentaries. The mandible opened and closed in a simple vertical or "orthal" up-and-down movement. The vertebral column of pterosaurs numbered between thirty-four and seventy vertebrae. The vertebrae in front of the tail were "procoelous": the cotyle (front of the vertebral body) was concave and into it fitted a convex extension at the rear of the preceding vertebra, the condyle. Advanced pterosaurs are unique in possessing special processes projecting adjacent to their condyle and cotyle, the exapophyses, and the cotyle also may possess a small prong on its midline called a hypapophysis. The neck of pterosaurs was relatively long and straight. In pterodactyloids, the neck is typically longer than the torso. This length is not caused by an increase of the number of vertebrae, which is invariably seven. Some researchers include two transitional "cervicodorsals" which brings the number to nine. Instead, the vertebrae themselves became more elongated, up to eight times longer than wide. Nevertheless, the cervicals were wider than high, implying a better vertical than horizontal neck mobility. Pterodactyloids have lost all neck ribs. Pterosaur necks were probably rather thick and well-muscled, especially vertically. The torso was relatively short and egg-shaped. The vertebrae in the back of pterosaurs originally might have numbered eighteen. With advanced species a growing number of these tended to be incorporated into the sacrum. Such species also often show a fusion of the front dorsal vertebrae into a rigid whole which is called the notarium after a comparable structure in birds. This was an adaptation to withstand the forces caused by flapping the wings. The notarium included three to seven vertebrae, depending on the species involved but also on individual age. These vertebrae could be connected by tendons or a fusion of their neural spines into a "supraneural plate". Their ribs also would be tightly fused into the notarium. In general, the ribs are double-headed. The sacrum consisted of three to ten sacral vertebrae. They too, could be connected via a supraneural plate that, however, would not contact the notarium. The tails of pterosaurs were always rather slender. This means that the caudofemoralis retractor muscle which in most basal Archosauria provides the main propulsive force for the hindlimb, was relatively unimportant. The tail vertebrae were amphicoelous, the vertebral bodies on both ends being concave. Early species had long tails, containing up to fifty caudal vertebrae, the middle ones stiffened by elongated articulation processes, the zygapophyses, and chevrons. Such tails acted as rudders, sometimes ending at the rear in a vertical diamond-shaped or oval vane. In pterodactyloids, the tails were much reduced and never stiffened, with some species counting as few as ten vertebrae. The shoulder girdle was a strong structure that transferred the forces of flapping flight to the thorax. It was probably covered by thick muscle layers. The upper bone, the shoulder blade, was a straight bar. It was connected to a lower bone, the coracoid that is relatively long in pterosaurs. In advanced species, their combined whole, the scapulocoracoid, was almost vertically oriented. The shoulder blade in that case fitted into a recess in the side of the notarium, while the coracoid likewise connected to the breastbone. This way, both sides together made for a rigid closed loop, able to withstand considerable forces. A peculiarity was that the breastbone connections of the coracoids often were asymmetrical, with one coracoid attached in front of the other. In advanced species the shoulder joint had moved from the shoulder blade to the coracoid. The joint was saddle-shaped and allowed considerable movement to the wing. It faced sideways and somewhat upwards. The breastbone, formed by fused paired "sterna", was wide. It had only a shallow keel. Via sternal ribs, it was at its sides attached to the dorsal ribs. At its rear, a row of belly ribs or gastralia was present, covering the entire belly. To the front, a long point, the "cristospina", jutted obliquely upwards. The rear edge of the breastbone was the deepest point of the thorax. Clavicles or interclavicles were completely absent. Pterosaur wings were formed by bones and membranes of skin and other tissues. The primary membranes attached to the extremely long fourth finger of each arm and extended along the sides of the body. Where they ended has been very controversial but since the 1990s a dozen specimens with preserved soft tissue have been found that seem to show they attached to the ankles. The exact curvature of the trailing edge, however, is still equivocal. While historically thought of as simple leathery structures composed of skin, research has since shown that the wing membranes of pterosaurs were highly complex dynamic structures suited to an active style of flight. The outer wings (from the tip to the elbow) were strengthened by closely spaced fibers called "actinofibrils". The actinofibrils themselves consisted of three distinct layers in the wing, forming a crisscross pattern when superimposed on one another. The function of the actinofibrils is unknown, as is the exact material from which they were made. Depending on their exact composition (keratin, muscle, elastic structures, etc.), they may have been stiffening or strengthening agents in the outer part of the wing. The wing membranes also contained a thin layer of muscle, fibrous tissue, and a unique, complex circulatory system of looping blood vessels. The combination of actinofibrils and muscle layers may have allowed the animal to adjust the wing slackness and camber. As shown by cavities in the wing bones of larger species and soft tissue preserved in at least one specimen, some pterosaurs extended their system of respiratory air sacs into the wing membrane. The pterosaur wing membrane is divided into three basic units. The first, called the "propatagium" ("fore membrane"), was the forward-most part of the wing and attached between the wrist and shoulder, creating the "leading edge" during flight. The "brachiopatagium" ("arm membrane") was the primary component of the wing, stretching from the highly elongated fourth finger of the hand to the hindlimbs. Finally, at least some pterosaur groups had a membrane that stretched between the legs, possibly connecting to or incorporating the tail, called the uropatagium; the extent of this membrane is not certain, as studies on "Sordes" seem to suggest that it simply connected the legs but did not involve the tail (rendering it a cruropatagium). A common interpretation is that non-pterodactyloid pterosaurs had a broader uro/cruropatagium stretched between their long fifth toes, with pterodactyloids, lacking such toes, only having membranes running along the legs. There has been considerable argument among paleontologists about whether the main wing membranes (brachiopatagia) attached to the hindlimbs, and if so, where. Fossils of the rhamphorhynchoid "Sordes", the anurognathid "Jeholopterus", and a pterodactyloid from the Santana Formation seem to demonstrate that the wing membrane did attach to the hindlimbs, at least in some species. However, modern bats and flying squirrels show considerable variation in the extent of their wing membranes and it is possible that, like these groups, different species of pterosaur had different wing designs. Indeed, analysis of pterosaur limb proportions shows that there was considerable variation, possibly reflecting a variety of wing-plans. The bony elements of the arm formed a mechanism to support and extend the wing. Near the body, the humerus or upper arm bone is short but powerfully built. It sports a large deltopectoral crest, to which the major flight muscles are attached. Despite the considerable forces exerted on it, the humerus is hollow or pneumatised inside, reinforced by bone struts. The long bones of the lower arm, the ulna and radius, are much longer than the humerus. They were probably incapable of pronation. A bone unique to pterosaurs, known as the pteroid, connected to the wrist and helped to support the forward membrane (the propatagium) between the wrist and shoulder. Evidence of webbing between the three free fingers of the pterosaur forelimb suggests that this forward membrane may have been more extensive than the simple pteroid-to-shoulder connection traditionally depicted in life restorations. The position of the pteroid bone itself has been controversial. Some scientists, notably Matthew Wilkinson, have argued that the pteroid pointed forward, extending the forward membrane and allowing it to function as an adjustable flap. This view was contradicted in a 2007 paper by Chris Bennett, who showed that the pteroid did not articulate as previously thought and could not have pointed forward, but rather was directed inward toward the body as traditionally interpreted. David Peters in 2009 proposed that the pteroid articulated with the 'saddle' of the radiale (proximal syncarpal) and both the pteroid and preaxial carpal were migrated centralia. This seems to be confirmed by specimens of "Changchengopterus pani" and "Darwinopterus linglongtaensis" showing the pteroid in articulation with the proximal syncarpal. The pterosaur wrist consists of two inner (proximal, at the side of the long bones of the arm) and four outer (distal, at the side of the hand) carpals (wrist bones), excluding the pteroid bone, which may itself be a modified distal carpal. The proximal carpals are fused together into a "syncarpal" in mature specimens, while three of the distal carpals fuse to form a distal syncarpal. The remaining distal carpal, referred to here as the medial carpal, but which has also been termed the distal lateral, or pre-axial carpal, articulates on a vertically elongate biconvex facet on the anterior surface of the distal syncarpal. The medial carpal bears a deep concave fovea that opens anteriorly, ventrally and somewhat medially, within which the pteroid articulates, according to Wilkinson. In derived pterodactyloids like pteranodontians and azhdarchoids, metacarpals I-III are small and do not connect to the carpus, instead hanging in contact with the fourth metacarpal. With these derived species, the fourth metacarpal has been enormously elongated, typically equalling or exceeding the length of the long bones of the lower arm. The fifth metacarpal had been lost. In all species, the first to third fingers are much smaller than the fourth, the "wingfinger", and contain two, three and four phalanges respectively. The smaller fingers are clawed, with the ungual size varying among species. In nyctosaurids the forelimb digits besides the wingfinger have been lost altogether. The wingfinger accounts for about half or more of the total wing length. It normally consists of four phalanges. Their relative lengths tend to vary among species, which has often been used to distinguish related forms. The fourth phalanx is usually the shortest. It lacks a claw and has been lost completely by nyctosaurids. It is curved to behind, resulting in a rounded wing tip, which reduces induced drag. The wingfinger is also bent somewhat downwards. When standing, pterosaurs probably rested on their metacarpals, with the outer wing folded to behind. In this position, the "anterior" sides of the metacarpals were rotated to the rear. This would point the smaller fingers obliquely to behind. According to Bennett, this would imply that the wingfinger, able to describe the largest arc of any wing element, up to 175°, was not folded by flexion but by an extreme extension. The wing was automatically folded when the elbow was bowed. The pelvis of pterosaurs was of moderate size compared to the body as a whole. Often the three pelvic bones were fused. The ilium was long and low, its front and rear blades projecting horizontally beyond the edges of the lower pelvic bones. Despite this length, the rod-like form of these processes indicates that the hindlimb muscles attached to them were limited in strength. The, in side view narrow, pubic bone fused with the broad ischium into an ischiopubic blade. Sometimes, the blades of both sides were also fused, closing the pelvis from below and forming the pelvic canal. The hip joint was not perforated and allowed considerable mobility to the leg. It was directed obliquely upwards, preventing a perfectly vertical position of the leg. The front of the pubic bones articulated with a unique structure, the paired prepubic bones. Together these formed a cusp covering the rear belly, between the pelvis and the belly ribs. The vertical mobility of this element suggests a function in breathing, compensating the relative rigidity of the chest cavity. The hindlimbs of pterosaurs were strongly built, yet relative to their wingspans smaller than those of birds. They were long in comparison to the torso length. The thighbone was rather straight, with the head making only a small angle with the shaft. This implies that the legs were not held vertically below the body but were somewhat sprawling. The shinbone was often fused with the upper ankle bones into a tibiotarsus that was longer than the thighbone. It could attain a vertical position when walking. The calf bone tended to be slender, especially at its lower end that in advanced forms did not reach the ankle, sometimes reducing total length to a third. Typically it was fused to the shinbone. The ankle was a simple, "mesotarsal", hinge. The, rather long and slender, metatarsus was always splayed to some degree. The foot was plantigrade, meaning that during the walking cycle the sole of the metatarsus was pressed onto the soil. There was a clear difference between early pterosaurs and advanced species regarding the form of the fifth digit. Originally, the fifth metatarsal was robust and not very shortened. It was connected to the ankle in a higher position than the other metatsarsals. It bore a long, and often curved, mobile clawless fifth toe consisting of two phalanges. The function of this element has been enigmatic. It used to be thought that the animals slept upside-down like bats, hanging from branches and using the fifth toes as hooks. Another hypothesis held that they stretched the brachiopatagia, but in articulated fossils the fifth digits are always flexed towards the tail. Later it became popular to assume that these toes extended an uropatagium or cruropatagium between them. As the fifth toes were on the outside of the feet, such a configuration would only have been possible if these rotated their fronts outwards in flight. Such a rotation could be caused by an abduction of the thighbone, meaning that the legs would be spread. This would also turn the feet into a vertical position. They then could act as rudders to control yaw. Some specimens show membranes between the toes, allowing them to function as flight control surfaces. The (cr)uroptagium would control pitch. When walking the toes could flex upwards to lift the membrane from the ground. In Pterodactyloidea, the fifth metatarsal was much reduced and the fifth toe, if present, little more than a stub. This suggests that their membranes were split, increasing flight manoeuvrability. The first to fourth toes were long. They had two, three, four and five phalanges respectively. Often the third toe was longest; sometimes the fourth. Flat joints indicate a limited mobility. These toes were clawed but the claws were smaller than the hand claws. The rare conditions that allowed for the fossilisation of pterosaur remains, sometimes also preserved soft tissues. Modern synchrotron or ultraviolet light photography has revealed many traces not visible to the naked eye. These are often imprecisely called "impressions" but mostly consist of petrifications, natural casts and transformations of the original material. They may include horn crests, beaks or claw sheaths as well as the various flight membranes. Exceptionally, muscles were preserved. Skin patches show small round non-overlapping scales on the soles of the feet, the ankles and the ends of the metacarpals. They covered pads cushioning the impact of walking. Scales are unknown from other parts of the body. Most or all pterosaurs had hair-like filaments known as pycnofibers on the head and torso. The term "pycnofiber", meaning "dense filament", was coined by palaeontologist Alexander Kellner and colleagues in 2009. Pycnofibers were unique structures similar to, but not homologous (sharing a common origin) with, mammalian hair, an example of convergent evolution. A fuzzy integument was first reported from a specimen of "Scaphognathus crassirostris" in 1831 by Georg Augustus Goldfuss, but had been widely doubted. Since the 1990s, pterosaur finds and histological and ultraviolet examination of pterosaur specimens have provided incontrovertible proof: pterosaurs had pycnofiber coats. "Sordes pilosus" (which translates as "hairy demon") and "Jeholopterus ninchengensis" show pycnofibers on the head and body. The presence of pycnofibers strongly indicates that pterosaurs were endothermic (warm-blooded). They aided thermoregulation, as is common in warm-blooded animals who need insulation to prevent excessive heat-loss. Pycnofibers were flexible, short filaments, about five to seven millimetres long and rather simple in structure with a hollow central canal. Pterosaur pelts might have been comparable in density to many Mesozoic mammals. Pterosaur filaments could share a common origin with feathers, as speculated in 2002 by Czerkas and Ji. In 2009, Kellner concluded that pycnofibers were structured similarly to theropod proto-feathers. Others were unconvinced, considering the difference with the "quills" found on many of the bird-like maniraptoran specimens too fundamental. A 2018 study of the remains of two small Jurassic-age pterosaurs from Inner Mongolia, China, found that pterosaurs had a wide array of pycnofiber shapes and structures, as opposed to the homogeneous structures that had generally been assumed to cover them. Some of these had frayed ends, very similar in structure to four different feather types known from birds or other dinosaurs but almost never known from pterosaurs prior to the study, suggesting homology. Pterosaur fossils are very rare, due to their light bone construction. Complete skeletons can generally only be found in geological layers with exceptional preservation conditions, the so-called "Lagerstätten". The pieces from one such "Lagerstätte", the Late Jurassic Solnhofen Limestone in Bavaria, became much sought after by rich collectors. In 1784, the Italian naturalist Cosimo Alessandro Collini was the first scientist in history to describe a pterosaur fossil. At that time the concepts of evolution and extinction were only imperfectly developed. The bizarre build of the pterosaur was therefore shocking, as it could not clearly be assigned to any existing animal group. The discovery of pterosaurs would thus play an important role in the progress of modern paleontology and geology. If such creatures were still alive, only the sea was a credible habitat and Collini suggested it might be a swimming animal that used its long front limbs as paddles. A few scientists continued to support the aquatic interpretation even until 1830, when the German zoologist Johann Georg Wagler suggested that "Pterodactylus" used its wings as flippers and was affiliated with Ichthyosauria and Plesiosauria. In 1800, Johann Hermann first suggested that it represented a flying creature in a letter to Georges Cuvier. Cuvier agreed in 1801, understanding it was an extinct flying reptile. In 1809, he coined the name "Ptéro-Dactyle", "wing-finger". This was in 1815 Latinised to "Pterodactylus". At first most species were assigned to this genus and ultimately "pterodactyl" was popularly and incorrectly applied to all members of Pterosauria. Today, paleontologists limit the term to the genus "Pterodactylus" or members of the Pterodactyloidea. In 1812 and 1817, Samuel Thomas von Soemmerring redescribed the original specimen and an additional one. He saw them as affiliated to birds and bats. Although he was mistaken in this, his "bat model" would be very influential during the 19th century. In 1843, Edward Newman thought pterosaurs were flying marsupials. As the bat model correctly depicted pterosaurs as furred and warm-blooded, it better approached the true physiology of pterosaurs than Cuvier's "reptile model". In 1834, Johann Jakob Kaup coined the term Pterosauria. In 1828, Mary Anning in England found the first pterosaur genus outside Germany, by Richard Owen named as "Dimorphodon", also the first non-pterodactyloid pterosaur known. Later in the century, the Early Cretaceous Cambridge Greensand produced thousands of pterosaur fossils, that however, were of poor quality, consisting mostly of strongly eroded fragments. Based on these, nevertheless numerous genera and species would be named. Many were described by Harry Govier Seeley, at the time the main English expert on the subject, who also wrote the first pterosaur book, "Ornithosauria", and in 1901 the first popular book, "Dragons of the Air". Seeley thought that pterosaurs were warm-blooded and dynamic creatures, closely related to birds. Earlier, the evolutionist St. George Jackson Mivart had suggested pterosaurs were the direct ancestors of birds. Owen opposed the views of both men, seeing pterosaurs as cold-blooded "true" reptiles. In the US, Othniel Charles Marsh in 1870 discovered "Pteranodon" in the Niobrara Chalk, then the largest known pterosaur, the first toothless one and the first from America. These layers too rendered thousands of fossils, also including relatively complete skeletons that were three-dimensionally preserved instead of being strongly compressed as with the Solnhofen specimens. This led to a much better understanding of many anatomical details, such as the hollow nature of the bones. Meanwhile, finds from the Solnhofen had continued, accounting for the majority of complete high quality specimens discovered. They allowed to identify most new basal taxa, such as "Rhamphorhynchus", "Scaphognathus" and "Dorygnathus". This material gave birth to a German school of pterosaur research, which saw flying reptiles as the warm-blooded, furry and active Mesozoic counterparts of modern bats and birds. In 1882, Marsh and Karl Alfred Zittel published studies about the wing membranes of specimens of "Rhamphorhynchus". German studies continued well into the 1930s, describing new species such as "Anurognathus". In 1927, Ferdinand Broili discovered hair follicles in pterosaur skin, and paleoneurologist Tilly Edinger determined that the brains of pterosaurs more resembled those of birds than modern cold-blooded reptiles. In contrast, English and American paleontologists by the middle of the twentieth century largely lost interest in pterosaurs. They saw them as failed evolutionary experiments, cold-blooded and scaly, that hardly could fly, the larger species only able to glide, being forced to climb trees or throw themselves from cliffs to achieve a take-off. In 1914, for the first time pterosaur aerodynamics were quantitatively analysed, by Ernest Hanbury Hankin and David Meredith Seares Watson, but they interpreted "Pteranodon" as a pure glider. Little research was done on the group during the 1940s and 1950s. The situation for dinosaurs was comparable. From the 1960s onwards, a dinosaur renaissance took place, a quick increase in the number of studies and critical ideas, influenced by the discovery of additional fossils of "Deinonychus", whose spectacular traits refuted what had become entrenched orthodoxy. In 1970, likewise the description of the furry pterosaur "Sordes" began what Robert Bakker named a renaissance of pterosaurs. Especially Kevin Padian propagated the new views, publishing a series of studies depicting pterosaurs as warm-blooded, active and running animals. This coincided with a revival of the German school through the work of Peter Wellnhofer, who in 1970s laid the foundations of modern pterosaur science. In 1978, he published the first pterosaur textbook, the "Handbuch der Paläoherptologie, Teil 19: Pterosauria", and in 1991 the second ever popular science pterosaur book, the "Encyclopedia of Pterosaurs". This development accelerated through the exploitation of two new "Lagerstätten". During the 1970s, the Early Cretaceous Santana Formation in Brazil began to produce chalk nodules that, though often limited in size and the completeness of the fossils they contained, perfectly preserved three-dimensional pterosaur skeletal parts. German and Dutch institutes bought such nodules from fossil poachers and prepared them in Europe, allowing their scientists to describe many new species and revealing a whole new fauna. Soon, Brazilian researchers, among them Alexander Kellner, intercepted the trade and named even more species. Even more productive was the Early Cretaceous Chinese Jehol Biota of Liaoning that since the 1990s has brought forth hundreds of exquisitely preserved two-dimensional fossils, often showing soft tissue remains. Chinese researchers such as Lü Junchang have again named many new taxa. As discoveries also increased in other parts of the world, a sudden surge in the total of named genera took place. By 2009, when they had increased to about ninety, this growth showed no sign of levelling-off. In 2013, M.P. Witton indicated that the number of discovered pterosaur species had risen to 130. Over ninety percent of known taxa has been named during the "renaissance". Many of these were from groups the existence of which had been unknown. Advances in computing power allowed to determine their complex relationships through the quantitative method of cladistics. New and old fossils yielded much more information when subjected to modern ultraviolet light or roentgen photography, or CAT-scans. Insights from other fields of biology were applied to the data obtained. All this resulted in a substantial progress in pterosaur research, rendering older accounts in popular science books completely outdated. Because pterosaur anatomy has been so heavily modified for flight, and immediate transitional fossil predecessors have not so far been described, the ancestry of pterosaurs is not fully understood. The oldest known pterosaurs were already fully adapted to a flying lifestyle. Since Seeley, it was recognised that pterosaurs were likely to have had their origin in the "archosaurs", what today would be called the Archosauromorpha. In the 1980s, early cladistic analyses found that they were Avemetatarsalians (archosaurs closer to dinosaurs than to crocodilians). As this would make them also rather close relatives of the dinosaurs, these results were seen by Kevin Padian as confirming his interpretation of pterosaurs as bipedal warm-blooded animals. Because these early analyses were based on a limited number of taxa and characters, their results were inherently uncertain. Several influential researchers who rejected Padian's conclusions offered alternative hypotheses. David Unwin proposed an ancestry among the basal Archosauromorpha, specifically long-necked forms ("protorosaurs") such as tanystropheids. A placement among basal archosauriforms like "Euparkeria" was also suggested. Some basal archosauromorphs seem at first glance to be good candidates for close pterosaur relatives due to their long-limbed anatomy; one example is "Sharovipteryx", a "protorosaur" with skin membranes on its hindlimbs likely used for gliding. A 1999 study by Michael Benton found that pterosaurs were avemetatarsalians closely related to "Scleromochlus," and named the group Ornithodira to encompass pterosaurs and dinosaurs"." Two researchers, S. Christopher Bennett in 1996, and paleoartist David Peters in 2000, published analyses finding pterosaurs to be protorosaurs or closely related to them. However, Peters gathered novel anatomical data using an unverified technique called "Digital Graphic Segregation" (DGS), which involves digitally tracing over images of pterosaur fossils using photo editing software. Bennett only recovered pterosaurs as close relatives of the protorosaurs after removing characteristics of the hindlimb from his analysis, to test the possibility of locomotion-based convergent evolution between pterosaurs and dinosaurs. A 2007 reply by Dave Hone and Michael Benton could not reproduce this result, finding pterosaurs to be closely related to dinosaurs even without hindlimb characters. They also criticized David Peters for drawing conclusions without access to the primary evidence, that is, the pterosaur fossils themselves. Hone and Benton concluded that, although more basal pterosauromorphs are needed to clarify their relationships, current evidence indicates that pterosaurs are avemetatarsalians, as either the sister group of "Scleromochlus" or a branch between the latter and "Lagosuchus". An 2011 archosaur-focused phylogenetic analysis by Sterling Nesbitt benefited from far more data and found strong support for pterosaurs being avemetatarsalians, though "Scleromochlus" was not included due to its poor preservation. A 2016 archosauromorph-focused study by Martin Ezcurra included various proposed pterosaur relatives, yet also found pterosaurs to be closer to dinosaurs and unrelated to more basal taxa. Working from his 1996 analysis, Bennett published a 2020 study on "Scleromochlus" which argued that both "Scleromochlus" and pterosaurs were non-archosaur archosauromorphs, albeit not particularly closely related to each other. A related problem is the origin of pterosaur flight. Like with birds, hypotheses can be ordered into two main varieties: "ground up" or "tree down". Climbing a tree would cause height and gravity provide both the energy and a strong selection pressure for incipient flight. Rupert Wild in 1983 proposed a hypothetical "propterosaurus": a lizard-like arboreal animal developing a membrane between its limbs, first to safely parachute and then, gradually elongating the fourth finger, to glide. However, subsequent cladistic results did not fit this model well. Neither protorosaurs nor ornithodirans are biologically equivalent to lizards. Furthermore, the transition between gliding and flapping flight is not well-understood. More recent studies on basal pterosaur hindlimb morphology seem to vindicate a connection to "Scleromochlus". Like this archosaur, basal pterosaur lineages have plantigrade hindlimbs that show adaptations for saltation. It was once thought that competition with early bird species might have resulted in the extinction of many of the pterosaurs. Part of this is due to the fact it used to be thought that by the end of the Cretaceous, only large species of pterosaurs were present (no longer true; see below). The smaller species were thought to have become extinct, their niche filled by birds. However, pterosaur decline (if actually present) seems unrelated to bird diversity, as ecological overlap between the two groups appears to be minimal. In fact, at least some avian niches were reclaimed by pterosaurs prior to the KT event. At the end of the Cretaceous period, the Cretaceous–Paleogene extinction event, which wiped out all non-avian dinosaurs and most avian dinosaurs as well, and many other animals, seems also to have taken the pterosaurs. In the early 2010s, several new pterosaur taxa were discovered dating to the Campanian/Maastrichtian, such as the ornithocheirids "Piksi" and ""Ornithocheirus"", possible pteranodontids and nyctosaurids, several tapejarids and the indeterminate non-azhdarchid "Navajodactylus". Small azhdarchoid pterosaurs were also present in the Campanian. This suggests that late Cretaceous pterosaur faunas were far more diverse than previously thought, possibly not even having declined significantly from the early Cretaceous. Small-sized pterosaur species apparently were present in the Csehbánya Formation, indicating a higher diversity of Late Cretaceous pterosaurs than previously accounted for. The recent findings of a small cat-sized adult azhdarchid further indicate that small pterosaurs from the Late Cretaceous might actually have simply been rarely preserved in the fossil record, helped by the fact that there is a strong bias against terrestrial small sized vertebrates such as juvenile dinosaurs, and that their diversity might actually have been much larger than previously thought. At least some non-pterodactyloid pterosaurs survived into the Late Cretaceous, postulating a Lazarus taxa situation for late Cretaceous pterosaur faunas. In phylogenetic taxonomy, the clade Pterosauria has usually been defined as node-based and anchored to several extensively studied taxa as well as those thought to be primitive. One 2003 study defined Pterosauria as "The most recent common ancestor of the Anurognathidae, "Preondactylus" and "Quetzalcoatlus" and all their descendants." However, these types of definition would inevitably leave any related species that are slightly more primitive out of the Pterosauria. To remedy this, a new definition was proposed that would anchor the name not to any particular species but to an anatomical feature, the presence of an enlarged fourth finger that supports a wing membrane. A broader clade, Pterosauromorpha, has been defined as all ornithodirans more closely related to pterosaurs than to dinosaurs. The internal classification of pterosaurs has historically been difficult, because there were many gaps in the fossil record. Starting from the 21st century, new discoveries are now filling in these gaps and giving a better picture of the evolution of pterosaurs. Traditionally, they were organized into two suborders: the Rhamphorhynchoidea, a "primitive" group of long-tailed pterosaurs, and the Pterodactyloidea, "advanced" pterosaurs with short tails. However, this traditional division has been largely abandoned. Rhamphorhynchoidea is a paraphyletic (unnatural) group, since the pterodactyloids evolved directly from them and not from a common ancestor, so, with the increasing use of cladistics, it has fallen out of favor among most scientists. The precise relationships between pterosaurs is still unsettled. Many studies of pterosaur relationships in the past have included limited data and were highly contradictory. However, newer studies using larger data sets are beginning to make things clearer. The cladogram (family tree) below follows a phylogenetic analysis presented by Longrich, Martill and Andres in 2018. The mechanics of pterosaur flight are not completely understood or modeled at this time. Katsufumi Sato, a Japanese scientist, did calculations using modern birds and concluded that it was impossible for a pterosaur to stay aloft. In the book "Posture, Locomotion, and Paleoecology of Pterosaurs" it is theorized that they were able to fly due to the oxygen-rich, dense atmosphere of the Late Cretaceous period. However, both Sato and the authors of "Posture, Locomotion, and Paleoecology of Pterosaurs" based their research on the now-outdated theories of pterosaurs being seabird-like, and the size limit does not apply to terrestrial pterosaurs, such as azhdarchids and tapejarids. Furthermore, Darren Naish concluded that atmospheric differences between the present and the Mesozoic were not needed for the giant size of pterosaurs. Another issue that has been difficult to understand is how they took off. Earlier suggestions were that pterosaurs were largely cold-blooded gliding animals, deriving warmth from the environment like modern lizards, rather than burning calories. In this case, it was unclear how the larger ones of enormous size, with an inefficient cold-blooded metabolism, could manage a bird-like takeoff strategy, using only the hind limbs to generate thrust for getting airborne. Later research shows them instead as being warm-blooded and having powerful flight muscles, and using the flight muscles for walking as quadrupeds. Mark Witton of the University of Portsmouth and Mike Habib of Johns Hopkins University suggested that pterosaurs used a vaulting mechanism to obtain flight. The tremendous power of their winged forelimbs would enable them to take off with ease. Once aloft, pterosaurs could reach speeds of up to and travel thousands of kilometres. In 1985, the Smithsonian Institution commissioned aeronautical engineer Paul MacCready to build a half-scale working model of "Quetzalcoatlus northropi". The replica was launched with a ground-based winch. It flew several times in 1986 and was filmed as part of the Smithsonian's IMAX film "On the Wing". A 2009 study showed that pterosaurs had a lung-and-air-sac system and a precisely controlled skeletal breathing pump, which supports a flow-through pulmonary ventilation model in pterosaurs, analogous to that of birds. The presence of a subcutaneous air sac system in at least some pterodactyloids would have further reduced the density of the living animal. Like modern crocodilians, pterosaurs appeared to have had a hepatic piston, seeing as their shoulder-pectoral girdles were too inflexible to move the sternum as in birds, and they possessed strong gastralia. Thus, their respiratory system had characteristics comparable to both modern archosaur clades. An X-ray study of pterosaur brain cavities revealed that the animals ("Rhamphorhynchus muensteri" and "Anhanguera santanae") had massive flocculi. The flocculus is a brain region that integrates signals from joints, muscles, skin and balance organs. The pterosaurs' flocculi occupied 7.5% of the animals' total brain mass, more than in any other vertebrate. Birds have unusually large flocculi compared with other animals, but these only occupy between 1 and 2% of total brain mass. The flocculus sends out neural signals that produce small, automatic movements in the eye muscles. These keep the image on an animal's retina steady. Pterosaurs may have had such a large flocculus because of their large wing size, which would mean that there was a great deal more sensory information to process. The low relative mass of the flocculi in birds is also a result of birds having a much larger brain overall; though this has been considered an indication that pterosaurs lived in a structurally simpler environment or had less complex behaviour compared to birds, recent studies of crocodilians and other reptiles show that it is common for sauropsids to achieve high intelligence levels with small brains. Studies on the endocast of "Allkaruen" show that brain evolution in pterodactyloids was a modular process. Pterosaurs' hip sockets are oriented facing slightly upwards, and the head of the femur (thigh bone) is only moderately inward facing, suggesting that pterosaurs had an erect stance. It would have been possible to lift the thigh into a horizontal position during flight, as gliding lizards do. There was considerable debate whether pterosaurs ambulated as quadrupeds or as bipeds. In the 1980s, paleontologist Kevin Padian suggested that smaller pterosaurs with longer hindlimbs, such as "Dimorphodon", might have walked or even ran bipedally, in addition to flying, like road runners. However, a large number of pterosaur trackways were later found with a distinctive four-toed hind foot and three-toed front foot; these are the unmistakable prints of pterosaurs walking on all fours. Fossil footprints show that pterosaurs stood with the entire foot in contact with the ground (plantigrade), in a manner similar to many mammals like humans and bears. Footprints from azhdarchids and several unidentified species show that pterosaurs walked with an erect posture with their four limbs held almost vertically beneath the body, an energy-efficient stance used by most modern birds and mammals, rather than the sprawled limbs of modern reptiles. Indeed, erect-limbs may be omnipresent in pterosaurs. Though traditionally depicted as ungainly and awkward when on the ground, the anatomy of some pterosaurs (particularly pterodactyloids) suggests that they were competent walkers and runners. Early pterosaurs have long been considered particularly cumbersome locomotors due to the presence of large cruropatagia, but they too appear to have been generally efficient on the ground. The forelimb bones of azhdarchids and ornithocheirids were unusually long compared to other pterosaurs, and, in azhdarchids, the bones of the arm and hand (metacarpals) were particularly elongated. Furthermore, as a whole, azhdarchid front limbs were proportioned similarly to fast-running ungulate mammals. Their hind limbs, on the other hand, were not built for speed, but they were long compared with most pterosaurs, and allowed for a long stride length. While azhdarchid pterosaurs probably could not run, they would have been relatively fast and energy efficient. The relative size of the hands and feet in pterosaurs (by comparison with modern animals such as birds) may indicate the type of lifestyle pterosaurs led on the ground. Azhdarchid pterosaurs had relatively small feet compared to their body size and leg length, with foot length only about 25%–30% the length of the lower leg. This suggests that azhdarchids were better adapted to walking on dry, relatively solid ground. "Pteranodon" had slightly larger feet (47% the length of the tibia), while filter-feeding pterosaurs like the ctenochasmatoids had very large feet (69% of tibial length in "Pterodactylus", 84% in "Pterodaustro"), adapted to walking in soft muddy soil, similar to modern wading birds. Though clearly forelimb-based launchers, basal pterosaurs have hindlimbs well adapted for hopping, suggesting a connection with archosaurs such as "Scleromochlus". Tracks made by ctenochasmatoids indicate that these pterosaurs swam using their hindlimbs. In general, these have large hindfeet and long torsos, indicating that they were probably more adapted for swimming than other pterosaurs. Pteranodontians conversely have several speciations in their humeri interpreted to have been suggestive of a water-based version of the typical quadrupedal launch, and several like boreopterids must have foraged while swimming, as they seem incapable of frigatebird-like aerial hawking. These adaptations are also seen in terrestrial pterosaurs like azhdarchids, which presumably still needed to launch from water in case they found themselves in it. The nyctosaurid "Alcione" may display adaptations for wing-propelled diving like modern gannets and tropicbirds. Traditionally, almost all pterosaurs were seen as surface-feeding piscivores or fish-eaters, a view that still dominates popular science. Today, many pterosaurs groups are thought to have been terrestrial carnivores, omnivores or insectivores. Early-on it was recognised that the small Anurognathidae were nocturnal, aerial insectivores. With highly flexible joints on the wing finger, a broad, triangular wing shape, large eyes and short tail, these pterosaurs were likely analogous to nightjars or extant insectivorous bats, being capable of high manoeuvrability at relatively low speeds. Interpretations of the habits of basal groups have changed profoundly. "Dimorphodon", envisioned as a puffin analogue in the past, is indicated by its jaw structure, gait, and poor flight capabilities, as a terrestrial/semiarboreal predator of small mammals, squamates, and large insects. Its robust dentition caused "Campylognathoides" to be seen as a generalist or a terrestrial predator of small vertebrates, but the highly robust humerus and high-aspect wing morphology, suggest it may have been capable of grabbing prey on the wing. The small insectivorous "Carniadactylus" and the larger "Eudimorphodon" were highly aerial animals and fast, agile flyers with long robust wings. "Eudimorphodon" has been found with fish remains in its stomach, but its dentition suggests an opportunistic diet. Slender-winged "Austriadactylus" and "Caviramus" were likely terrestrial/semiarboreal generalists. "Caviramus" likely had a strong bite force, indicating an adaptation towards hard food items that might have been chewed in view of the tooth wear. Some Rhamphorhynchidae, such as "Rhamphorhynchus" itself or "Dorygnathus", were fish-eaters with long, slender wings, needle-like dentition and long, thin jaws. "Sericipterus", "Scaphognathus" and "Harpactognathus" had more robust jaws and teeth (which were ziphodont, dagger-shaped, in " Sericipterus"), and shorter, broader wings. These were either terrestrial/aerial predators of vertebrates or corvid-like generalists. Wukongopteridae like "Darwinopterus" were first considered aerial predators. Lacking a robust jaw structure or powerful flying muscles, they are now seen as arboreal or semiterrestrial insectivores. "Darwinopterus robustidens", in particular, seems to have been a beetle specialist. Among pterodactyloids, a greater variation in diet is present. Pteranodontia contained many piscivorous taxa, such as the Ornithocheirae, Boreopteridae, Pteranodontidae and Nyctosauridae. Niche partitioning caused ornithocheirs and the later nyctosaurids to be aerial dip-feeders like today's frigatebirds (with the exception of the plunge-diving adapted "Alcione elainus"), while boreopterids were freshwater diving animals similar to cormorants, and pteranodonts pelagic plunge-divers akin to boobies and gannets. The istiodactylids were likely primarily scavengers. Archaeopterodactyloidea obtained food in coastal or freshwater habitats. "Germanodactylus" and "Pterodactylus" were piscivores, while the Ctenochasmatidae were suspension feeders, using their numerous fine teeth to filter small organisms from shallow water. "Pterodaustro" was adaptated for flamingo-like filter-feeding. In contrast, Azhdarchoidea mostly were terrestrial pterosaurs. Tapejaridae were arboreal omnivores, supplementing seeds and fruits with small insects and vertebrates. Dsungaripteridae were specialist molluscivores, using their powerful jaws to crush the shells of molluscs and crustaceans. Thalassodromidae were likely terrestrial carnivores. "Thalassodromeus" itself was named after a fishing method known as "skim-feeding", later understood to be biomechanically impossible. Perhaps it pursued relatively large prey, in view of its reinforced jaw joints and relatively high bite force. Azhdarchidae are now understood to be terrestrial predators akin to ground hornbills or some storks, eating any prey item they could swallow whole. "Hatzegopteryx" was a robustly built predator of relatively large prey, including medium-sized dinosaurs. "Alanqa" may have been a specialist molluscivore. Pterosaurs are known to have been eaten by theropods. In the 1 July 2004 edition of "Nature", paleontologist Eric Buffetaut discusses an Early Cretaceous fossil of three cervical vertebrae of a pterosaur with the broken tooth of a spinosaur, most likely "Irritator", embedded in it. The vertebrae are known not to have been eaten and exposed to digestion, as the joints are still articulated. While very little is known about pterosaur reproduction, it is believed that, similar to all dinosaurs, all pterosaurs reproduced by laying eggs, though such findings are very rare. The first known pterosaur egg was found in the quarries of Liaoning, the same place that yielded feathered dinosaurs. The egg was squashed flat with no signs of cracking, so evidently the eggs had leathery shells, as in modern lizards. This was supported by the description of an additional pterosaur egg belonging to the genus "Darwinopterus", described in 2011, which also had a leathery shell and, also like modern reptiles but unlike birds, was fairly small compared to the size of the mother. In 2014 five unflattened eggs from the species "Hamipterus tianshanensis" were found in an Early Cretaceous deposit in northwest China. Examination of the shells by scanning electron microscopy showed the presence of a thin calcareous eggshell layer with a membrane underneath. A study of pterosaur eggshell structure and chemistry published in 2007 indicated that it is likely pterosaurs buried their eggs, like modern crocodiles and turtles. Egg-burying would have been beneficial to the early evolution of pterosaurs, as it allows for more weight-reducing adaptations, but this method of reproduction would also have put limits on the variety of environments pterosaurs could live in, and may have disadvantaged them when they began to face ecological competition from birds. A "Darwinopterus" specimen showcases that at least some pterosaurs had a pair of functional ovaries, as opposed to the single functional ovary in birds, dismissing the reduction of functional ovaries as a requirement for powered flight. Wing membranes preserved in pterosaur embryos are well developed, suggesting that pterosaurs were ready to fly soon after birth. However, tomography scans of fossilised "Hamipterus" eggs suggests that the young pterosaurs had well-developed thigh bones for walking, but weak chests for flight. It is unknown if this holds true for other pterosaurs. Fossils of pterosaurs only a few days to a week old (called "flaplings") have been found, representing several pterosaur families, including pterodactylids, rhamphorhinchids, ctenochasmatids and azhdarchids. All preserve bones that show a relatively high degree of hardening ("ossification") for their age, and wing proportions similar to adults. In fact, many pterosaur flaplings have been considered adults and placed in separate species in the past. Additionally, flaplings are normally found in the same sediments as adults and juveniles of the same species, such as the "Pterodactylus" and "Rhamphorhynchus" flaplings found in the Solnhofen limestone of Germany, and "Pterodaustro" flaplings from Argentina. All are found in deep aquatic environment far from shore. For the majority of pterosaur species, it is not known whether they practiced any form of parental care, but their ability to fly as soon as they emerged from the egg and the numerous flaplings found in environments far from nests and alongside adults has led most researchers, including Christopher Bennett and David Unwin, to conclude that the young were dependent on their parents for a relatively short period of time, during a period of rapid growth while the wings grew long enough to fly, and then left the nest to fend for themselves, possibly within days of hatching. Alternatively, they may have used stored yolk products for nourishment during their first few days of life, as in modern reptiles, rather than depend on parents for food. Fossilised "Hamipterus" nests were shown preserving many male and female pterosaurs together with their eggs in a manner to a similar to that of modern seabird colonies. Due to how underdeveloped the chests of the hatchlings were for flying, it was suggested that "Hamipterus" may have practiced some form of parental care. However, this study has since been criticised, so currently most evidence leans towards supreprecociality analogous to that of megapodes, which also fly after birth and do not possess parental care. Growth rates of pterosaurs once they hatched varied across different groups. In more primitive, long-tailed pterosaurs ("rhamphorhynchoids"), such as "Rhamphorhynchus", the average growth rate during the first year of life was 130% to 173%, slightly faster than the growth rate of alligators. Growth in these species slowed after sexual maturity, and it would have taken more than three years for "Rhamphorhynchus" to attain maximum size. In contrast, the more advanced, large pterodactyloid pterosaurs, such as "Pteranodon", grew to adult size within the first year of life. Additionally, pterodactyloids had "determinate growth", meaning that the animals reached a fixed maximum adult size and stopped growing. Comparisons between the scleral rings of pterosaurs and modern birds and reptiles have been used to infer daily activity patterns of pterosaurs. The pterosaur genera "Pterodactylus", "Scaphognathus", and "Tupuxuara" have been inferred to be diurnal, "Ctenochasma", "Pterodaustro", and "Rhamphorhynchus" have been inferred to be nocturnal, and "Tapejara" has been inferred to be cathemeral, being active throughout the day for short intervals. As a result, the possibly fish-eating "Ctenochasma" and "Rhamphorhynchus" may have had similar activity patterns to modern nocturnal seabirds, and the filter-feeding "Pterodaustro" may have had similar activity patterns to modern anseriform birds that feed at night. The differences between activity patterns of the Solnhofen pterosaurs "Ctenochasma", "Rhamphorhynchus", "Scaphognathus", and "Pterodactylus" may also indicate niche partitioning between these genera. Pterosaurs have been a staple of popular culture for as long as their cousins the dinosaurs, though they are usually not featured as prominently in films, literature or other art. While the depiction of dinosaurs in popular media has changed radically in response to advances in paleontology, a mainly outdated picture of pterosaurs has persisted since the mid-20th century. The vague generic term "pterodactyl" is often used for these creatures. The animals depicted frequently represent either "Pteranodon" or (non-pterodactyloid) "Rhamphorhynchus", or a fictionalized hybrid of the two. Many children's toys and cartoons feature "pterodactyls" with "Pteranodon"-like crests and long, "Rhamphorhynchus"-like tails and teeth, a combination that never existed in nature. However, at least one pterosaur "did" have both the "Pteranodon"-like crest and teeth: "Ludodactylus", whose name means "toy finger" for its resemblance to old, inaccurate children's toys. Pterosaurs have sometimes been incorrectly identified as (the ancestors of) "birds", though birds are actually theropod dinosaurs, not closely related to pterosaurs. Pterosaurs were used in fiction in Sir Arthur Conan Doyle's 1912 novel "The Lost World", and subsequent 1925 film adaptation. They appeared in a number of films and television programs since, including the 1933 film "King Kong", and 1966's "One Million Years B.C.". In the latter, animator Ray Harryhausen had to add inaccurate bat-like wing fingers to his stop motion models in order to keep the membranes from falling apart, though this particular error was common in art even before the film was made. The Japanese character Rodan, a fictional giant monster (or "kaiju") which first appeared in the 1956 film "Rodan", is portrayed as an enormous irradiated species of "Pteranodon". Rodan has appeared in multiple "Godzilla" films released during the 1960s, 1970s, 1990s, and 2000s, and also appeared in the 2019 American-produced film "". After the 1960s, pterosaurs remained mostly absent from notable American film appearances until 2001, with "Jurassic Park III". Paleontologist Dave Hone noted that the pterosaurs in this film had not been significantly updated to reflect modern research. Errors persisting were teeth while toothless "Pteranodon" was intended to be depicted, nesting behavior that was known to be inaccurate by 2001, and leathery wings, rather than the taut membranes of muscle fiber required for pterosaur flight. In most media appearances, pterosaurs are depicted as piscivores, not reflecting their full dietary variation. They are also often shown as aerial predators similar to birds of prey, grasping human victims with talons on their feet. However, only the small anurognathid "Vesperopterylus" is known to possesses prehensile feet; all other pterosaurs have flat, plantigrade feet with no opposable toes, and the feet are generally proportionally small, at least in the case of the Pteranodontia.
https://en.wikipedia.org/wiki?curid=24824
Pteranodon Pteranodon (; from Greek πτερόν (pteron, "wing") and ἀνόδων (anodon, "toothless")) is a genus of pterosaur that included some of the largest known flying reptiles, with wingspans over 7 meters (23 feet). They lived during the late Cretaceous geological period of North America in present-day Kansas, Alabama, Nebraska, Wyoming, and South Dakota. More fossil specimens of "Pteranodon" have been found than any other pterosaur, with about 1,200 specimens known to science, many of them well preserved with nearly complete skulls and articulated skeletons. It was an important part of the animal community in the Western Interior Seaway. "Pteranodon" was a pterosaur, meaning that it is not a dinosaur. By definition, all dinosaurs belong to one of the two groups within Dinosauria, i.e. Saurischia or Ornithischia. As such, this excludes pterosaurs. Nonetheless, "Pteranodon" is frequently featured in dinosaur media and are strongly associated with dinosaurs by the general public. While not dinosaurs, pterosaurs such as "Pteranodon" form a sister clade to dinosaurs within the clade Avemetatarsalia. "Pteranodon" was the first pterosaur found outside of Europe. Its fossils first were found by Othniel Charles Marsh in 1870, in the Late Cretaceous Smoky Hill Chalk deposits of western Kansas. These chalk beds were deposited at the bottom of what was once the Western Interior Seaway, a large shallow sea over what now is the midsection of the North American continent. These first specimens, YPM 1160 and YPM 1161, consisted of partial wing bones, as well as a tooth from the prehistoric fish "Xiphactinus", which Marsh mistakenly believed to belong to this new pterosaur (all known pterosaurs up to that point had teeth). In 1871, Marsh named the find ""Pterodactylus oweni"", assigning it to the well-known (but much smaller) European genus "Pterodactylus". Marsh also collected more wing bones of the large pterosaur in 1871. Realizing that the name he had chosen had already been used for Harry Seeley's European pterosaur species "Pterodactylus oweni" in 1864, Marsh re-named his giant North American pterosaur Pterodactylus occidentalis, meaning "Western wing finger," in his 1872 description of the new specimen. He also named two additional species, based on size differences: Pterodactylus ingens (the largest specimen so far), and Pterodactylus velox (the smallest). Meanwhile, Marsh's rival Edward Drinker Cope also had unearthed several specimens of the large North American pterosaur. Based on these specimens, Cope named two new species, Ornithochirus umbrosus and Ornithochirus harpyia, in an attempt to assign them to the large European genus "Ornithocheirus", though he misspelled the name (forgetting the 'e'). Cope's paper naming his species was published in 1872, just five days after Marsh's paper. This resulted in a dispute, fought in the published literature, over whose names had priority in what obviously were the same species. Cope conceded in 1875 that Marsh's names did have priority over his, but maintained that "Pterodactylus umbrosus" was a distinct species (but not genus) from any that Marsh had named previously. Re-evaluation by later scientists has supported Marsh's case, and found that Cope's assertion that "P. umbrosus" was a larger, distinct species were incorrect. While the first "Pteranodon" wing bones were collected by Marsh and Cope in the early 1870s, the first "Pteranodon" skull was found on May 2, 1876, along the Smoky Hill River in Wallace County (now Logan County), Kansas, USA, by Samuel Wendell Williston, a fossil collector working for Marsh. A second, smaller skull soon was discovered as well. These skulls showed that the North American pterosaurs were different from any European species, in that they lacked teeth and had bony crests on their skulls. Marsh recognized this major difference, describing the specimens as "distinguished from all previously known genera of the order Pterosauria by the entire absence of teeth." Marsh recognized that this characteristic warranted a new genus, and he coined the name "Pteranodon" ("wing without tooth") in 1876. Marsh reclassified all the previously named North American species from "Pterodactylus" to "Pteranodon". He considered the smaller skull to belong to "Pteranodon occidentalis", based on its size. Marsh classified the larger skull, YPM 1117, in the new species "Pteranodon longiceps", which he thought to be a medium-sized species in between the small "P. occidentalis" and the large "P. ingens". Marsh also named several additional species: Pteranodon comptus and "Pteranodon nanus" were named for fragmentary skeletons of small individuals, while "Pteranodon gracilis" was based on a wing bone that he mistook for a pelvic bone. He soon realized his mistake, and re-classified that specimen again into a separate genus, which he named "Nyctosaurus". "P. nanus" was also later recognized as a "Nyctosaurus" specimen. In 1892, Samuel Williston examined the question of "Pteranodon" classification. He noticed that, in 1871, Seeley had mentioned the existence of a partial set of toothless pterosaur jaws from the Cambridge Greensand of England, which he named "Ornithostoma". Because the primary characteristic Marsh had used to separate "Pteranodon" from other pterosaurs was its lack of teeth, Williston concluded that "Ornithostoma" must be considered the senior synonym of "Pteranodon". However, in 1901, Pleininger pointed out that "Ornithostoma" had never been scientifically described or even assigned a species name until Williston's work, and therefore had been a "nomen nudum" and could not beat out "Pteranodon" for naming priority. Williston accepted this conclusion and went back to calling the genus "Pteranodon". However, both Williston and Pleininger were incorrect, because unnoticed by both of them was the fact that, in 1891, Seeley himself had finally described and properly named "Ornithostoma", assigning it to the species "O. sedgwicki". In the 2010s, more research on the identity of "Ornithostoma" showed that it was probably not "Pteranodon" or even a close relative, but may in fact have been an azhdarchoid, a different type of toothless pterosaur. Williston was also the first scientist to critically evaluate all of the pteranodont species classified by Cope and Marsh. He agreed with most of Marsh's classification, with a few exceptions. First, he did not believe that "P. ingens" and "P. umbrosus" could be considered synonyms, which even Cope had come to believe. He considered both "P. velox" and "P. longiceps" to be dubious; the first was based on non-diagnostic fragments, and the second, though known from a complete skull, probably belonged to one of the other, previously-named species. In 1903, Williston revisited the question of "Pteranodon" classification, and revised his earlier conclusion that there were seven species down to just three. He considered both "P. comptus" and "P. nanus" to be specimens of "Nyctosaurus", and divided the others into small ("P. velox"), medium ("P. occidentalis"), and large species ("P. ingens"), based primarily on the shape of their upper arm bones. He thought "P. longiceps", the only one known from a skull, could be a synonym of either "P. velox" or "P. occidentalis", based on its size. In 1910, Eaton became the first scientist to publish a more detailed description of the entire "Pteranodon" skeleton, as it was known at the time. He used his findings to revise the classification of the genus once again based on a better understanding of the differences in pteranodont anatomy. Eaton conducted experiments using clay models of bones to help determine the effects of crushing and flattening on the shapes of the arm bones Williston had used in his own classification. Eaton found that most of the differences in bone shapes could be easily explained by the pressures of fossilization, and concluded that no "Pteranodon" skeletons had any significant differences from each other besides their size. Therefore, Eaton was left to decide his classification scheme based on differences in the skulls alone, which he assigned to species just as Marsh did, by their size. In the end, Eaton recognized only three valid species: "P. occidentalis", "P. ingens", and "P. longiceps". The discovery of specimens with upright crests, classified by Harksen in 1966 as the new species "Pteranodon sternbergi", complicated the situation even further, prompting another revision of the genus by Miller in 1972. Because it was impossible to determine crest shape for all of the species based on headless skeletons, Miller concluded that all "Pteranodon" species except the two based on skulls ("P. longiceps" and "P. sternbergi") must be considered "nomena dubia" and abandoned. The skull Eaton thought belonged to "P. ingens" was placed in the new species Pteranodon marshi, and the skull Eaton assigned to "P. occidentalis" was re-named Pteranodon eatoni. Miller also recognized another species based on a skull with a crest similar to that of "P. sternbergi"; Miller named this Pteranodon walkeri. To help bring order to this tangle of names, Miller created three categories or "subgenera" for them. "P. marshi" and "P. longiceps" were placed in the subgenus "Longicepia", though this was later changed to simply "Pteranodon" due to the rules of priority. "P. sternbergi" and "P. walkeri", the upright-crested species, were given the subgenus "Sternbergia", which was later changed to "Geosternbergia" because "Sternbergia" was already in use ("preoccupied"). Finally, Miller named the subgenus "Occidentalia" for "P. eatoni", the skull formerly associated with "P. occidentalis". Miller further expanded the concept of "Pteranodon" to include "Nyctosaurus" as a fourth subgenus. Miller considered these to be an evolutionary progression, with the primitive "Nyctosaurus", at the time thought to be crestless, giving rise to "Occidentalia" (with a small crest), which in turn gave rise to "Pteranodon" with its long backwards crest, finally leading to "Geosternbergia" with its large, upright crest. However, Miller made several mistakes in his study concerning which specimens Marsh had assigned to which species, and most scientists disregarded his work on the subject in their later research, though Wellnhofer (1978) followed Miller's species list. and Schoch (1984) somewhat oddly published another revision that essentially returned to Marsh's original classification scheme, most notably sinking "P. longiceps" as a synonym of "P. ingens". During the early 1990s, S. Christopher Bennett also published several major papers reviewing the anatomy, taxonomy and life history of "Pteranodon". Fragmentary fossils assigned to "Pteranodon" have also been discovered in Skåne, Sweden. "Pteranodon" species are extremely well represented in the fossil record, allowing for detailed descriptions of their anatomy and analysis of their life history. Over 1,000 specimens have been identified, though less than half are complete enough to give researchers good anatomical information. Still, this is more fossils material than is known for any other pterosaur, and it includes both male and female specimens of various age groups and possibly species. Adult "Pteranodon" specimens from the two major species can be divided into two distinct size classes. The smaller class of specimens have small, rounded head crests and very wide pelvic canals, even wider than those of the much larger size class. The size of the pelvic canal probably allowed the laying of eggs, indicating that these smaller adults are females. The larger size class, representing male individuals, have narrow hips and very large crests, which were probably for display. Adult male "Pteranodon" were among the largest pterosaurs, and were the largest flying animals known until the late 20th century, when the giant azhdarchid pterosaurs were discovered. The wingspan of an average adult male "Pteranodon" was . Adult females were much smaller, averaging in wingspan. The largest specimen of "Pteranodon longiceps" from the Niobrara Formation measured from wingtip to wingtip. An even larger specimen is known from the Pierre Shale Formation, with a wingspan of , though this specimen may belong to the distinct genus and species "Geosternbergia maysei". While most specimens are found crushed, enough fossils exist to put together a detailed description of the animal. Methods used to estimate the mass of large male "Pteranodon" specimens (those with wingspans of about 7 meters) have been notoriously unreliable, producing a wide range of estimates from as low as to as high as . In a review of pterosaur size estimates published in 2010, researchers Mark Witton and Mike Habib demonstrated that the latter, largest estimates are almost certainly incorrect given the total volume of a "Pteranodon" body, and could only be correct if the animal "was principally aluminium". Witton and Habib considered the methods used by researchers who obtained smaller mass estimates equally flawed. Most have been produced by scaling modern animals such as bats and birds up to "Pteranodon" size, despite the fact that pterosaurs have vastly different body proportions and soft tissue anatomy from any living animal. Other distinguishing characteristics that set "Pteranodon" apart from other pterosaurs include narrow neural spines on the vertebrae, plate-like bony ligaments strengthening the vertebrae above the hip, and a relatively short tail in which the last few vertebrae are fused into a long rod. The entire length of the tail was about 3.5% as long as the wingspan, or up to in the largest males. Unlike earlier pterosaurs, such as "Rhamphorhynchus" and "Pterodactylus", "Pteranodon" had toothless beaks, similar to those of birds. "Pteranodon" beaks were made of solid, bony margins that projected from the base of the jaws. The beaks were long, slender, and ended in thin, sharp points. The upper jaw, which was longer than the lower jaw, was curved upward; while this normally has been attributed only to the upward-curving beak, one specimen (UALVP 24238) has a curvature corresponding with the beak widening towards the tip. While the tip of the beak is not known in this specimen, the level of curvature suggests it would have been extremely long. The unique form of the beak in this specimen led Alexander Kellner to assign it to a distinct genus, "Dawndraco", in 2010. The most distinctive characteristic of "Pteranodon" is its cranial crest. These crests consisted of skull bones (frontals) projecting upward and backward from the skull. The size and shape of these crests varied due to a number of factors, including age, sex, and species. Male "Pteranodon sternbergi", the older species of the two described to date (and sometimes placed in the distinct genus "Geosternbergia"), had a more vertical crest with a broad forward projection, while their descendants, "Pteranodon longiceps", evolved a narrower, more backward-projecting crest. Females of both species were smaller and bore small, rounded crests. The crests were probably mainly display structures, though they may have had other functions as well. The wing shape of "Pteranodon" suggests that it would have flown rather like a modern-day albatross. This is based on the fact that "Pteranodon" had a high aspect ratio (wingspan to chord length) similar to that of the albatross — 9:1 for "Pteranodon", compared to 8:1 for an albatross. Albatrosses spend long stretches of time at sea fishing, and use a flight pattern called "dynamic soaring" which exploits the vertical gradient of wind speed near the ocean surface to travel long distances without flapping, and without the aid of thermals (which do not occur over the open ocean the same way they do over land). While most of a "Pteranodon" flight would have depended on soaring, like long-winged seabirds, it probably required an occasional active, rapid burst of flapping, and studies of "Pteranodon" wing loading (the strength of the wings vs. the weight of the body) indicate that they were capable of substantial flapping flight, contrary to some earlier suggestions that they were so big they could only glide. Like other pterosaurs, "Pteranodon" probably took off from a standing, quadrupedal position. Using their long forelimbs for leverage, they would have vaulted themselves into the air in a rapid leap. Almost all of the energy would have been generated by the forelimbs. The upstroke of the wings would have occurred when the animal cleared the ground followed by a rapid down-stroke to generate additional lift and complete the launch into the air. Historically, the terrestrial locomotion of "Pteranodon", especially whether it was bipedal or quadrupedal, has been the subject of debate. Today, most pterosaur researchers agree that pterosaurs were quadrupedal, thanks largely to the discovery of pterosaur trackways. The possibility of aquatic locomotion via swimming has been discussed briefly in several papers (Bennett 2001, 1994, and Bramwell & Whitfield 1974). The diet of "Pteranodon" is known to have included fish; fossilized fish bones have been found in the stomach area of one "Pteranodon", and a fossilized fish bolus has been found between the jaws of another "Pteranodon", specimen AMNH 5098. Numerous other specimens also preserve fragments of fish scales and vertebrae near the torso, indicating that fish made up a majority of the diet of "Pteranodon" (though they may also have taken invertebrates). Traditionally, most researchers have suggested that "Pteranodon" would have taken fish by dipping their beaks into the water while in low, soaring flight. However, this was probably based on the assumption that the animals could not take off from the water surface. It is more likely that "Pteranodon" could take off from the water, and would have dipped for fish while swimming rather than while flying. Even a small, female "Pteranodon" could have reached a depth of at least with its long bill and neck while floating on the surface, and they may have reached even greater depths by plunge-diving into the water from the air like some modern long-winged seabirds. In 1994, Bennett noted that the head, neck, and shoulders of "Pteranodon" were as heavily built as diving birds, and suggested that they could dive by folding back their wings like the modern gannet. "Pteranodon" was notable for its skull crest, though the function of this crest has been a subject of debate. Most explanations have focused on the blade-like, backward pointed crest of male "P. longiceps", however, and ignored the wide range of variation across age and sex. The fact that the crests vary so much rules out most practical functions other than for use in mating displays. Therefore, display was probably the main function of the crest, and any other functions were secondary. Scientific interpretations of the crest's function began in 1910, when George Francis Eaton proposed two possibilities: an aerodynamic counterbalance and a muscle attachment point. He suggested that the crest might have anchored large, long jaw muscles, but admitted that this function alone could not explain the large size of some crests. Bennett (1992) agreed with Eaton's own assessment that the crest was too large and variable to have been a muscle attachment site. Eaton had suggested that a secondary function of the crest might have been as a counterbalance against the long beak, reducing the need for heavy neck muscles to control the orientation of the head. Wind tunnel tests showed that the crest did function as an effective counterbalance to a degree, but Bennett noted that, again, the hypothesis focuses only on the long crests of male "P. longiceps", not on the larger crests of "P. sternbergi" and very small crests that existed among the females. Bennett found that the crests of females had no counterbalancing effect, and that the crests of male "P. sternbergi" would, by themselves, have a negative effect on the balance of the head. In fact, side to side movement of the crests would have required more, not less, neck musculature to control balance. In 1943, Dominik von Kripp suggested that the crest may have served as a rudder, an idea embraced by several later researchers. One researcher, Ross S. Stein, even suggested that the crest may have supported a membrane of skin connecting the backward-pointing crest to the neck and back, increasing its surface area and effectiveness as a rudder. The rudder hypothesis, again, does not take into account females nor "P. sternbergi", which had an upward-pointing, not backward-pointing crest. Bennett also found that, even in its capacity as a rudder, the crest would not provide nearly so much directional force as simply maneuvering the wings. The suggestion that the crest was an air brake, and that the animals would turn their heads to the side in order to slow down, suffers from a similar problem. Additionally, the rudder and air brake hypotheses do not explain why such large variation exists in crest size even among adults. Alexander Kellner suggested that the large crests of the pterosaur "Tapejara", as well as other species, might be used for heat exchange, allowing these pterosaurs to absorb or shed heat and regulate body temperature, which also would account for the correlation between crest size and body size. There is no evidence of extra blood vessels in the crest for this purpose, however, and the large, membranous wings filled with blood vessels would have served that purpose much more effectively. With these hypotheses ruled out, the best-supported hypothesis for crest function seems to be as a sexual display. This is consistent with the size variation seen in fossil specimens, where females and juveniles have small crests and males large, elaborate, variable crests. Adult "Pteranodon" specimens may be divided into two distinct size classes, small and large, with the large size class being about one and a half times larger than the small class, and the small class being twice as common as the large class. Both size classes lived alongside each other, and while researchers had previously suggested that they represent different species, Christopher Bennett showed that the differences between them are consistent with the concept that they represent females and males, and that "Pteranodon" species were sexually dimorphic. Skulls from the larger size class preserve large, upward and backward pointing crests, while the crests of the smaller size class are small and triangular. Some larger skulls also show evidence of a second crest that extended long and low, toward the tip of the beak, which is not seen in smaller specimens. The sex of the different size classes was determined, not from the skulls, but from the pelvic bones. Contrary to what may be expected, the smaller size class had disproportionately large and wide-set pelvic bones. Bennett interpreted this as indicating a more spacious birth canal, through which eggs would pass. He concluded that the small size class with small, triangular crests represent females, and the larger, large-crested specimens represent males. Note that the overall size and crest size also corresponds to age. Immature specimens are known from both females and males, and immature males often have small crests similar to adult females. Therefore, it seems that the large crests only developed in males when they reached their large, adult size, making the sex of immature specimens difficult to establish from partial remains. The fact that females appear to have outnumbered males two to one suggests that, as with modern animals with size-related sexual dimorphism, such as sea lions and other pinnipeds, "Pteranodon" might have been polygynous, with a few males competing for association with groups consisting of large numbers of females. Similar to modern pinnipeds, "Pteranodon" may have competed to establish territory on rocky, offshore rookeries, with the largest, and largest-crested, males gaining the most territory and having more success mating with females. The crests of male "Pteranodon" would not have been used in competition, but rather as "visual dominance-rank symbols", with display rituals taking the place of physical competition with other males. If this hypothesis is correct, it also is likely that male "Pteranodon" played little to no part in rearing the young; such a behavior is not found in the males of modern polygynous animals who father many offspring at the same time. Specimens assigned to "Pteranodon" have been found in both the Smoky Hill Chalk deposits of the Niobrara Formation, and the slightly younger Sharon Springs deposits of the Pierre Shale Formation. When "Pteranodon" was alive, this area was covered by a large inland sea, known as the Western Interior Seaway. Famous for fossils collected since 1870, these formations extend from as far south as Kansas in the United States to Manitoba in Canada. However, "Pteranodon" specimens (or any pterosaur specimens) have only been found in the southern half of the formation, in Kansas, Wyoming, and South Dakota. Despite the fact that numerous fossils have been found in the contemporary parts of the formation in Canada, no pterosaur specimens have ever been found there. This strongly suggests that the natural geographic range of "Pteranodon" covered only the southern part of the Niobrara, and that its habitat did not extend farther north than South Dakota. Some very fragmentary fossils belonging to pteranodontian pterosaurs, and possibly "Pteranodon" itself, have also been found on the Gulf Coast and East Coast of the United States. For example, some bone fragments from the Mooreville Formation of Alabama and the Merchantville Formation of Delaware may have come from "Pteranodon", though they are too incomplete to make a definite identification. Some remains from Japan have also been tentatively attributed to "Pteranodon", but their distance from its known Western Interior Seaway habitat makes this identification unlikely. "Pteranodon longiceps" would have shared the sky with the giant-crested pterosaur "Nyctosaurus". Compared to "P. longiceps", which was a very common species, "Nyctosaurus" was rare, making up only 3% of pterosaur fossils from the formation. Also less common was the early toothed bird, "Ichthyornis". It is likely that, as in other polygynous animals (in which males compete for association with harems of females), "Pteranodon" lived primarily on offshore rookeries, where they could nest away from land-based predators and feed far from shore; most "Pteranodon" fossils are found in locations which at the time, were hundreds of kilometres from the coastline. Below the surface, the sea was populated primarily by invertebrates such as ammonites and squid. Vertebrate life, apart from basal fish, included sea turtles, such as "Toxochelys", the plesiosaur "Styxosaurus", and the flightless diving bird "Parahesperornis". Mosasaurs were the most common marine reptiles, with genera including "Clidastes" and "Tylosaurus". At least some of these marine reptiles are known to have fed on "Pteranodon". Barnum Brown, in 1904, reported plesiosaur stomach contents containing "pterodactyl" bones, most likely from "Pteranodon". Fossils from terrestrial dinosaurs also have been found in the Niobrara Chalk, suggesting that animals who died on shore must have been washed out to sea (one specimen of a hadrosaur appears to have been scavenged by a shark). "Pteranodon" fossils are known primarily from the Niobrara Formation of the central United States. Broadly defined, "Pteranodon" existed for more than four million years, during the late Coniacian to early Campanian stages of the Cretaceous period. The genus is present in most layers of the Niobrara Formation except for the upper two; in 2003, Kenneth Carpenter surveyed the distribution and dating of fossils in this formation, demonstrating that "Pteranodon sternbergi" existed there from 88 to 85 million years ago, while "P. longiceps" existed between 86 and 84.5 million years ago. A possible third species, which Kellner named "Geosternbergia maysei" in 2010, is known from the Sharon Springs member of the Pierre Shale Formation in Kansas, Wyoming, and South Dakota, dating to between 81.5 and 80.5 million years ago. In the early 1990s, Bennett noted that the two major morphs of pteranodont present in the Niobrara Formation were precisely separated in time with little, if any, overlap. Due to this, and to their gross overall similarity, he suggested that they probably represent "chronospecies" within a single evolutionary lineage lasting about 4 million years. In other words, only one species of "Pteranodon" would have been present at any one time, and "P. sternbergi" (or "Geosternbergia") in all likelihood was the direct ancestor species of "P. longiceps". Many researchers consider there to be at least two species of "Pteranodon". However, aside from the differences between males and females described above, the post-cranial skeletons of "Pteranodon" show little to no variation between species or specimens, and the bodies and wings of all pteranodonts were essentially identical. Two species of "Pteranodon" are traditionally recognized as valid: "Pteranodon longiceps", the type species, and "Pteranodon sternbergi". The species differ only in the shape of the crest in adult males (described above), and possibly in the angle of certain skull bones. Because well-preserved "Pteranodon" skull fossils are extremely rare, researchers use stratigraphy (i.e. which rock layer of the geologic formation a fossil is found in) to determine species identity in most cases. "Pteranodon sternbergi" is the only known species of "Pteranodon" with an upright crest. The lower jaw of "P. sternbergi" was long. It was collected by George F. Sternberg in 1952 and described by John Christian Harksen in 1966, from the lower portion of the Niobrara Formation. It was older than "P. longiceps" and is considered by Bennett to be the direct ancestor of the later species. Because fossils identifiable as "P. sternbergi" are found exclusively in the lower layers of the Niobrara Formation, and "P. longiceps" fossils exclusively in the upper layers, a fossil lacking the skull can be identified based on its position in the geologic column (though for many early fossil finds, precise data about its location was not recorded, rendering many fossils unidentifiable). Below is a cladogram showing the phylogenetic placement of this genus within Pteranodontia from Andres and Myers (2013). Due to the subtle variations between specimens of pteranodontid from the Niobrara Formation, most researchers have assigned all of them to the single genus "Pteranodon", in at least two species ("P. longiceps" and "P. sternbergi") distinguished mainly by the shape of the crest. However, the classification of these two forms has varied from researcher to researcher. In 1972, Halsey Wilkinson Miller published a paper arguing that the various forms of "Pteranodon" were different enough to be placed in distinct subgenera. He named these "Pteranodon (Occidentalia) occidentalis" (for the now-disused species "P. occidentalis") and "Pteranodon (Sternbergia) sternbergi". However, the name "Sternbergia" was preoccupied, and in 1978 Miller re-named the species "Pteranodon (Geosternbergia) sternbergi", and named a third subgenus/species combination for "P. longiceps", as "Pteranodon (Longicepia) longiceps". Most prominent pterosaur researchers of the late 20th century however, including S. Christopher Bennett and Peter Wellnhofer, did not adopt these subgeneric names, and continued to place all pteranodont species into the single genus "Pteranodon". In 2010, pterosaur researcher Alexander Kellner revisited H.W. Miller's classification. Kellner followed Miller's opinion that the differences between the "Pteranodon" species were great enough to place them into different genera. He placed "P. sternbergi" into the genus named by Miller, "Geosternbergia", along with the Pierre Shale skull specimen which Bennett had previously considered to be a large male "P. longiceps". Kellner argued that this specimen's crest, though incompletely preserved, was most similar to "Geosternbergia". Because the specimen was millions of years younger than any known "Geosternbergia", he assigned it to the new species "Geosternbergia maysei". Numerous other pteranodont specimens are known from the same formation and time period, and Kellner suggested they may belong to the same species as "G. maysei", but because they lack skulls, he could not confidently identify them. A number of additional species of "Pteranodon" have been named since the 1870s, although most now are considered to be junior synonyms of two or three valid species. The best-supported is the type species, "P. longiceps", based on the well-preserved specimen including the first-known skull found by S. W. Williston. This individual had a wingspan of . Other valid species include the possibly larger "P. sternbergi", with a wingspan originally estimated at . "P. oweni" ("P. occidentalis"), "P. velox", "P. umbrosus", "P. harpyia", and "P. comptus" are considered to be "nomina dubia" by Bennett (1994) and others who question their validity. All probably are synonymous with the more well-known species. Because the key distinguishing characteristic Marsh noted for "Pteranodon" was its lack of teeth, any toothless pterosaur jaw fragment, wherever it was found in the world, tended to be attributed to "Pteranodon" during the late nineteenth and early twentieth centuries. This resulted in a plethora of species and a great deal of confusion. The name became a wastebasket taxon, rather like the dinosaur "Megalosaurus", to label any pterosaur remains that could not be distinguished other than by the absence of teeth. Species (often dubious ones now known to be based on sexual variation or juvenile characters) have been reclassified a number of times, and several subgenera have in the 1970s been erected by Halsey Wilkinson Miller to hold them in various combinations, further confusing the taxonomy (subgenera include "Longicepia", "Occidentalia", and "Geosternbergia"). Notable authors who have discussed the various aspects of "Pteranodon" include Bennett, Padian, Unwin, Kellner, and Wellnhofer. Two species, "P. orogensis" and "P. orientalis", are not pteranodontids and have been renamed "Bennettazhia oregonensis" and "Bogolubovia orientalis" respectively. Status of names listed below follow a survey by Bennett, 1994 unless otherwise noted.
https://en.wikipedia.org/wiki?curid=24825
Passive voice A passive voice construction is a grammatical voice construction that is found in many languages. In a clause with passive voice, the grammatical subject expresses the "theme" or "patient" of the main verb – that is, the person or thing that undergoes the action or has its state changed. This contrasts with active voice, in which the subject has the agent role. For example, in the passive sentence "The tree was pulled down", the subject ("the tree") denotes the patient rather than the agent of the action. In contrast, the sentences "Someone pulled down the tree" and "The tree is down" are active sentences. Typically, in passive clauses, what is usually expressed by the object (or sometimes another argument) of the verb is now expressed by the subject, while what is usually expressed by the subject is either deleted or is indicated by some adjunct of the clause. Thus, turning an active verb into a passive verb is a valence-decreasing process ("detransitivizing process"), because it turns transitive verbs into intransitive verbs. This is not always the case; for example in Japanese a passive-voice construction does not necessarily decrease valence. Many languages have both an active and a passive voice; this allows for greater flexibility in sentence construction, as either the semantic agent or patient may take the syntactic role of subject. The use of passive voice allows speakers to organize stretches of discourse by placing figures other than the agent in subject position. This may be done to foreground the patient, recipient, or other thematic role; it may also be useful when the semantic patient is the topic of on-going discussion. The passive voice may also be used to avoid specifying the agent of an action. Different languages use various grammatical forms to indicate passive voice. In some languages, passive voice is indicated by verb conjugation, specific forms of the verb. Examples of languages that indicate voice through conjugation include Latin and North Germanic languages such as Swedish. Norwegian (Nynorsk) and Icelandic have a similar system, but the usage of the passive is more restricted. The passive forms in Nynorsk are restricted to only be accompanied by an auxiliary verb, which is not the case in Swedish and Danish. In Latin, the agent of a passive sentence (if indicated) is expressed using a noun in the ablative case, in this case "servō" (the ablative of "servus"). Different languages use different methods for expressing the agent in passive clauses. In Swedish, the agent can be expressed by means of a prepositional phrase with the preposition "av" (equivalent here to the English "by"). The Austronesian language Kimaragang Dusun also indicates passive voice by verb conjugation using the infix, "-in-". Other languages, including English, express the passive voice periphrastically, using an auxiliary verb. English, like some other languages, uses a periphrastic passive. Rather than conjugating directly for voice, English uses the past participle form of the verb plus an auxiliary verb, either "be" or "get" (called linking verbs in traditional grammar), to indicate passive voice. If the agent is mentioned, it usually appears in a prepositional phrase introduced by the preposition "by". The subject of the passive voice usually corresponds to the direct object of the corresponding active-voice formulation (as in the above examples), but English also allows passive constructions in which the subject corresponds to an indirect object or preposition complement: In sentences of the second type, a stranded preposition is left. This is called the "prepositional passive" or "pseudo-passive" (although the latter term can also be used with other meanings). The active voice is the dominant voice used in English. Many commentators, notably George Orwell in his essay "Politics and the English Language" and Strunk & White in "The Elements of Style", have urged minimizing use of the passive voice, but this is almost always based on these commentators' misunderstanding of what the passive voice is. Contrary to common critiques, the passive voice has important uses, with virtually all writers using the passive voice (including Orwell and Strunk & White). There is general agreement that the passive voice is useful for emphasis, or when the receiver of the action is more important than the actor. "Merriam–Webster's Dictionary of English Usage" refers to three statistical studies of passive versus active sentences in various periodicals, stating: "the highest incidence of passive constructions was 13 percent. Orwell runs to a little over 20 percent in "Politics and the English Language". Clearly he found the construction useful in spite of his advice to avoid it as much as possible". In the field of linguistics, the term "passive" is applied to a wide range of grammatical structures. Linguists therefore find it difficult to define the term in a way that makes sense across all human languages. The canonical passive in European languages has the following properties: The problem arises with non-European languages. Many constructions in these languages share at least one property with the canonical European passive, but not all. While it seems justified to call these constructions "passive" when comparing them to European languages' passive constructions, as a whole the passives of the world's languages do not share a single common feature. R. M. W. Dixon has defined four criteria for determining whether a construction is a passive: Dixon acknowledges that this excludes some constructions labeled as "passive" by some linguists. In some languages, including several Southeast Asian languages, the passive voice is sometimes used to indicate that an action or event was unpleasant or undesirable. This so-called "adversative passive" works like the ordinary passive voice in terms of syntactic structure—that is, a theme or instrument acts as subject. In addition, the construction indicates adversative affect, suggesting that someone was negatively affected. In Japanese, for example, the adversative passive (also called indirect passive) indicates adversative affect. The indirect or adversative passive has the same form as the direct passive. Unlike the direct passive, the indirect passive may be used with intransitive verbs. Yup'ik, from the Eskimo-Aleut family, has two different suffixes that can indicate passive, "-cir-" and "-ma-". The morpheme "-cir-" has an adversative meaning. If an agent is included in a passive sentence with the "-cir" passive, the noun is usually in the allative (oblique) case. In some languages, for example English, there is often a similarity between clauses expressing an action or event in the passive voice and clauses expressing a state. For example, the string of words "The dog is fed" can have the following two different meanings: The additions in parentheses "force" the same string of words to clearly show only one of their two possible grammatical functions and the related meaning. In the first sentence, the combination of the auxiliary verb "is" and the past participle "fed" is a regular example of the construction of the passive voice in English. In the second sentence, "is" can however be interpreted as an ordinary copula and the past participle as an adjective. Sentences of the second type are called "false passives" by some linguists, who feel that such sentences are simply confused with the passive voice due to their outward similarity. Other linguists consider the second type to be a different kind of passive – a "stative passive" (rarely called "statal", "static", or "resultative passive"), in contrast to the "dynamic" or "eventive" passive illustrated by the first sentence. Some languages express or can express these different meanings using different constructions. The difference between dynamic and stative passives is more evident in languages such as German that use different words or constructions for the two. In German, the auxiliary verb "sein" marks static passive (German: "Zustandspassiv", rarely "statisches Passiv", in referring to German also called ""sein"-Passiv" or "Sein-Passiv"), while "werden" marks the dynamic passive ("Vorgangspassiv" or "Handlungspassiv", rarely "dynamisches Passiv", in referring to German also called ""werden"-Passiv" or "Werden-Passiv" or simply "Passiv" or "Passivum"). The English string of words "the lawn is mown" has two possible meanings corresponding to the example "the dog is fed" above. It can be used in the following two different senses: German uses two different grammatical constructions for these sentences: Further examples and explanations: A number of German verbs such as "bedecken" ("cover"), "erfüllen" ("fill"), and "trennen" ("separate"), when used as stative verbs, usually only form static passives. In English, the passive voice expressed with the auxiliary verb "get" rather than "be" ("get-passive") expresses a dynamic rather than a static meaning. But when the auxiliary verb "be" is used, the main verb can have either a dynamic or static meaning as shown below (including copies of some examples from above): Verbs that typically express static meaning can show dynamic meaning when used in the passive formed with "get", for example "be known" (static) vs. "get known" (dynamic): (https://wals.info/chapter/107)
https://en.wikipedia.org/wiki?curid=24826
Primitive recursive function In computability theory, a primitive recursive function is roughly speaking a function that can be computed by a computer program whose loops are all "for" loops (that is, an upper bound of the number of iterations of every loop can be determined before entering the loop). Primitive recursive functions form a strict subset of those general recursive functions that are also total functions. The importance of primitive recursive functions lies on the fact that most computable functions that are studied in number theory (and more generally in mathematics) are primitive recursive. For example, addition and division, the factorial and exponential function, and the function which returns the "n"th prime are all primitive recursive. In fact, for showing that a computable function is primitive recursive, it suffices to show that its computational complexity is bounded above by a primitive recursive function of the input size. It follows that it is difficult to devise a computable function that is "not" primitive recursive, although some are known (see the section on Limitations below). The set of primitive recursive functions is known as PR in computational complexity theory. The primitive recursive functions are among the number-theoretic functions, which are functions from the natural numbers (nonnegative integers) {0, 1, 2, ...} to the natural numbers. These functions take "n" arguments for some natural number "n" and are called "n"-ary. The basic primitive recursive functions are given by these axioms: More complex primitive recursive functions can be obtained by applying the operations given by these axioms: Example. We take "f"("x") as the "S"("x") defined above. This f is a 1-ary primitive recursive function. And so is "g"("x") = "S"("x"). So "h"("x") defined as "f"("g"("x")) = "S"("S"("x")) is a primitive recursive 1-ary function too. Informally speaking, "h"("x") is the function that turns "x" into "x"+2. Example. Suppose "f"("x") = "P"11("x") = "x" and "g"("x","y","z")= "S"("P"23("x","y","z")) = "S"("y"). Then "h"(0,"x") = "x" and "h"("S"("y"),"x") = "g"("y","h"("y","x"),"x") = "S"("h"("y","x")). Now "h"(0,1) = 1, "h"(1,1) = "S"("h"(0,1)) = 2, "h"(2,1) = "S"("h"(1,1)) = 3. This "h" is a 2-ary primitive recursive function. We can call it 'addition'. Interpretation. The function "h" acts as a for loop from 0 up to the value of its first argument. The rest of the arguments for "h", denoted here with "x""i"’s ("i" = 1, ..., "k"), are a set of initial conditions for the For loop which may be used by it during calculations but which are immutable by it. The functions "f" and "g" on the right side of the equations which define "h" represent the body of the loop, which performs calculations. Function "f" is only used once to perform initial calculations. Calculations for subsequent steps of the loop are performed by "g". The first parameter of "g" is fed the “current” value of the For loop’s index. The second parameter of "g" is fed the result of the For loop’s previous calculations, from previous steps. The rest of the parameters for "g" are those immutable initial conditions for the For loop mentioned earlier. They may be used by "g" to perform calculations but they will not themselves be altered by "g". The primitive recursive functions are the basic functions and those obtained from the basic functions by applying these operations a finite number of times. The projection functions can be used to avoid the apparent rigidity in terms of the arity of the functions above; by using compositions with various projection functions, it is possible to pass a subset of the arguments of one function to another function. For example, if "g" and "h" are 2-ary primitive recursive functions then is also primitive recursive. One formal definition using projection functions is In some settings it is natural to consider primitive recursive functions that take as inputs tuples that mix numbers with truth values (that is "t" for true and "f" for false), or that produce truth values as outputs. This can be accomplished by identifying the truth values with numbers in any fixed manner. For example, it is common to identify the truth value "t" with the number 1 and the truth value "f" with the number 0. Once this identification has been made, the characteristic function of a set "A", which always returns 1 or 0, can be viewed as a predicate that tells whether a number is in the set "A". Such an identification of predicates with numeric functions will be assumed for the remainder of this article. An example of a primitive recursive programming language is one that contains basic arithmetic operators (e.g. + and −, or ADD and SUBTRACT), conditionals and comparison (IF-THEN, EQUALS, LESS-THAN), and bounded loops, such as the basic for loop, where there is a known or calculable upper bound to all loops (FOR i FROM 1 TO n, with neither i nor n modifiable by the loop body). No control structures of greater generality, such as while loops or IF-THEN plus GOTO, are admitted in a primitive recursive language. Douglas Hofstadter's BlooP in "Gödel, Escher, Bach" is such a language. Adding unbounded loops (WHILE, GOTO) makes the language partially recursive, or Turing-complete; Floop is an example, as are almost all real-world computer programming languages. Arbitrary computer programs, or Turing machines, cannot in general be analyzed to see if they halt or not (the halting problem). However, all primitive recursive functions halt. This is not a contradiction; primitive recursive programs are a non-arbitrary subset of all possible programs, constructed specifically to be analyzable. Most number-theoretic functions definable using recursion on a single variable are primitive recursive. Basic examples include the addition and truncated subtraction functions. Intuitively, addition can be recursively defined with the rules: To fit this into a strict primitive recursive definition, define: Here S("n") is "the successor of "n"" (i.e., "n"+1), "P"11 is the identity function, and "P"23 is the projection function that takes 3 arguments and returns the second one. Functions "f" and "g" required by the above definition of the primitive recursion operation are respectively played by "P"11 and the composition of "S" and "P"23. Because primitive recursive functions use natural numbers rather than integers, and the natural numbers are not closed under subtraction, a truncated subtraction function (also called "proper subtraction") is studied in this context. This limited subtraction function sub("a", "b") [or "b" ∸ "a"] returns "b" - "a" if this is nonnegative and returns "0" otherwise. The predecessor function acts as the opposite of the successor function and is recursively defined by the rules: These rules can be converted into a more formal definition by primitive recursion: The limited subtraction function is definable from the predecessor function in a manner analogous to the way addition is defined from successor: Here sub("a", "b") corresponds to "b" ∸ "a"; for the sake of simplicity, the order of the arguments has been switched from the "standard" definition to fit the requirements of primitive recursion. This could easily be rectified using composition with suitable projections. Exponentiation and primality testing are primitive recursive. Given primitive recursive functions "e", "f", "g", and "h", a function that returns the value of "g" when "e"≤"f" and the value of "h" otherwise is primitive recursive. By using Gödel numberings, the primitive recursive functions can be extended to operate on other objects such as integers and rational numbers. If integers are encoded by Gödel numbers in a standard way, the arithmetic operations including addition, subtraction, and multiplication are all primitive recursive. Similarly, if the rationals are represented by Gödel numbers then the field operations are all primitive recursive. In first-order Peano arithmetic, there are infinitely many variables (0-ary symbols) but no k-ary non-logical symbols with k>0 other than S, +, *, and ≤. Thus in order to define primitive recursive functions one has to use the following trick by Gödel. By using a Gödel numbering for sequences, for example Gödel's β function, any finite sequence of numbers can be encoded by a single number. Such a number can therefore represent the primitive recursive function until a given n. Let "h" be a 1-ary primitive recursion function defined by: where C is a constant and "g" is an already defined function. Using Gödel's β function, for any sequence of natural numbers (k0, k1, …, kn), there are natural numbers b and c such that, for every i ≤ n, β(b, c, i) = ki. We may thus use the following formula to define "h"; more precisely, "m"="h"("n") is a shorthand for the following: and the equating to "g", being already defined, is in fact shorthand for some other already defined formula (as is β, whose formula is given here). The generalization to any k-ary primitive recursion function is trivial. The broader class of partial recursive functions is defined by introducing an unbounded search operator. The use of this operator may result in a partial function, that is, a relation with "at most" one value for each argument, but does not necessarily have "any" value for any argument (see domain). An equivalent definition states that a partial recursive function is one that can be computed by a Turing machine. A total recursive function is a partial recursive function that is defined for every input. Every primitive recursive function is total recursive, but not all total recursive functions are primitive recursive. The Ackermann function "A"("m","n") is a well-known example of a total recursive function (in fact, provable total), that is not primitive recursive. There is a characterization of the primitive recursive functions as a subset of the total recursive functions using the Ackermann function. This characterization states that a function is primitive recursive if and only if there is a natural number "m" such that the function can be computed by a Turing machine that always halts within A("m","n") or fewer steps, where "n" is the sum of the arguments of the primitive recursive function. An important property of the primitive recursive functions is that they are a recursively enumerable subset of the set of all total recursive functions (which is not itself recursively enumerable). This means that there is a single computable function "f"("m","n") that enumerates the primitive recursive functions, namely: "f" can be explicitly constructed by iteratively repeating all possible ways of creating primitive recursive functions. Thus, it is provably total. One can use a diagonalization argument to show that "f" is not recursive primitive in itself: had it been such, so would be "h"("n") = "f"("n","n")+1. But if this equals some primitive recursive function, there is an "m" such that "h"("n") = "f"("m","n") for all "n", and then "h"("m") = "f"("m","m"), leading to contradiction. However, the set of primitive recursive functions is not the "largest" recursively enumerable subset of the set of all total recursive functions. For example, the set of provably total functions (in Peano arithmetic) is also recursively enumerable, as one can enumerate all the proofs of the theory. While all primitive recursive functions are provably total, the converse is not true. Primitive recursive functions tend to correspond very closely with our intuition of what a computable function must be. Certainly the initial functions are intuitively computable (in their very simplicity), and the two operations by which one can create new primitive recursive functions are also very straightforward. However, the set of primitive recursive functions does not include every possible total computable function—this can be seen with a variant of Cantor's diagonal argument. This argument provides a total computable function that is not primitive recursive. A sketch of the proof is as follows: This argument can be applied to any class of computable (total) functions that can be enumerated in this way, as explained in the article Machine that always halts. Note however that the "partial" computable functions (those that need not be defined for all arguments) can be explicitly enumerated, for instance by enumerating Turing machine encodings. Other examples of total recursive but not primitive recursive functions are known: In the following we observe that primitive recursive functions can be of four types: In the following the mark " ' ", e.g. a', is the primitive mark meaning "the successor of", usually thought of as " +1", e.g. a +1 =def a'. The functions 16-20 and #G are of particular interest with respect to converting primitive recursive predicates to, and extracting them from, their "arithmetical" form expressed as Gödel numbers. Some additional forms of recursion also define functions that are in fact primitive recursive. Definitions in these forms may be easier to find or more natural for reading or writing. Course-of-values recursion defines primitive recursive functions. Some forms of mutual recursion also define primitive recursive functions. The functions that can be programmed in the LOOP programming language are exactly the primitive recursive functions. This gives a different characterization of the power of these functions. The main limitation of the LOOP language, compared to a Turing-complete language, is that in the LOOP language the number of times that each loop will run is specified before the loop begins to run. The primitive recursive functions are closely related to mathematical finitism, and are used in several contexts in mathematical logic where a particularly constructive system is desired. Primitive recursive arithmetic (PRA), a formal axiom system for the natural numbers and the primitive recursive functions on them, is often used for this purpose. PRA is much weaker than Peano arithmetic, which is not a finitistic system. Nevertheless, many results in number theory and in proof theory can be proved in PRA. For example, Gödel's incompleteness theorem can be formalized into PRA, giving the following theorem: Similarly, many of the syntactic results in proof theory can be proved in PRA, which implies that there are primitive recursive functions that carry out the corresponding syntactic transformations of proofs. In proof theory and set theory, there is an interest in finitistic consistency proofs, that is, consistency proofs that themselves are finitistically acceptable. Such a proof establishes that the consistency of a theory "T" implies the consistency of a theory "S" by producing a primitive recursive function that can transform any proof of an inconsistency from "S" into a proof of an inconsistency from "T". One sufficient condition for a consistency proof to be finitistic is the ability to formalize it in PRA. For example, many consistency results in set theory that are obtained by forcing can be recast as syntactic proofs that can be formalized in PRA. Recursive definitions had been used more or less formally in mathematics before, but the construction of primitive recursion is traced back to Richard Dedekind's theorem 126 of his "Was sind und was sollen die Zahlen?" (1888). This work was the first to give a proof that a certain recursive construction defines a unique function. Primitive recursive arithmetic was first proposed by Thoralf Skolem in 1923. The current terminology was coined by Rózsa Péter (1934) after Ackermann had proved in 1928 that the function which today is named after him was not primitive recursive, an event which prompted the need to rename what until then were simply called recursive functions.
https://en.wikipedia.org/wiki?curid=24829
Prime Minister of Japan The is the head of government of Japan and the commander in chief of the Japanese Armed Forces. The prime minister is appointed by the Emperor of Japan after being designated by the National Diet and must enjoy the confidence of the House of Representatives to remain in office. He is the head of the Cabinet and appoints and dismisses the other ministers of state. The literal translation of the Japanese name for the office is "Minister for the Comprehensive Administration of ("or" the Presidency over) the Cabinet". Before the adoption of the Meiji Constitution, Japan had in practice no written constitution. Originally, a Chinese-inspired legal system known as "ritsuryō" was enacted in the late Asuka period and early Nara period. It described a government based on an elaborate and rational meritocratic bureaucracy, serving, in theory, under the ultimate authority of the Emperor; although in practice, real power was often held elsewhere, such as in the hands of the Fujiwara clan, who intermarried with the Imperial Family in the Heian period, or by the ruling "shōgun". Theoretically, the last "ritsuryō" code, the Yōrō Code enacted in 752, was still in force at the time of the Meiji Restoration. Under this system, the was the head of the "Daijō-kan" (Department of State), the highest organ of Japan's pre-modern Imperial government during the Heian period and until briefly under the Meiji Constitution with the appointment of Sanjō Sanetomi in 1871. The office was replaced in 1885 with the appointment of Itō Hirobumi to the new position of Prime Minister, four years before the enactment of the Meiji Constitution, which mentions neither the Cabinet nor the position of Prime Minister explicitly. It took its current form with the adoption of the Constitution of Japan in 1947. To date, 62 people have served this position. Shinzō Abe is the current prime minister. He re-took the office on December 26, 2012. He is the first former Prime Minister to return to office since 1948, and the longest serving Prime Minister to date. The Prime Minister is designated by both houses of the Diet, before the conduct of any other business. For that purpose, each conducts a ballot under the run-off system. If the two houses choose different individuals, then a joint committee of both houses is appointed to agree on a common candidate. Ultimately, however, if the two houses do not agree within ten days, the decision of the House of Representatives is deemed to be that of the Diet. Therefore, the House of Representatives can theoretically ensure the appointment of any Prime Minister it wants. The candidate is then presented with his or her commission, and formally appointed to office by the Emperor. In practice, the Prime Minister is almost always the leader of the majority party in the House of Representatives, or the leader of the senior partner in the governing coalition. Unlike most of his counterparts in constitutional monarchies, the Prime Minister is both "de jure" and "de facto" chief executive. In most other constitutional monarchies, the monarch is nominal chief executive, while being bound by convention to act on the advice of the cabinet. In contrast, the Constitution of Japan explicitly vests executive power in the Cabinet, of which the Prime Minister is the leader. His countersignature is required for all laws and Cabinet orders. While most ministers in parliamentary democracies have some freedom of action within the bounds of cabinet collective responsibility, the Japanese Cabinet is effectively an extension of the Prime Minister's authority. Located near the Diet building, the Office of the Prime Minister of Japan is called the . The original Kantei served from 1929 until 2002, when a new building was inaugurated to serve as the current Kantei. The old Kantei was then converted into the Official Residence, or . The Kōtei lies to the southwest of the Kantei, and is linked by a walkway. The Prime Minister of Japan travels in a Lexus LS 600h L, the official transport for the head of government, or an unmodified Toyota Century escorted by a police motorcade of numerous Toyota Celsiors. For long distance air travel, Japan maintains two Boeing 747-400 aircraft mostly for the Prime Minister of Japan, the Emperor, Empress and other members of the Imperial Family, operated by the Japan Air Self-Defense Force. They have the radio callsigns Japanese Air Force One and Japanese Air Force Two when operating on official business, and Cygnus One and Cygnus Two when operating outside of official business (e.g., on training flights). The aircraft always fly together on government missions, with one serving as the primary transport and the other serving as a backup with maintenance personnel on board. The aircraft are officially referred to as . The aircraft were constructed at the Boeing factory at the same time as the U.S. Air Force One VC-25s, though the U.S. aircraft were built to the 747-200 design, while the Japanese aircraft were built to the more contemporary 747-400 design. Both Japanese aircraft were delivered in 1990. The 747s will be replaced by new Boeing 777-300ER aircraft in fiscal year 2019. Until the mid-1930s, the Prime Minister of Japan was normally granted a hereditary peerage ("kazoku") prior to leaving office if he had not already been ennobled. Titles were usually bestowed in the ranks of count, viscount or baron, depending on the relative accomplishments and status of the Prime Minister. The two highest ranks, marquess and prince, were only bestowed upon highly distinguished statesmen, and were not granted to a Prime Minister after 1928. The last Prime Minister who was a peer was Baron Kijūrō Shidehara, who served as Prime Minister from October 1945 to May 1946. The peerage was abolished when the Constitution of Japan came into effect in May 1947. Certain eminent Prime Ministers have been awarded the Order of the Chrysanthemum, typically in the degree of Grand Cordon. The highest honour in the Japanese honours system, the Collar of the Order of the Chrysanthemum, has only been conferred upon select Prime Ministers and eminent statesmen; the last such award to a living Prime Minister was to Saionji Kinmochi in 1928. More often, the Order of the Chrysanthemum has been a posthumous distinction; the Collar of the order was last awarded posthumously to former Prime Minister Eisaku Satō in June 1975. The Grand Cordon has typically been posthumously awarded; the most recent such award was to Ryutaro Hashimoto in July 2006. After relinquishing office, the Prime Minister is normally accorded the second or senior third rank in the court order of precedence, and is usually raised to the senior second rank posthumously. Certain distinguished Prime Ministers have been posthumously raised to the first rank; the last such award was to Sato Eisaku in 1975. Since the 1920s, following their tenure in office, Prime Ministers have typically been conferred with the Grand Cordon of the Order of the Paulownia Flowers (until 2003 a special higher class of the Order of the Rising Sun), depending on tenure and eminence. However, honours may be withheld due to misconduct or refusal on the part of the Prime Minister (for example, Kiichi Miyazawa).
https://en.wikipedia.org/wiki?curid=24833
Protein targeting Protein targeting or protein sorting is the biological mechanism by which proteins are transported to their appropriate destinations in the cell or outside it. Proteins can be targeted to the inner space of an organelle, different intracellular membranes, plasma membrane, or to exterior of the cell via secretion. This delivery process is carried out based on information contained in the protein itself. Correct sorting is crucial for the cell; errors can lead to diseases. Targeting signals are the pieces of information that enable the cellular transport machinery to correctly position a protein inside or outside the cell. This information is contained in the polypeptide chain or in the folded protein. The continuous stretch of amino acid residues in the chain that enables targeting are called signal peptides or targeting peptides. There are two types of targeting peptides, the presequences and the internal targeting peptides. The presequences of the targeting peptide are often found at the N-terminal extension and is composed of between 6-136 basic and hydrophobic amino acids. In case of peroxisomes the targeting sequence is on the C-terminal extension mostly. Other signals, known as signal patches, are composed of parts which are separate in the primary sequence. They become functional when folding brings them together on the protein surface. In addition, protein modifications like glycosylations can induce targeting. In 1970, Günter Blobel conducted experiments on the translocation of proteins across membranes. He was awarded the 1999 Nobel prize for his findings. He discovered that many proteins have a signal sequence, that is, a short amino acid sequence at one end that functions like a postal code for the target organelle. The translation of mRNA into protein by a ribosome takes place within the cytosol. If the synthesized proteins "belong" in a different organelle, they can be transported there in either of two ways depending on the protein: Co-translational translocation (translocation during the process of translation), and post-translational translocation (translocation after the process of translation is complete). Most proteins that are secretory, membrane-bound, or reside in the endoplasmic reticulum (ER), golgi or endosomes use the co-translational translocation pathway. This process begins with the N-terminal signal peptide of the protein being recognized by a signal recognition particle (SRP) "while the protein is still being synthesized on the ribosome". The synthesis pauses while the ribosome-protein complex is transferred to an SRP receptor on the ER in eukaryotes, and the plasma membrane in prokaryotes. There, the nascent protein is inserted into the translocon, a membrane-bound protein conducting channel composed of the Sec61 translocation complex in eukaryotes, and the homologous SecYEG complex in prokaryotes. In secretory proteins and type I transmembrane proteins, the signal sequence is immediately cleaved from the nascent polypeptide once it has been translocated into the membrane of the ER (eukaryotes) or plasma membrane (prokaryotes) by signal peptidase. The signal sequence of type II membrane proteins and some polytopic membrane proteins are not cleaved off and therefore are referred to as signal anchor sequences. Within the ER, the protein is first covered by a chaperone protein to protect it from the high concentration of other proteins in the ER, giving it time to fold correctly. Once folded, the protein is modified as needed (for example, by glycosylation), then transported to the Golgi for further processing and goes to its target organelles or is retained in the ER by various ER retention mechanisms. The amino acid chain of transmembrane proteins, which often are transmembrane receptors, passes through a membrane one or several times. They are inserted into the membrane by translocation, until the process is interrupted by a stop-transfer sequence, also called a membrane anchor or signal-anchor sequence. These complex membrane proteins are at the moment mostly understood using the same model of targeting that has been developed for secretory proteins. However, many complex multi-transmembrane proteins contain structural aspects that do not fit the model. Seven transmembrane G-protein coupled receptors (which represent about 5% of the genes in humans) mostly do not have an amino-terminal signal sequence. In contrast to secretory proteins, the first transmembrane domain acts as the first signal sequence, which targets them to the ER membrane. This also results in the translocation of the amino terminus of the protein into the ER membrane lumen. This would seem to break the rule of "co-translational" translocation which has always held for mammalian proteins targeted to the ER. This has been demonstrated with opsin with in vitro experiments. A great deal of the mechanics of transmembrane topology and folding remains to be elucidated. Even though most secretory proteins are co-translationally translocated, some are translated in the cytosol and later transported to the ER/plasma membrane by a post-translational system. In prokaryotes this requires certain cofactors such as SecA and SecB. This pathway is facilitated by Sec62 and Sec63, two membrane-bound proteins. The Sec63 complex is embedded in the ER membrane. The Sec63 complex causes hydrolysis of ATP, which allows chaperone proteins to bind to an exposed peptide chain and slide the polypeptide into the ER lumen. Once in the lumen the polypeptide chain can be folded properly. This occurs in only unfolded proteins that are in the cytosol. In addition, proteins targeted to other destinations, such as mitochondria, chloroplasts, or peroxisomes, use specialized post-translational pathways. Also, proteins targeted for the nucleus are translocated post-translation. They pass through the nuclear envelope via nuclear pores. Most mitochondrial proteins are synthesized as cytosolic precursors containing uptake peptide signals. Cytosolic chaperones deliver preproteins to channel linked receptors in the mitochondrial membrane. The preprotein with presequence targeted for the mitochondria is bound by receptors and the General Import Pore (GIP) (Receptors and GIP are collectively known as Translocase of Outer Membrane or TOM) at the outer membrane. The preprotein is translocated through TOM as hairpin loops. The preprotein is transported through the intermembrane space by small TIMs (which also acts as molecular chaperones) to the TIM23 or 22 (Translocase of Inner Membrane) at the inner membrane. Within the matrix the targeting sequence is cleaved off by mtHsp70. Three mitochondrial outer membrane receptors are known: The TOM channel (TOM40) is a cation specific high conductance channel with a molecular weight of 410 kDa and a pore diameter of 21Å. The presequence translocase23 (TIM23) is localized to the mitochondrial inner membrane and acts a pore forming protein which binds precursor proteins with its N-terminus. TIM23 acts a translocator for preproteins for the mitochondrial matrix, the inner mitochondrial membrane as well as for the intermembrane space. TIM50 is bound to TIM23 at the inner mitochondrial side and found to bind presequences. TIM44 is bound on the matrix side and found binding to mtHsp70. The presequence translocase22 (TIM22) binds preproteins exclusively bound for the inner mitochondrial membrane. Mitochondrial matrix targeting sequences are rich in positively charged amino acids and hydroxylated ones. Proteins are targeted to submitochondrial compartments by multiple signals and several pathways. Targeting to the outer membrane, intermembrane space, and inner membrane often requires another signal sequence in addition to the matrix targeting sequence. The preprotein for chloroplasts may contain a stromal import sequence or a stromal and thylakoid targeting sequence. The majority of preproteins are translocated through the Toc and Tic complexes located within the chloroplast envelope. In the stroma the stromal import sequence is cleaved off and folded as well as intra-chloroplast sorting to thylakoids continues. Proteins targeted to the envelope of chloroplasts usually lack cleavable sorting sequence. Many proteins are needed in both mitochondria and chloroplasts. In general the targeting peptide is of intermediate character to the two specific ones. The targeting peptides of these proteins have a high content of basic and hydrophobic amino acids, a low content of negatively charged amino acids. They have a lower content of alanine and a higher content of leucine and phenylalanine. The dual targeted proteins have a more hydrophobic targeting peptide than both mitochondrial and chloroplastic ones. All peroxisomal proteins are encoded by nuclear genes. To date there are two types of known Peroxisome Targeting Signals (PTS): Peroxisome targeting signal 1 (PTS1): a C-terminal tripeptide with a consensus sequence (S/A/C)-(K/R/H)-(L/A). The most common PTS1 is serine-lysine-leucine (SKL). Most peroxisomal matrix proteins possess a PTS1 type signal. Peroxisome targeting signal 2 (PTS2): a nonapeptide located near the N-terminus with a consensus sequence (R/K)-(L/V/I)-XXXXX-(H/Q)-(L/A/F) (where X can be any amino acid). There are also proteins that possess neither of these signals. Their transport may be based on a so-called "piggy-back" mechanism: such proteins associate with PTS1-possessing matrix proteins and are translocated into the peroxisomal matrix together with them. Peroxisomal protein transport is defective in the following genetic diseases: As discussed above (see protein translocation), most prokaryotic membrane-bound and secretory proteins are targeted to the plasma membrane by either a co-translation pathway that uses bacterial SRP or a post-translation pathway that requires SecA and SecB. At the plasma membrane, these two pathways deliver proteins to the SecYEG translocon for translocation. Bacteria may have a single plasma membrane (Gram-positive bacteria), or an inner membrane plus an outer membrane separated by the periplasm (Gram-negative bacteria). Besides the plasma membrane the majority of prokaryotes lack membrane-bound organelles as found in eukaryotes, but they may assemble proteins onto various types of inclusions such as gas vesicles and storage granules. In gram-negative bacteria proteins may be incorporated into the plasma membrane, the outer membrane, the periplasm or secreted into the environment. Systems for secreting proteins across the bacterial outer membrane may be quite complex and play key roles in pathogenesis. These systems may be described as type I secretion, type II secretion, etc. In most gram-positive bacteria, certain proteins are targeted for export across the plasma membrane and subsequent covalent attachment to the bacterial cell wall. A specialized enzyme, sortase, cleaves the target protein at a characteristic recognition site near the protein C-terminus, such as an LPXTG motif (where X can be any amino acid), then transfers the protein onto the cell wall. Several analogous systems are found that likewise feature a signature motif on the extracytoplasmic face, a C-terminal transmembrane domain, and cluster of basic residues on the cytosolic face at the protein's extreme C-terminus. The PEP-CTERM/exosortase system, found in many Gram-negative bacteria, seems to be related to extracellular polymeric substance production. The PGF-CTERM/archaeosortase A system in archaea is related to S-layer production. The GlyGly-CTERM/rhombosortase system, found in the Shewanella, Vibrio, and a few other genera, seems involved in the release of proteases, nucleases, and other enzymes. Minimotif Miner is a bioinformatics tool that searches protein sequence queries for a known protein targeting sequence motifs.
https://en.wikipedia.org/wiki?curid=24834
Pinochle Pinochle (), also called pinocle or penuchle, is a trick-taking, Ace-Ten card game typically for two to four players and played with a 48-card deck. It is derived from the card game bezique; players score points by trick-taking and also by forming combinations of cards into melds. It is thus considered part of a "trick-and-meld" category which also includes the game belote. Each hand is played in three phases: bidding, melds, and tricks. The standard game today is called "partnership auction pinochle." There are two possible origins in looking at pinochle's ancestry. One is that it is a cousin of binokel with both games evolving from the game of bezique. A second alternative is that pinochle actually developed from the Swiss and, later, South German, game of binocle or binokel which in turn is a descendant of bezique. The word pinochle has several different potential derivations. It may come from the French word "binocle" meaning "eyeglasses" or "binoculars". There are suggestions that it comes from "bis" (until) and "knochle" (knuckle) because originally the game ended when a player rapped their knuckles on the table. The term may also be related to the French word "binage" for the combination of cards called "binocle". This latter pronunciation of the game was adopted by German speakers. German immigrants brought the game to America in the latter quarter of the 19th century, where it was mispronounced and misspelled "pinochle." Pinochle was the favorite card game of American Jewish and Irish immigrants, while skat was the preferred game of a majority of German immigrants. Auction pinochle for three players has some similarities with the German game skat, although the bidding is more similar to that of bid whist. During World War I, the city of Syracuse, New York outlawed the playing of pinochle in a gesture of anti-German sentiment. A pinochle deck consists of two copies of each of the 9, 10, jack, queen, king, and ace cards of all four suits, for 48 cards per deck. Aces are always considered high. Pinochle follows a nonstandard card ordering. The complete ordering from highest to lowest is A, 10, K, Q, J, 9. The game can also be played using standard ranking with a simple change to scoring. Originally, the deck had to be composed by combining two poker, piquet or euchre decks and removing unneeded cards (a piquet deck does not have the 2–6, making it easier to modify, and a euchre deck is exactly half a pinochle deck), but with the game's popularity in the United States in the early 1900s, a single boxed deck with the necessary cards was marketed, and these specialized pinochle decks are now widely available in similar styles to common 52-card counterparts. Variants of pinochle can be played with five, six, eight or more players. These larger variations can combine two pinochle decks called a "double deck". The double deck can also be used when playing with four players; hand sizes, average scores and minimum bids are doubled. The game is played with a deck of 48 cards and four players; one player is the dealer. After the shuffle, the dealer will offer a cut to the player on their right, then distribute the cards. All the cards are dealt in partnership pinochle. In variations for odd numbers of players like three, a "widow's hand" (also called a "kitty", "talon", or "stock") of cards remain. Traditionally, the deal is done clockwise, dealing a packet of three or four cards at a time, starting with the player to the left (the eldest hand) and ending with the dealer. The deal rotates clockwise, so the dealer's left-hand opponent will deal next. In auction pinochle, players bid for the points they predict their hand could earn. The highest bidder earns the right to declare the trump suit. One of the players, usually the player to the left of the dealer, or the dealer themselves, is obligated to open with a first bid. The size of bids is based on the point scale and number of decks used; traditionally, points are in multiples of 10, thus a minimum opening bid might be agreed to be 100 or 250. However, many alternate scoring rules drop the unnecessary trailing zero; in that case, bids of 10 and 25, respectively, have the same values. When a player has the turn to bid, the player may either bid or pass. A popular variation for four (or three) player pinochle involves dealing a 4 card kitty (3 or 6 cards for three players), with the bid winner taking the kitty and discarding 4 (3 or 6) cards from his hand. The point value of the discards can sometimes be added to the bid winner's total trick count or not, depending on the pre-established rules. In three player games the 6 card kitty can often lead to very competitive and extravagant bidding. Each bid must be greater than the previous one, and be a multiple of 10 or 25 (if playing without trailing zeroes, the bid must be one or two greater respectively). When a player passes, they can no longer bid. The auction ends when all subsequent players in rotation have passed after the last bid. The last bid becomes the "contract". The player that made this final bid will then declare trump in the suit that is desired. In some house rules, trump cannot be declared in any suit not containing a "run", "marriage" or "dix" meld. In order for the winning bidder to win the hand, the combined total of melding and trick points must be equal to or greater than the winning bid. Thus bidding involves anticipating the points that will be accumulated from melds and from the points accumulated from winning tricks. If the combined score is lower than the bid, then the bidding team or player has been "set". This means that the total bid amount is subtracted from the total game score, often accompanied by losing the points scored in meld for that hand as well. This can result in a negative score. A related though different style of bidding is for the players to bid individually, and partners' bids are then summed. The winning bid only decides trump; both (or all) teams' bids become their contract, meaning any team can score or be set. This creates a more balanced game. In some versions of pinochle, after the bid has been taken and trump declared in a partnership game, the bid winning team exchanges cards. It may be two, three, or four cards, depending on the version of the game. The partner of the bid winner passes first. The objective of the partner is either to add to the total points in meld or to pass trick-winning cards. After receiving the cards, the bid winner examines what will create the strongest hand and then discards an equal number of cards back to their partner. Variations are for the bid winner and partner to exchange the designated number of cards simultaneously, or for no passing to occur. Melding consists of displaying specific combinations of cards to all players. Typically this is done by placing the combination of cards face up on the playing surface until all players have had the opportunity to examine them. All players meld after the bid winner shows meld first. The types of melds include "arounds", "marriages", "flushes" and "pinochles". These melds are placed under "headings" where a card which is melded under a particular heading can be used again under another heading, but cannot be melded again under the same heading. The group melds containing four of the same face cards – ace, king, queen or jack – must include one card from each of the different suits. They are scored as follows: The marriages and flush are the "sequence melds": A marriage in each suit is worth 240 or 24 total points, which is nothing more than the sum of the marriages, plus kings around and queens around. As a shortcut, this is called a "roundtable", "marriages around", "round house", or a "round robin". The pinochle and dix are the "special melds". In the most common form of the game (see variations below), any one card may be used in only one meld of each type. Thus, a queen can be used in one marriage with one king, regardless of if the player has the other king of the same suit. However, a queen can be used to score a marriage and a pinochle if the player also has the correct jack. After the melds are displayed, the points are counted and each player totals their individual meld scores. Because all of these values are multiples of ten, one can arrive at simplified values by removing the trailing zero from each point total. For instance, a pinochle has a simplified score of 4, a double Pinochle would score 30. In playing cards for tricks, there are strict rules of forced play, which limit a player's ability to strategically retain high cards. The high bidder leads the play with the first card, which can be any card in the contract winner's hand, although some rules require the first card led to be a trump card. Then there are two variations of following suit depending if you are playing post-1945 or pre-1945 rules. According to the pre-1945 rules, every player must follow the lead suit if possible. Usually every player must play a winning card against those played so far, if it is possible to do so, even when the current player expects a later player to win the hand with a better card. The only exception is if a player played a trump card when trump was not the suit led. In that case, those following that player may play any card of the lead suit, since they must follow the lead suit but are already losing to the player who played trump. Likewise, if a player cannot follow suit, but has trump, they must play trump. Again, if a player does not have any cards of the lead suit and can play a trump card higher than any other trump played so far, the player must do so, even if the player expects that a later player will beat the card. If another trump has already been played that a player cannot beat, then they can play any trump in their hand, but they still must play a trump card if they can. Only when a player has no cards in suit, and has no trump, can the player choose to play any card in their hand. Most books of post-1945 rules say that unless trump is led, there is no requirement to try to win the trick. It is only when trump is led that "heading" the trick is mandatory. In pinochle circles and tournaments the post-1945 rules are played about half of the time according to Pagat and Hoyle. If two identical cards are played, the first one outranks the second. After the first trick, the winner of each trick leads the first card for the next trick, until all the cards are played. Points are scored based on the tricks won in the hand. There are several ways to count up the points for play, but they always add up to 250 points. The last trick is always worth an additional 10 points added to any existing points in the actual trick cards. The classic counting system of pinochle is where aces are worth 11, tens are worth 10, kings are worth four, queens are worth three, jacks are worth two, and nines are worth zero. This method takes longer to count the score at the end of each hand. A simpler method is to count aces and tens for 10 points, kings and queens for five points, and jacks and nines are worth zero. An even simpler method has aces, tens, and kings worth 10 (and known as "counters"), and everything else zero ("garbage"). Since all points are multiples of ten in the third method, most players drop the redundant zero. Aces, tens, and kings won in tricks are worth one point. The meld scoring can also avoid the zero in the tenth place. Melds like 1,000 aces are thus worth 100. The terms "1,000 aces", "800 kings" and so on are often used, even though the point values are one-tenth. Two-handed pinochle is the original pinochle game, while partnership, auction, and all other variants are derived from it. It is the game most similar to the original Bèzique game, whence pinochle was derived, via the German game of Binokel. The only significant difference in its rules from Bèzique is the scoring. The original version of pinochle involves a partial deal of twelve cards to both players in packets of four, leaving a stock of 24 cards. A player can score one meld after each trick won of the first 12 tricks. Melded cards can even be used to win tricks. After each trick, players draw one card from the stock into their hand starting with the trick-winning player. For the last 12 tricks, melds are taken into each player's hand and are no longer announced by the player who wins the trick. The traditional trick-taking rules apply only for these last 12 tricks. In variations of two-handed play, no cards are initially dealt, a distinction from all other variations. Instead, the entire deck is placed face-down on the playing surface between the two players to form the widow. One player begins the hand-building process by drawing the top card of the widow. The player can either keep that card for her or his hand or reject the card. If the player chooses to hold the initial card, the player then draws a second card from the widow, then places it face-down, without looking at it, creating a discard pile. If the player rejects the first card, the card becomes the first card in the discard pile. The second card drawn from the widow must be kept, regardless of whether she or he preferred the first card. Players alternate turns in this hand-building process until all cards are chosen. With bidding, the player winning the bid declares trump, then lays all meld face-up on the table. The other player shows her or his melds as well. Meld points are tallied, and players return meld cards to their hands. Some varieties accept a "round house", kings and queens of each suit, and earn a bonus 10 points awarding a total of 250 points. Trick-taking commences and continues until all held cards have been played. One variation has no "leading" requirement for the bid winner or subsequent trick winner to lead a specific card, however the rules of "following" are still observed. When adding counters, cards from each player's discard pile are included in totals for a total of 240 counters per round, plus one counter for winning the final trick. One variation to make it more difficult for the bid-winning player, the discard pile created by drawing cards is used by the non-bidding player to score towards tricks. In Three-handed pinochle, each player plays for him or herself. The dealer deals 15 cards to each player and three cards to the kitty—a separate pile in the middle. All players review their cards and silently determine their bids. The player to the dealer's left initiates the bidding process. If the player has a meld, he or she is required to open the bidding; otherwise, they may pass or bid. If he or she passes, the obligation to bid passes to the next player, if meld is held. Once a player passes, he or she is out of the auction. Bidding begins at 20, or as little as 19 if dealer, and increases in multiples of 1. The highest bidder wins the auction and turns up the three-card kitty for all to see. The three widow cards are placed in the bid winner's hand. The bid winner then declares trump and lays down meld. The other two players also lay meld face-up for count. After the appropriate points have been tallied the bid winner must set aside any three cards that have not been melded. This will reduce the bid winner's hand to 15 cards. For all three players, meld is now returned to each respective player's hand, and the round is played. During the round, a player must take at least one trick to "save one's meld", even if the trick contains no points; otherwise, no meld points will be counted for that player during that round. After all tricks are taken, counters are tallied for each player. The three discards by the highest bidder count toward their counter score for the hand, so there is always a total of 25 points for the trick score among the three players. If the highest bidder fails to make their contract by adding meld points and trick points from the play, then their score is negative the amount of the bid for that hand. The meld count is cancelled. The game is won when one player reaches 100 points. It is possible for two or all three players to go over 100 on the same hand. There are three methods of resolving ties: Any time a player accidentally misplays during the play portion of the hand, it is called a renege. There are various forms of misplay: If the bidder reneges, they automatically takes a double set and the amount of the bid is subtracted from their score. The two opposing players get to count their meld points and the remainder of the hand is thrown in. If either of the two nonbidders misplay, the bidder automatically makes their bid. The bidder gets to score the amount of their bid and meld, the player that misplayed loses all meld and takes a single set, and the third player scores only their meld. Card-Fault Misdeal If at any point during melding or play it is determined that a non-standard deck is being used, any player may declare a card-fault misdeal. This results in the nullification of the entire hand including all meld and points obtained. Similar to three-handed pinochle, cutthroat is a simple modification. The dealer deals the entire deck out (16 cards to each player), in packets of four. The player to the left of the dealer begins the bidding once meld has been silently determined by all players. Play continues normally in terms of scoring and trick taking. The only way to win in cutthroat pinochle, however, is to "bid and out", or to have taken the bid and surpassed the predetermined winning score. It is then possible for multiple players to go over the winning score, yet if none has taken a bid and met the resulting contract, a win has not happened and play continues. It is also possible for a person to lose with the high score if they do not take a winning bid. Four-handed pinochle, or "partnership pinochle" is played with two teams of two players each. Partners are seated opposite from each other. Each player is dealt 12 cards. The opening bid is typically 150, but can be a higher agreed on value. All four players may bid. Both the bidder and his partner have their score count towards making the contract. High bidder names trump. There typically is no kitty. With a kitty, the four cards are distributed, one to each player, by the bid winner. Each hand must meld separately. As in the three-handed version, the first player is forced to bid when holding meld. Play is often to 1000 but can increase to 1500 during partnership. Games with five hands or more typically modify partnership pinochle in a variety of ways. They are generally played with 1 1/2 or doubled decks, with extra dix added or withheld to make an even deal. With an odd number of players, the bidder asks for a desired card in the trump suit, with the first matching player being partner for that hand. Everyone else plays against the team. In larger groups, one or more players can sit out each hand allowing the remaining players to follow the appropriate rules for the respective number of players. Check pinochle is a gambling variant of three-hand. It is the same as to 1000, except that players keep track of "checks". If playing for $1 stakes, each check gained means that the other two players owe a dollar. The following events cause a gain or loss of checks. Today "double-deck" pinochle is a popular form of the game, exclusively played by the National Pinochle Association, the American Pinochle Association, the Cambridge Pinochle Association, and in the "World Series of Pinochle". Double-deck pinochle is played with two pinochle decks, without the nines. This makes for an 80 card deck. Play is similar to regular pinochle, except 20 cards are dealt to each person and minimum bid is increased to 500 points. In some variations, bids are made in increments of 10 or more points until 600 is reached, then by 50 points. This version often features "meld bidding", a bid made to let a partner know what is in the bidder's hand. The only communication during bidding should be a numerical number or "pass", any other way of communicating is called "talking across the table" and is forbidden. There are occasionally different meld values for a run and a pinochle; a run being 250 or 150 points, and a pinochle being 150 or 40 points. All other aspects of the game generally remain the same. Technical Misdeal If a player is dealt 13 or more non-counters and no aces, the player may declare a technical misdeal. This must declared before he or she plays the first trick. A technical misdeal nullifies all points melded for all players. The hand is then re-dealt by the original dealer of that hand. In triple-deck pinochle six play in two partnerships of three each; each player has an opponent at their right and left. Three pinochle decks with no nines are mixed together, making a pack of 120 cards. Each player is dealt 20 cards, and the rules of double deck pinochle apply, except that the minimum bid is 75, and the last trick is worth 3 points. most of the extra melds made possible by the triple pack do not count extra. i.e. if a player should hold twenty aces, five of each suit, the value would be that of double aces and triple aces combined. Internet pinochle is almost always "double deck" except for a few applications for some smart phones. Today the Internet is host to many live professional cash tournaments, although many are still cautious about playing online because of potential cheating. Note that this use of the term "racehorse" is inconsistent with the commonly understood meaning of the term when applied to Pinochle. As summarized by Dave LeVasseur: "Racehorse means that, after the winning bidder has named trump, that player's partner passes cards across the table" Played much the same as "double deck" but to six hands, the point values are inflated. Two teams are formed, 20 cards are then dealt to each player and 4 cards are dealt to the blind. Bidding commences with the person immediately to the left of the dealer automatically bidding 500. The winner of the bid includes the blind into their hand, calls trump and melds. Note: all runs, double, triple, and quadruple, marriages must be in trump The game continues with the standard rules of play. When the play is over each team adds up their points in the count with kings, 10s, and aces worth ten points, while queens and jacks are worth zero. If a team count plus meld does not equal their bid, they "go set". By going set the amount of the bid is subtracted from the team's score and their count is discarded. The other team retains both their meld and their count provided they took at least 10 points in the count. Two full decks are dealt between eight players, forming four teams. Team members are spaced so that they are not able to see any other hands. The game is usually played to a score of 5,000 or higher. Other than this, the four player rules apply, and any variations may also be used. There is an increased possibility that when one team declares trump another team may have an equal number of trump also, which may lead to an interesting game. An optional scoring rule rewards 1,000 points for a quadruple pinochle—four jacks of diamonds and four queens of spades in a meld. Alternate end games One variation on winning allows a team or individual to win instantly from any score by taking all the tricks in a hand. To win in this fashion, the winning player or team must play very skillfully to prevent opposing players from taking even one lowly (even zero-point) trick. This victory is known as "pinochling". A player or team can play for this victory even if they are not the highest bidder. "pinochling" does not require a bidder to make their bid. They also can play for this victory even if their bid cannot be made with the maximum number of trick points available plus their meld. However, the highest bidding player or team can prevent other players from attempting this if they elect to "throw in" the hand before the first card is played. When playing "bid-out" rules, a team can win without bidding if their score reaches (and remains above) the agreed upon game-ending score while their opponents fail to make their bid three times. This is known as a "slide-out".
https://en.wikipedia.org/wiki?curid=24837
Peptidoglycan Peptidoglycan or murein is a polymer consisting of sugars and amino acids that forms a mesh-like layer outside the plasma membrane of most bacteria, forming the cell wall. The sugar component consists of alternating residues of β-(1,4) linked "N"-acetylglucosamine (NAG) and "N"-acetylmuramic acid (NAM). Attached to the "N"-acetylmuramic acid is a peptide chain of three to five amino acids. The peptide chain can be cross-linked to the peptide chain of another strand forming the 3D mesh-like layer. Peptidoglycan serves a structural role in the bacterial cell wall, giving structural strength, as well as counteracting the osmotic pressure of the cytoplasm. Peptidoglycan is also involved in binary fission during bacterial cell reproduction. The peptidoglycan layer is substantially thicker in Gram-positive bacteria (20 to 80 nanometers) than in Gram-negative bacteria (7 to 8 nanometers). Peptidoglycan forms around 90% of the dry weight of Gram-positive bacteria but only 10% of Gram-negative strains. Thus, presence of high levels of peptidoglycan is the primary determinant of the characterisation of bacteria as Gram-positive. In Gram-positive strains, it is important in attachment roles and serotyping purposes. For both Gram-positive and Gram-negative bacteria, particles of approximately 2 nm can pass through the peptidoglycan. The peptidoglycan layer in the bacterial cell wall is a crystal lattice structure formed from linear chains of two alternating amino sugars, namely "N"-acetylglucosamine (GlcNAc or NAGA) and "N"-acetylmuramic acid (MurNAc or NAMA). The alternating sugars are connected by a β-(1,4)-glycosidic bond. Each MurNAc is attached to a short (4- to 5-residue) amino acid chain, containing L-alanine, D-glutamic acid, "meso"-diaminopimelic acid, and D-alanine in the case of "Escherichia coli" (a Gram-negative bacterium) or L-alanine, D-glutamine, L-lysine, and D-alanine with a 5-glycine interbridge between tetrapeptides in the case of "Staphylococcus aureus" (a Gram-positive bacterium). Peptidoglycan is one of the most important sources of D-amino acids in nature. Cross-linking between amino acids in different linear amino sugar chains occurs with the help of the enzyme DD-transpeptidase and results in a 3-dimensional structure that is strong and rigid. The specific amino acid sequence and molecular structure vary with the bacterial species. The peptidoglycan monomers are synthesized in the cytosol and are then attached to a membrane carrier bactoprenol. Bactoprenol transports peptidoglycan monomers across the cell membrane where they are inserted into the existing peptidoglycan. In the first step of peptidoglycan synthesis, glutamine, which is an amino acid, donates an amino group to a sugar, fructose 6-phosphate. This turns fructose 6-phosphate into glucosamine-6-phosphate. In step two, an acetyl group is transferred from acetyl CoA to the amino group on the glucosamine-6-phosphate creating "N"-acetyl-glucosamine-6-phosphate. In step three of the synthesis process, the "N"-acetyl-glucosamine-6-phosphate is isomerized, which will change "N"-acetyl-glucosamine-6-phosphate to "N"-acetyl-glucosamine-1-phosphate. In step 4, the "N"-acetyl-glucosamine-1-phosphate, which is now a monophosphate, attacks UTP. Uridine triphosphate, which is a pyrimidine nucleotide, has the ability to act as an energy source. In this particular reaction, after the monophosphate has attacked the UTP, an inorganic pyrophosphate is given off and is replaced by the monophosphate, creating UDP-N-acetylglucosamine (2,4). (When UDP is used as an energy source, it gives off an inorganic phosphate.) This initial stage, is used to create the precursor for the NAG in peptidoglycan. In step 5, some of the UDP-N-acetylglucosamine (UDP-GlcNAc) is converted to UDP-MurNAc (UDP-N-acetylmuramic acid) by the addition of a lactyl group to the glucosamine. Also in this reaction, the C3 hydroxyl group will remove a phosphate from the alpha carbon of phosphoenolpyruvate. This creates what is called an enol derivative that will be reduced to a “lactyl moiety” by NADPH in step six. In step 7, the UDP–MurNAc is converted to UDP-MurNAc pentapeptide by the addition of five amino acids, usually including the dipeptide D-alanyl-D-alanine. Each of these reactions requires the energy source ATP. This is all referred to as Stage one. Stage two occurs in the cytoplasmic membrane. It is in the membrane where a lipid carrier called bactoprenol carries peptidoglycan precursors through the cell membrane. Bactoprenol will attack the UDP-MurNAc penta, creating a PP-MurNac penta, which is now a lipid. UDP-GlcNAc is then transported to MurNAc, creating Lipid-PP-MurNAc penta-GlcNAc, a disaccharide, also a precursor to peptidoglycan. How this molecule is transported through the membrane is still not understood. However, once it is there, it is added to the growing glycan chain. The next reaction is known as tranglycosylation. In the reaction, the hydroxyl group of the GlcNAc will attach to the MurNAc in the glycan, which will displace the lipid-PP from the glycan chain. The enzyme responsible for this is transglycosylase. Some antibacterial drugs such as penicillin interfere with the production of peptidoglycan by binding to bacterial enzymes known as penicillin-binding proteins or DD-transpeptidases. Penicillin-binding proteins form the bonds between oligopeptide crosslinks in peptidoglycan. For a bacterial cell to reproduce through binary fission, more than a million peptidoglycan subunits (NAM-NAG+oligopeptide) must be attached to existing subunits. Mutations in genes coding for transpeptidases that lead to reduced interactions with an antibiotic are a significant source of emerging antibiotic resistance. Lysozyme, which is found in tears and constitutes part of the body's innate immune system exerts its antibacterial effect by breaking the β-(1,4)-glycosidic bonds in peptidoglycan (see above). Some archaea have a similar layer of pseudopeptidoglycan (also known as pseudomurein), in which the sugar residues are β-(1,3) linked "N"-acetylglucosamine and "N"-acetyltalosaminuronic acid. This makes the cell walls of such archaea insensitive to lysozyme.
https://en.wikipedia.org/wiki?curid=24838
Pope Sixtus IV Pope Sixtus IV (21 July 1414 – 12 August 1484), born Francesco della Rovere, was head of the Catholic Church and ruler of the Papal States from 9 August 1471 to his death. His accomplishments as pope included the construction of the Sistine Chapel and the creation of the Vatican Archives. A patron of the arts, he brought together the group of artists who ushered the Early Renaissance into Rome with the first masterpieces of the city's new artistic age. Sixtus aided the Spanish Inquisition though he fought to prevent abuses therein, and he annulled the decrees of the Council of Constance. He was noted for his nepotism and was personally involved in the infamous Pazzi conspiracy. Francesco was born to a family of modest means from Liguria, Italy, the son of Leonardo della Rovere and Luchina Monleoni. He was born in Celle Ligure, a town near Savona. As a young man, Della Rovere joined the Franciscan Order, an unlikely choice for a political career, and his intellectual qualities were revealed while he was studying philosophy and theology at the University of Pavia. He went on to lecture at Padua and many other Italian universities. In 1464, Della Rovere was elected Minister General of the Franciscan order at the age of 50. In 1467, he was appointed Cardinal by Pope Paul II with the titular church being the Basilica of San Pietro in Vincoli. Before his papal election, Cardinal della Rovere was renowned for his unworldliness and had written learned treatises, including "On the Blood of Christ" and "On the Power of God". His reputation for piety was one of the deciding factors that prompted the College of Cardinals to elect him Pope upon the unexpected death of Paul II at the age of fifty-four. Upon being elected Pope, Della Rovere adopted the name Sixtus, which had not been used since the 5th century. One of his first acts was to declare a renewed crusade against the Ottoman Turks in Smyrna. However, after the conquest of Smyrna, the fleet disbanded. Some fruitless attempts were made towards unification with the Greek Church. For the remainder of his pontificate, Sixtus turned to temporal issues and dynastic considerations. Sixtus IV sought to strengthen his position by surrounding himself with relatives and friends. In the fresco by Melozzo da Forlì, he is accompanied by his Della Rovere and Riario nephews, not all of whom were made cardinals; the protonotary apostolic Pietro Riario (on his right), the future Pope Julius II/ Giuliano Della Rovere standing before him; and Girolamo Riario and Giovanni della Rovere, behind the kneeling Platina, author of the first humanist history of the popes. His nephew Pietro Riario also benefited from his nepotism. Pietro became one of the richest men in Rome and was entrusted with Pope Sixtus' foreign policy. However, Pietro died prematurely in 1474, and his role passed to Giuliano Della Rovere. The secular fortunes of the Della Rovere family began when Sixtus invested his nephew Giovanni with the lordship of Senigallia and arranged his marriage to the daughter of Federico III da Montefeltro, duke of Urbino; from that union came a line of Della Rovere dukes of Urbino that lasted until the line expired, in 1631. Six of the thirty-four cardinals that he created were his nephews. In his territorial aggrandizement of the Papal States, his niece's son Cardinal Raffaele Riario, for whom the Palazzo della Cancelleria was constructed, was suspected of colluding in the failed Pazzi conspiracy of 1478 to assassinate both Lorenzo de' Medici and his brother Giuliano and replace them in Florence with Sixtus IV's other nephew, Girolamo Riario. Francesco Salviati, Archbishop of Pisa and a main organizer of the plot, was hanged on the walls of the Florentine Palazzo della Signoria. Sixtus IV replied with an interdict and two years of war with Florence. According to the later published chronicle of the Italian historian Stefano Infessura, "Diary of the City of Rome", Sixtus was a "lover of boys and sodomites", awarding benefices and bishoprics in return for sexual favours and nominating a number of young men as cardinals, some of whom were celebrated for their good looks. However, Infessura had partisan allegiances to the Colonna and so is not considered to be always reliable or impartial. The English churchman and Protestant polemicist John Bale, writing a century later, attributed to Sixtus "the authorisation to practice sodomy during periods of warm weather" to the "Cardinal of Santa Lucia". Although such accusations are easily dismissed as anti-Catholic propaganda, they still prompted the noted historian of the Catholic Church, Ludwig von Pastor, to issue a firm rebuttal. Sixtus continued a dispute with King Louis XI of France, who upheld the Pragmatic Sanction of Bourges (1438), which held that papal decrees needed royal assent before they could be promulgated in France. That was a cornerstone of the privileges claimed for the Gallican Church and could never be shifted as long as Louis XI manoeuvred to replace King Ferdinand I of Naples with a French prince. Louis was thus in conflict with the papacy, and Sixtus could not permit it. On 1 November 1478, Sixtus published the papal bull "Exigit Sincerae Devotionis Affectus" through which the Spanish Inquisition was established in the Kingdom of Castile. Sixtus consented under political pressure from Ferdinand of Aragon, who threatened to withhold military support from his kingdom of Sicily. Nevertheless, Sixtus IV quarrelled over protocol and prerogatives of jurisdiction; he was unhappy with the excesses of the Inquisition and condemned the most flagrant abuses in 1482. As a temporal prince who constructed stout fortresses in the Papal States, he encouraged the Venetians to attack Ferrara, which he wished to obtain for another nephew. Ercole I d'Este, Duke of Ferrara, was allied with the Sforzas of Milan, the Medicis of Florence along with the King of Naples, normally a hereditary ally and champion of the papacy. The angered Italian princes allied to force Sixtus IV to make peace to his great annoyance. For refusing to desist from the very hostilities that he himself had instigated and for being a dangerous rival to Della Rovere dynastic ambitions in the Marche, Sixtus placed Venice under interdict in 1483. He also lined the coffers of the state by unscrupulously selling high offices and privileges. In ecclesiastical affairs, Sixtus promoted the dogma of the Immaculate Conception, which had been confirmed at the Council of Basle in 1439, and he designated 8 December as its feastday. In 1476, he issued the apostolic constitution "Cum Praeexcelsa", establishing a Mass and Office for the feast. He formally annulled the decrees of the Council of Constance in 1478. The two papal bulls issued by Pope Nicholas V, "Dum Diversas" of 1452 and "Romanus Pontifex" of 1455, had effectively given the Portuguese the rights to acquire slaves along the African Coast by force or trade. Those concessions were confirmed by Sixtus in his own bull, "Aeterni regis", of 21 June 1481. Arguably the "ideology of conquest" expounded in those texts became the means by which commerce and conversion were facilitated. In November 1476, Isabel and Fernando ordered an investigation into rights of conquest in the Canary Islands, and in the spring of 1478, they sent Juan Rejon with sixty soldiers and thirty cavalry to the Grand Canary, where the natives retreated inland. Sixtus's earlier threats to excommunicate all captains or pirates who enslaved Christians in the bull "Regimini Gregis" of 1476 could have been intended to emphasise the need to convert the natives of the Canary Islands and Guinea and establish a clear difference in status between those who had converted and those who resisted. The ecclesiastical penalties were directed towards those who were enslaving the recent converts. As a civic patron in Rome, even the anti-papal chronicler Stefano Infessura agreed that Sixtus should be admired. The dedicatory inscription in the fresco by Melozzo da Forlì in the Vatican Palace records: "You gave your city temples, streets, squares, fortifications, bridges and restored the Acqua Vergine as far as the Trevi..." In addition to restoring the aqueduct that provided Rome an alternative to the river water, which had made the city famously unhealthy, he restored or rebuilt over 30 of Rome's dilapidated churches such as San Vitale (1475) and Santa Maria del Popolo, and he added seven new ones. The Sistine Chapel was sponsored by Sixtus IV, as was the "Ponte Sisto", the Sistine Bridge (the first new bridge across the Tiber since Antiquity) and the building of "Via Sistina" (later named "Borgo Sant'Angelo"), a road leading from Castel Sant'Angelo to Saint Peter. All of that was done to facilitate the integration of the Vatican Hill and Borgo with the heart of Old Rome. That was part of a broader scheme of urbanization carried out under Sixtus IV, who swept the long-established markets from the Campidoglio in 1477 and decreed in a bull of 1480 the widening of streets and the first post-Roman paving, the removal of porticoes and other post-classical impediments to free public passage. At the beginning of his papacy, in 1471, Sixtus had donated several historically important Roman sculptures that founded a papal collection of art, which would eventually develop into the collections of the Capitoline Museums. He also refounded, enriched and enlarged the Vatican Library. He had Regiomontanus attempt the first sanctioned reorganisation of the Julian calendar and increased the size and prestige of the papal chapel choir, bringing singers and some prominent composers (Gaspar van Weerbeke, Marbrianus de Orto and Bertrandus Vaqueras) to Rome from the north. In addition to being a patron of the arts, Sixtus was a patron of the sciences. Before he became pope, he had spent time at the very liberal and cosmopolitan University of Padua, which maintained considerable independence from the Church and had a very international character. As Pope, he issued a papal bull allowing local bishops to give the bodies of executed criminals and unidentified corpses to physicians and artists for dissection. It was that access to corpses which allowed the anatomist Vesalius, along with Titian's pupil Jan Stephen van Calcar, to complete the revolutionary medical/anatomical text "De humani corporis fabrica". The Pope created 34 cardinals in eight consistories held during his reign, among them three nephews, one grandnephew and one other relative, thus continuing the practice of nepotism that he and his successors would engage in during this period. Sixtus IV named seven new saints with the most notable being Bonaventure (1482); he also beatified one person: John Buoni (1483). In 1477, Sixtus IV issued a papal bull, authorizing the creation of Uppsala University - the first university in Sweden and in the whole of Scandinavia. The choice of this location for the university derived from the fact that the archbishopric of Uppsala had been one of the most important sees in Sweden proper since Christianity first spread to this region in the ninth century, as well as Uppsala being long-standing hub for regional trade. Uppsala's bull, which granted the university its corporate rights, established a number of provisions. Among the most important of these was that the university was officially given the same freedoms and privileges as the University of Bologna. This included the right to establish the four traditional faculties of theology, law (Canon Law and Roman law), medicine, and philosophy, and to award the bachelor's, master's, licentiate, and doctoral degrees. The archbishop of Uppsala was also named as the university's Chancellor, and was charged with maintaining the rights and privileges of the university and its members. This act of Sixtus IV had a profound long-term effect on the society and culture of Sweden, an effect which continues up to the present. Sixtus IV became ill on 8 August 1484; this illness worsened on 10 August while the pope was attending an event in Rome. He felt unwell that evening and was forced to cancel a meeting he was to hold with his cardinals the following morning. The Pope grew weaker during the night of 11 August and he was unable to sleep. Pope Sixtus IV died the following evening - 12 August. The envoy of the Medici family summed up Sixtus' reign in the announcement to his master 'Today at 5 o'clock His Holiness Sixtus IV departed this life-may God forgive him!' Pope Sixtus's tomb was destroyed in the Sack of Rome in 1527. Today, his remains, along with the remains of his nephew Pope Julius II (Giuliano della Rovere), are interred in St. Peter's Basilica, in the floor in front of the monument to Pope Clement X. A marble tombstone marks the site. His bronze funerary monument, now in the basement Treasury of St. Peter's Basilica, made like a giant casket of goldsmith's work, is by Antonio Pollaiuolo. The top of the casket is a lifelike depiction of the Pope lying in state. Around the sides are bas-relief panels depicting allegorical female figures representing Grammar, Rhetoric, Arithmetic, Geometry, Music, Painting, Astronomy, Philosophy and Theology—the classical liberal arts, with the addition of painting and theology. Each figure incorporates the oak tree ("rovere" in Italian), symbol of Sixtus IV. The overall program of the panels, their beauty, complex symbolism, classical references and their relative arrangement are compelling and comprehensive illustrations of the Renaissance worldview. None of them actually states how he died. Sixtus created an unusually large number of cardinals during his pontificate (23) who were drawn from the roster of the princely houses of Italy, France and Spain, thus ensuring that many of his policies continued after his death: Pope Sixtus is portrayed by Arthur Grosser in the short film , a prequel to the video game Assassin's Creed II. Pope Sixtus is portrayed by James Faulkner in the historical fantasy "Da Vinci's Demons" as having an identical twin, Alessandro. Shortly after the true Pope Sixtus, Francesco, was elected on conclave, Alessandro usurped the Holy See and had his brother locked up in Castel Sant'Angelo. The series implies that many of the more unsavoury parts of Sixtus' reign were really the work of his evil twin, who was out to gain power for himself. Pope Sixtus is portrayed by Raul Bova in the second season, and John Lynch in the third season of the TV series "".
https://en.wikipedia.org/wiki?curid=24845
Panama Canal The Panama Canal () is an artificial waterway in Panama that connects the Atlantic Ocean with the Pacific Ocean. The canal cuts across the Isthmus of Panama and is a conduit for maritime trade. Canal locks are at each end to lift ships up to Gatun Lake, an artificial lake created to reduce the amount of excavation work required for the canal, 26 m (85 ft) above sea level, and then lower the ships at the other end. The original locks are 32.5 m (110 ft) wide. A third, wider lane of locks was constructed between September 2007 and May 2016. The expanded canal began commercial operation on June 26, 2016. The new locks allow transit of larger, neo-Panamax ships, capable of handling more cargo. France began work on the canal in 1881, but stopped because of engineering problems and a high worker mortality rate. The United States took over the project in 1904 and opened the canal on August 15, 1914. One of the largest and most difficult engineering projects ever undertaken, the Panama Canal shortcut greatly reduced the time for ships to travel between the Atlantic and Pacific oceans, enabling them to avoid the lengthy, hazardous Cape Horn route around the southernmost tip of South America via the Drake Passage or Strait of Magellan and the even less popular route through the Arctic Archipelago and the Bering Strait. Colombia, France, and later the United States controlled the territory surrounding the canal during construction. The US continued to control the canal and surrounding Panama Canal Zone until the 1977 Torrijos–Carter Treaties provided for handover to Panama. After a period of joint American–Panamanian control, in 1999, the canal was taken over by the Panamanian government. It is now managed and operated by the government-owned Panama Canal Authority. Annual traffic has risen from about 1,000 ships in 1914, when the canal opened, to 14,702 vessels in 2008, for a total of 333.7 million Panama Canal/Universal Measurement System (PC/UMS) tons. By 2012, more than 815,000 vessels had passed through the canal. It takes 11.38 hours to pass through the Panama Canal. The American Society of Civil Engineers has ranked the Panama Canal one of the seven wonders of the modern world. The earliest record related to a canal across the Isthmus of Panama was in 1534, when Charles V, Holy Roman Emperor and King of Spain, ordered a survey for a route through the Americas in order to ease the voyage for ships traveling between Spain and Peru. The Spanish were seeking to gain a military advantage over the Portuguese. In 1668, the English physician and philosopher Sir Thomas Browne speculated in his encyclopaedic work, "Pseudodoxia Epidemica," that "some Isthmus have been eat through by the Sea, and others cut by the spade: And if policy would permit, that of Panama in America were most worthy the attempt: it being but few miles over, and would open a shorter cut unto the East Indies and China". In 1788, American Thomas Jefferson, then Minister to France, suggested that the Spanish should build the canal, since they controlled the colonies where it would be built. He said that this would be a less treacherous route for ships than going around the southern tip of South America, and that tropical ocean currents would naturally widen the canal after construction. During an expedition from 1788 to 1793, Alessandro Malaspina outlined plans for construction of a canal. Given the strategic location of Panama, and the potential of its narrow isthmus separating two great oceans, other trade links in the area were attempted over the years. The ill-fated Darien scheme was launched by the Kingdom of Scotland in 1698 to set up an overland trade route. Generally inhospitable conditions thwarted the effort, and it was abandoned in April 1700. Numerous canals were built in other countries in the late 18th and early 19th centuries. The success of the Erie Canal through central New York in the United States in the 1820s, and the collapse of the Spanish Empire in Latin America resulted in growing American interest in building an inter-oceanic canal. Beginning in 1826, US officials began negotiations with Gran Colombia (present-day Colombia, Venezuela, Ecuador, and Panama), hoping to gain a concession to build a canal. Jealous of their newly gained independence and fearing domination by the more powerful United States, president Simón Bolívar and New Granada officials declined American offers. The new nation was politically unstable, and Panama rebelled several times during the 19th century. Great Britain attempted to develop a canal in 1843. According to the "New York Daily Tribune", August 24, 1843, Barings of London and the Republic of New Granada entered into a contract for construction of a canal across the Isthmus of Darien (Isthmus of Panama). They referred to it as the Atlantic and Pacific Canal, and it was a wholly British endeavor. Projected for completion in five years, the plan was never carried out. At nearly the same time, other ideas were floated, including a canal (and/or a railroad) across Mexico's Isthmus of Tehuantepec. That did not develop, either. In 1846, the Mallarino–Bidlack Treaty, negotiated between the US and New Granada, granted the United States transit rights and the right to intervene militarily in the isthmus. In 1848, the discovery of gold in California, on the West Coast of the United States, generated renewed interest in a canal crossing between the Atlantic and Pacific oceans. William H. Aspinwall, who had won the federal subsidy to build and operate the Pacific mail steamships at around the same time, benefited from the gold discovery. Aspinwall's route included steamship legs from New York City to Panama, and from Panama to California, with an overland portage through Panama. This route with an overland leg in Panama was soon frequently traveled, as it provided one of the fastest connections between San Francisco, California, and the East Coast cities, about 40 days' transit in total. Nearly all the gold that was shipped out of California went by the fast Panama route. Several new and larger paddle steamers were soon plying this new route, including private steamship lines owned by American entrepreneur Cornelius Vanderbilt that made use of an overland route through Nicaragua. In 1850 the United States began construction of the Panama Railroad (now called the Panama Railway) to cross the isthmus; it opened in 1855. This overland link became a vital piece of Western Hemisphere infrastructure, greatly facilitating trade. The later canal route was constructed parallel to it, as it had helped clear dense forests. An all-water route between the oceans was still the goal. In 1855 William Kennish, a Manx-born engineer working for the United States government, surveyed the isthmus and issued a report on a route for a proposed Panama Canal. His report was published as a book entitled "The Practicability and Importance of a Ship Canal to Connect the Atlantic and Pacific Oceans". In 1877, Armand Reclus, an officer with the French Navy, and Lucien Napoléon Bonaparte Wyse, both engineers, surveyed the route and published a French proposal for a canal. The French had achieved success in building the Suez Canal in the Mideast. While it was a lengthy project, they were encouraged to plan for a canal to cross the Panamanian isthmus. The first attempt to construct a canal through what was then Colombia's province of Panama began on January 1, 1881. The project was inspired by the diplomat Ferdinand de Lesseps, who was able to raise considerable funds in France as a result of the huge profits generated by his successful construction of the Suez Canal. Although the Panama Canal needed to be only 40 percent as long as the Suez Canal, it was much more of an engineering challenge due to the combination of tropical rain forests, debilitating climate, the need for canal locks, and the lack of any ancient route to follow. De Lesseps wanted a sea-level canal (like the Suez), but he visited the site only a few times, during the dry season which lasts only four months of the year. His men were totally unprepared for the rainy season, during which the Chagres River, where the canal started, became a raging torrent, rising up to 10 m (35 ft). The dense jungle was alive with venomous snakes, insects, and spiders, but the worst challenges were yellow fever, malaria, and other tropical diseases, which killed thousands of workers; by 1884, the death rate was over 200 per month. Public health measures were ineffective because the role of the mosquito as a disease vector was then unknown. Conditions were downplayed in France to avoid recruitment problems, but the high mortality rate made it difficult to maintain an experienced workforce. Workers had to continually widen the main cut through the mountain at Culebra and reduce the angles of the slopes to minimize landslides into the canal. Steam shovels were used in the construction of the canal, purchased from Bay City Industrial Works, a business owned by William L. Clements in Bay City, Michigan. Bucket chain excavators manufactured by both Alphonse Couvreux and Wehyer & Richemond and Buette were also used. Other mechanical and electrical equipment was limited in capabilities, and steel equipment rusted rapidly in the rainy climate. In France, de Lesseps kept the investment and supply of workers flowing long after it was obvious that the targets were not being met, but eventually the money ran out. The French effort went bankrupt in 1889 after reportedly spending US$287,000,000; an estimated 22,000 men died from disease and accidents, and the savings of 800,000 investors were lost. Work was suspended on May 15, and in the ensuing scandal, known as the Panama affair, some of those deemed responsible were prosecuted, including Gustave Eiffel. De Lesseps and his son Charles were found guilty of misappropriation of funds and sentenced to five years' imprisonment. This sentence was later overturned, and the father, at age 88, was never imprisoned. In 1894, a second French company, the Compagnie Nouvelle du Canal de Panama, was created to take over the project. A minimal workforce of a few thousand people was employed primarily to comply with the terms of the Colombian Panama Canal concession, to run the Panama Railroad, and to maintain the existing excavation and equipment in saleable condition. The company sought a buyer for these assets, with an asking price of US$109,000,000. In the meantime, they continued with enough activity to maintain their franchise. Phillipe Bunau-Varilla, the French manager of the New Panama Canal Company, eventually managed to persuade de Lesseps that a lock-and-lake canal was more realistic than a sea-level canal. At this time, the President and the Senate of the United States were interested in establishing a canal across the isthmus, with some favoring a canal across Nicaragua and others advocating the purchase of the French interests in Panama. Bunau-Varilla, who was seeking American involvement, asked for $100 million, but accepted $40 million in the face of the Nicaraguan option. In June 1902, the US Senate voted in favor of the Spooner Act, to pursue the Panamanian option, provided the necessary rights could be obtained. On January 22, 1903, the Hay–Herrán Treaty was signed by United States Secretary of State John M. Hay and Colombian Chargé Dr. Tomás Herrán. For $10 million and an annual payment, it would have granted the United States a renewable lease in perpetuity from Colombia on the land proposed for the canal. The treaty was ratified by the US Senate on March 14, 1903, but the Senate of Colombia did not ratify it. Bunau-Varilla told President Theodore Roosevelt and Hay of a possible revolt by Panamanian rebels who aimed to separate from Colombia, and hoped that the United States would support the rebels with US troops and money. Roosevelt changed tactics, based in part on the Mallarino–Bidlack Treaty of 1846, and actively supported the separation of Panama from Colombia. Shortly after recognizing Panama, he signed a treaty with the new Panamanian government under terms similar to the Hay–Herrán Treaty. On November 2, 1903, US warships blocked sea lanes against possible Colombian troop movements en route to put down the Panama rebellion. Panama declared independence on November 3, 1903. The United States quickly recognized the new nation. On November 6, 1903, Philippe Bunau-Varilla, as Panama's ambassador to the United States, signed the Hay–Bunau-Varilla Treaty, granting rights to the United States to build and indefinitely administer the Panama Canal Zone and its defenses. This is sometimes misinterpreted as the "99-year lease" because of misleading wording included in article 22 of the agreement. Almost immediately, the treaty was condemned by many Panamanians as an infringement on their country's new national sovereignty. This would later become a contentious diplomatic issue among Colombia, Panama, and the United States. President Roosevelt famously stated, "I took the Isthmus, started the canal and then left Congress not to debate the canal, but to debate me." Several parties in the United States called this an act of war on Colombia: The "New York Times" described the support given by the United States to Bunau-Varilla as an "act of sordid conquest." The "New York Evening Post" called it a "vulgar and mercenary venture." The US maneuvers are often cited as the classic example of US gunboat diplomacy in Latin America, and the best illustration of what Roosevelt meant by the old African adage, "Speak softly and carry a big stick [and] you will go far." After the revolution in 1903, the Republic of Panama became a US protectorate until 1939. In 1904, the United States purchased the French equipment and excavations, including the Panama Railroad, for US$40 million, of which $30 million related to excavations completed, primarily in the Gaillard Cut (then called the Culebra Cut), valued at about $1.00 per cubic yard. The United States also paid the new country of Panama $10 million and a $250,000 payment each following year. In 1921, Colombia and the United States entered into the Thomson–Urrutia Treaty, in which the United States agreed to pay Colombia $25 million: $5 million upon ratification, and four-$5 million annual payments, and grant Colombia special privileges in the Canal Zone. In return, Colombia recognized Panama as an independent nation. The US formally took control of the canal property on May 4, 1904, inheriting from the French a depleted workforce and a vast jumble of buildings, infrastructure, and equipment, much of it in poor condition. A US government commission, the Isthmian Canal Commission (ICC), was established to oversee construction; it was given control of the Panama Canal Zone, over which the United States exercised sovereignty. The commission reported directly to Secretary of War William Howard Taft and was directed to avoid the inefficiency and corruption that had plagued the French 15 years earlier. On May 6, 1904, President Theodore Roosevelt appointed John Findley Wallace, formerly chief engineer and finally general manager of the Illinois Central Railroad, as chief engineer of the Panama Canal Project. Overwhelmed by the disease-plagued country and forced to use often dilapidated French infrastructure and equipment, as well as being frustrated by the overly bureaucratic ICC, Wallace resigned abruptly in June 1905. He was succeeded by John Frank Stevens, a self-educated engineer who had built the Great Northern Railroad. Stevens was not a member of the ICC; he increasingly viewed its bureaucracy as a serious hindrance, bypassing the commission and sending requests and demands directly to the Roosevelt administration in Washington, DC. One of Stevens' first achievements in Panama was in building and rebuilding the housing, cafeterias, hotels, water systems, repair shops, warehouses, and other infrastructure needed by the thousands of incoming workers. Stevens began the recruitment effort to entice thousands of workers from the United States and other areas to come to the Canal Zone to work, and tried to provide accommodation in which the incoming workers could work and live in reasonable safety and comfort. He also re-established and enlarged the railway, which was to prove crucial in transporting millions of tons of soil from the cut through the mountains to the dam across the Chagres River. Colonel William C. Gorgas had been appointed chief sanitation officer of the canal construction project in 1904. Gorgas implemented a range of measures to minimize the spread of deadly diseases, particularly yellow fever and malaria, which had recently been shown to be mosquito-borne following the work of Dr. Carlos Finlay and Dr. Walter Reed. Investment was made in extensive sanitation projects, including city water systems, fumigation of buildings, spraying of insect-breeding areas with oil and larvicide, installation of mosquito netting and window screens, and elimination of stagnant water. Despite opposition from the commission (one member said his ideas were barmy), Gorgas persisted, and when Stevens arrived, he threw his weight behind the project. After two years of extensive work, the mosquito-spread diseases were nearly eliminated. Even after all that effort, about 5,600 workers died of disease and accidents during the US construction phase of the canal. In 1905, a US engineering panel was commissioned to review the canal design, which had not been finalized. The panel recommended to President Roosevelt a sea-level canal, as had been attempted by the French. But in 1906 Stevens, who had seen the Chagres in full flood, was summoned to Washington; he declared a sea-level approach to be "an entirely untenable proposition". He argued in favor of a canal using a lock system to raise and lower ships from a large reservoir above sea level. This would create both the largest dam (Gatun Dam) and the largest man-made lake (Gatun Lake) in the world at that time. The water to refill the locks would be taken from Gatun Lake by opening and closing enormous gates and valves and letting gravity propel the water from the lake. Gatun Lake would connect to the Pacific through the mountains at the Gaillard (Culebra) Cut. Stevens successfully convinced Roosevelt of the necessity and feasibility of this alternative scheme. The construction of a canal with locks required the excavation of more than of material over and above the excavated by the French. As quickly as possible, the Americans replaced or upgraded the old, unusable French equipment with new construction equipment that was designed for a much larger and faster scale of work. 102 large, railroad-mounted steam shovels were purchased, 77 from Bucyrus-Erie, and 25 from the Marion Power Shovel Company. These were joined by enormous steam-powered cranes, giant hydraulic rock crushers, concrete mixers, dredges, and pneumatic power drills, nearly all of which were manufactured by new, extensive machine-building technology developed and built in the United States. The railroad also had to be comprehensively upgraded with heavy-duty, double-tracked rails over most of the line to accommodate new rolling stock. In many places, the new Gatun Lake flooded over the original rail line, and a new line had to be constructed above Gatun Lake's waterline. In 1907, Stevens resigned as chief engineer. His replacement, appointed by President Theodore Roosevelt, was US Army Major George Washington Goethals of the US Army Corps of Engineers. Soon to be promoted to lieutenant colonel and later to general, he was a strong, West Point–trained leader and civil engineer with experience of canals (unlike Stevens). Goethals directed the work in Panama to a successful conclusion in 1914, two years ahead of the target date of June 10, 1916. Goethals divided the engineering and excavation work into three divisions: Atlantic, Central, and Pacific. The Atlantic Division, under Major William L. Sibert, was responsible for construction of the massive breakwater at the entrance to Limon Bay, the Gatun locks, and their 3½ mi (5.6 km) approach channel, and the immense Gatun Dam. The Pacific Division, under Sydney B. Williamson (the only civilian member of this high-level team), was similarly responsible for the Pacific 3 mi (4.8 km) breakwater in Panama Bay, the approach channel to the locks, and the Miraflores and Pedro Miguel locks and their associated dams and reservoirs. The Central Division, under Major David du Bose Gaillard of the United States Army Corps of Engineers, was assigned one of the most difficult parts: excavating the Culebra Cut through the continental divide to connect Gatun Lake to the Pacific Panama Canal locks. On October 10, 1913, President Woodrow Wilson sent a signal from the White House by telegraph which triggered the explosion that destroyed the Gamboa Dike. This flooded the Culebra Cut, thereby joining the Atlantic and Pacific oceans via the Panama Canal. (a floating crane built by Lobnitz & Company and launched in 1887) was the first self-propelled vessel to transit the canal from ocean to ocean. This vessel crossed the canal from the Atlantic in stages during construction, finally reaching the Pacific on January 7, 1914. (a cargo and passenger ship built by Maryland Steel, and launched in 1902 as SS "Tremont") on August 3, 1914 was the first ship to transit the canal from ocean to ocean. The construction of the canal was completed in 1914, 401 years after Panama was first crossed by Vasco Núñez de Balboa. The United States spent almost $500,000,000 (roughly equivalent to $9,169,650,000 now) to finish the project. This was by far the largest American engineering project to date. The canal was formally opened on August 15, 1914, with the passage of the cargo ship . The opening of the Panama Canal in 1914 caused a severe drop in traffic along Chilean ports due to shifts in maritime trade routes. Throughout this time, Ernest "Red" Hallen was hired by the Isthmian Canal Commission to document the progress of the work. By the 1930s, water supply became an issue for the canal, prompting construction of the Madden Dam across the Chagres River above Gatun Lake. Completed in 1935, the dam created Madden Lake (later Alajeula Lake), which provides additional water storage for the canal. In 1939, construction began on a further major improvement: a new set of locks large enough to carry the larger warships that the United States was building at the time and planned to continue building. The work proceeded for several years, and significant excavation was carried out on the new approach channels, but the project was cancelled after World War II. After World War II, US control of the canal and the Canal Zone surrounding it became contentious; relations between Panama and the United States became increasingly tense. Many Panamanians felt that the Canal Zone rightfully belonged to Panama; student protests were met by the fencing-in of the zone and an increased military presence there. Demands for the United States to hand over the canal to Panama increased after the Suez Crisis in 1956, when the United States used financial and diplomatic pressure to force France and the UK to abandon their attempt to retake control of the Suez Canal, previously nationalized by the Nasser regime in Egypt. Panamanian unrest culminated in riots on Martyr's Day, January 9, 1964, when about 20 Panamanians and 3–5 US soldiers were killed. A decade later, in 1974, negotiations toward a settlement began and resulted in the Torrijos–Carter Treaties. On September 7, 1977, the treaty was signed by President of the United States Jimmy Carter and Omar Torrijos, "de facto" leader of Panama. This mobilized the process of granting the Panamanians free control of the canal so long as Panama signed a treaty guaranteeing the permanent neutrality of the canal. The treaty led to full Panamanian control effective at noon on December 31, 1999, and the Panama Canal Authority (ACP) assumed command of the waterway. The Panama Canal remains one of the chief revenue sources for Panama. Before this handover, the government of Panama held an international bid to negotiate a 25-year contract for operation of the container shipping ports located at the canal's Atlantic and Pacific outlets. The contract was not affiliated with the ACP or Panama Canal operations and was won by the firm Hutchison Whampoa, a Hong Kong–based shipping interest owned by Li Ka-shing. While globally the Atlantic Ocean is east of the isthmus and the Pacific is west, the general direction of the canal passage from the Atlantic to the Pacific is from northwest to southeast, because of the shape of the isthmus at the point the canal occupies. The Bridge of the Americas () at the Pacific side is about a third of a degree east of the Colón end on the Atlantic side. Still, in formal nautical communications, the simplified directions "southbound" and "northbound" are used. The canal consists of artificial lakes, several improved and artificial channels, and three sets of locks. An additional artificial lake, Alajuela Lake (known during the American era as Madden Lake), acts as a reservoir for the canal. The layout of the canal as seen by a ship passing from the Atlantic to the Pacific is: Thus, the total length of the canal is 50 miles. Artificially created in 1913 by damming the Chagres River, Gatun Lake is an essential part of the Panama Canal, providing the millions of litres of water necessary to operate the Panama Canal locks each time a ship passes through. At the time it was formed, Gatun Lake was the largest human-made lake in the world. The impassable rainforest around the lake has been the best defense of the Panama Canal. Today these areas remain practically unscathed by human interference and are one of the few accessible areas where various native Central American animal and plant species can be observed undisturbed in their natural habitat. The largest island on Gatun Lake is Barro Colorado Island. It was established for scientific study when the lake was formed, and is operated by the Smithsonian Institution. Many important scientific and biological discoveries of the tropical animal and plant kingdom originated here. Gatun Lake covers about , a vast tropical ecological zone and part of the Atlantic Forest Corridor. Ecotourism on the lake has become an industry for Panamanians. Gatun Lake also provides drinking water for Panama City and Colón. Fishing is one of the primary recreational pursuits on Gatun Lake. Non-native peacock bass were introduced by accident to Gatun Lake around 1967 by a local businessman, and have since flourished to become the dominant angling game fish in Gatun Lake. Locally called Sargento and believed to be the species "Cichla pleiozona", these peacock bass originate from the Amazon, Rio Negro, and Orinoco river basins, where they are considered a premier game fish. The size of the locks determines the maximum size ship that can pass through. Because of the importance of the canal to international trade, many ships are built to the maximum size allowed. These are known as Panamax vessels. A Panamax cargo ship typically has a deadweight tonnage (DWT) of 65,000–80,000 tonnes, but its actual cargo is restricted to about 52,500 tonnes because of the draft restrictions within the canal. The longest ship ever to transit the canal was the "San Juan Prospector" (now "Marcona Prospector"), an ore-bulk-oil carrier that is long with a beam of . Initially the locks at Gatun were designed to be wide. In 1908, the United States Navy requested that an increased width of at least to allow the passage of US naval ships. Eventually a compromise was made and the locks were built wide. Each lock is long, with the walls ranging in thickness from at the base to at the top. The central wall between the parallel locks at Gatun is thick and over high. The steel lock gates measure an average of thick, wide, and high. Panama Canal pilots were initially unprepared to handle the significant flight deck overhang of aircraft carriers. knocked over all the adjacent concrete lamp posts while passing through the Gatun Locks for the first time in 1928. It is the size of the locks, specifically the Pedro Miguel Locks, along with the height of the Bridge of the Americas at Balboa, that determine the Panamax metric and limit the size of ships that may use the canal. The 2006 third set of locks project has created larger locks, allowing bigger ships to transit through deeper and wider channels. The allowed dimensions of ships using these locks increased by 25 percent in length, 51 percent in beam, and 26 percent in draft, as defined by New Panamax metrics. As with a toll road, vessels transiting the canal must pay tolls. Tolls for the canal are set by the Panama Canal Authority and are based on vessel type, size, and the type of cargo. For container ships, the toll is assessed on the ship's capacity expressed in twenty-foot equivalent units (TEUs), one TEU being the size of a standard intermodal shipping container. Effective April 1, 2016, this toll went from US$74 per loaded container to $60 per TEU capacity plus $30 per loaded container for a potential $90 per TEU when the ship is full. A Panamax container ship may carry up to . The toll is calculated differently for passenger ships and for container ships carrying no cargo ("in ballast"). , the ballast rate is US$60, down from US$65.60 per TEU. Passenger vessels in excess of 30,000 tons (PC/UMS), known popularly as cruise ships, pay a rate based on the number of berths, that is, the number of passengers that can be accommodated in permanent beds. The per-berth charge since April 1, 2016 is $111 for unoccupied berths and $138 for occupied berths in the Panamax locks. Started in 2007, this fee has greatly increased the tolls for such ships. Passenger vessels of less than 30,000 tons or less than 33 tons per passenger are charged according to the same per-ton schedule as are freighters. Almost all major cruise ships have more than 33 tons per passenger; the rule of thumb for cruise line comfort is generally given as a minimum of 40 tons per passenger. Most other types of vessel pay a toll per PC/UMS net ton, in which one "ton" is actually a volume of . (The calculation of tonnage for commercial vessels is quite complex.) , this toll is US$5.25 per ton for the first 10,000 tons, US$5.14 per ton for the next 10,000 tons, and US$5.06 per ton thereafter. As with container ships, reduced tolls are charged for freight ships "in ballast", $4.19, $4.12, $4.05 respectively. On 1 April 2016, a more complicated toll system was introduced, having the neopanamax locks at a higher rate in some cases, natural gas transport as a new separate category and other changes. As of October 1, 2017, there are modified tolls and categories of tolls in effect. Small (less than 125 ft) vessels up to 583 PC/UMS net tons when carrying passengers or cargo, or up to 735 PC/UMS net tons when in ballast, or up to 1,048 fully loaded displacement tons, are assessed minimum tolls based upon their length overall, according to the following table (as of 29 April 2015): Morgan Adams of Los Angeles, California, holds the distinction of paying the first toll received by the United States Government for the use of the Panama Canal by a pleasure boat. His boat "Lasata" passed through the Zone on August 14, 1914. The crossing occurred during a 6,000-mile sea voyage from Jacksonville, Florida, to Los Angeles in 1914. The most expensive regular toll for canal passage to date was charged on April 14, 2010 to the cruise ship "Norwegian Pearl," which paid US$375,600. The average toll is around US$54,000. The highest fee for priority passage charged through the Transit Slot Auction System was US$220,300, paid on August 24, 2006, by the Panamax tanker "Erikoussa", bypassing a 90-ship queue waiting for the end of maintenance work on the Gatun Locks, and thus avoiding a seven-day delay. The normal fee would have been just US$13,430. The lowest toll ever paid was 36 cents (), by American Richard Halliburton who swam the Panama Canal in 1928. Opponents to the 1977 Torrijos-Carter Treaties feared that efficiency and maintenance would suffer following the US withdrawal from the Panama Canal Zone; however, this has been proven not to be the case. Capitalizing on practices developed during the American administration, canal operations are improving under Panamanian control. Canal Waters Time (CWT), the average time it takes a vessel to navigate the canal, including waiting time, is a key measure of efficiency; according to the ACP, since 2000, it has ranged between 20 and 30 hours. The accident rate has also not changed appreciably in the past decade, varying between 10 and 30 accidents each year from about 14,000 total annual transits. An official accident is one in which a formal investigation is requested and conducted. Increasing volumes of imports from Asia, which previously landed on US West Coast ports, are now passing through the canal to the American East Coast. The total number of ocean-going transits increased from 11,725 in 2003 to 13,233 in 2007, falling to 12,855 in 2009. (The canal's fiscal year runs from October through September.) This has been coupled with a steady rise in average ship size and in the numbers of Panamax vessels passing through the canal, so that the total tonnage carried rose from 227.9 million PC/UMS tons in fiscal year 1999 to a then record high of 312.9 million tons in 2007, and falling to 299.1 million tons in 2009. Tonnage for fiscal 2013, 2014 and 2015 was 320.6, 326.8 and 340.8 million PC/UMS tons carried on 13,660, 13,481 and 13,874 transits respectively. In the first decade after the transfer to Panamanian control, the Panama Canal Authority (ACP) invested nearly US$1 billion in widening and modernizing the canal, with the aim of increasing capacity by 20 percent. The ACP cites a number of major improvements, including the widening and straightening of the Gaillard Cut to reduce restrictions on passing vessels, the deepening of the navigational channel in Gatun Lake to reduce draft restrictions and improve water supply, and the deepening of the Atlantic and Pacific entrances to the canal. This is supported by new equipment, such as a new drill barge and suction dredger, and an increase of the tug boat fleet by 20 percent. In addition, improvements have been made to the canal's operating machinery, including an increased and improved tug locomotive fleet, the replacement of more than of locomotive track, and new lock machinery controls. Improvements have been made to the traffic management system to allow more efficient control over ships in the canal. In December 2010, record-breaking rains caused a 17-hour closure of the canal; this was the first closure since the United States invasion of Panama in 1989. The rains also caused an access road to the Centenario Bridge to collapse. The canal is currently handling more vessel traffic than had ever been envisioned by its builders. In 1934 it was estimated that the maximum capacity of the canal would be around 80 million tons per year; as noted above, canal traffic in 2015 reached 340.8 million tons of shipping. To improve capacity, a number of improvements have been made to the current canal system. These improvements aim to maximize the possible use of current locking system: These improvements enlarged the capacity from 300 million PCUMS (2008) to 340 PCUMS (2012). These improvements were started before the new locks project, and are complementary to it. Despite having enjoyed a privileged position for many years, the canal is increasingly facing competition from other quarters. Because canal tolls have risen as ships have become larger, some critics have suggested that the Suez Canal is now a viable alternative for cargo en route from Asia to the US East Coast. The Panama Canal, however, continues to serve more than 144 of the world's trade routes and the majority of canal traffic comes from the "all-water route" from Asia to the US East and Gulf Coasts. On June 15, 2013, Nicaragua awarded the Hong Kong-based HKND Group a 50-year concession to develop a canal through the country. The increasing rate of melting of ice in the Arctic Ocean has led to speculation that the Northwest Passage or Arctic Bridge may become viable for commercial shipping at some point in the future. This route would save on the route from Asia to Europe compared with the Panama Canal, possibly leading to a diversion of some traffic to that route. However, such a route is beset by unresolved territorial issues and would still hold significant problems owing to ice. Gatun Lake is filled with rainwater, and the lake accumulates excess water during wet months. The water is lost to the oceans at a rate of per downward lock cycle. Since a ship will have to go upward to Gatun Lake first and then descend, a single passing will cost double the amount; but the same waterflow cycle can be used for another ship passing in the opposite direction. The ship's submerged volume is not relevant to this amount of water. During the dry season, when there is less rainfall, there is also a shortfall of water in Gatun Lake. As a signatory to the United Nations Global Compact and member of the World Business Council for Sustainable Development, the ACP has developed an environmentally and socially sustainable program for expansion, which protects the aquatic and terrestrial resources of the canal watershed. The expansion guarantees the availability and quality of water resources by using water-saving basins at each new lock. These water-saving basins diminish water loss and preserve freshwater resources along the waterway by reusing water from the basins into the locks. Each lock chamber has three water-saving basins, which reuse 60 percent of the water in each transit. There are a total of nine basins for each of the two lock complexes, and a total of 18 basins for the entire project. The mean sea level at the Pacific side is about higher than that of the Atlantic side due to differences in ocean conditions such as water density and weather. As demand is rising for efficient global shipping of goods, the canal is positioned to be a significant feature of world shipping for the foreseeable future. However, changes in shipping patterns —particularly the increasing numbers of larger-than-Panamax ships— necessitated changes to the canal for it to retain a significant market share. In 2006 it was anticipated that by 2011, 37 percent of the world's container ships would be too large for the present canal, and hence a failure to expand would result in a significant loss of market share. The maximum sustainable capacity of the original canal, given some relatively minor improvement work, was estimated at 340 million PC/UMS tons per year; it was anticipated that this capacity would be reached between 2009 and 2012. Close to 50 percent of transiting vessels were already using the full width of the locks. An enlargement scheme similar to the 1939 Third Lock Scheme, to allow for a greater number of transits and the ability to handle larger ships, had been under consideration for some time, was approved by the government of Panama, The cost was estimated at , and the expansion allowed to double the canal's capacity, allowing more traffic and the passage of longer and wider Post-Panamax ships. The proposal to expand the canal was approved in a national referendum by about 80 percent on October 22, 2006. The canal expansion was built between 2007 and 2016, though completion was originally expected by the end of 2014. The expansion plan had two new flights of locks built parallel to, and operated in addition to, the old locks: one east of the existing Gatun locks, and one southwest of the Miraflores locks, each supported by approach channels. Each flight ascends from sea level directly to the level of Gatun Lake; the existing two-stage ascent at Miraflores and Pedro Miguel locks was not replicated. The new lock chambers feature sliding gates, doubled for safety, and are long, wide, and deep. This allows the transit of vessels with a beam of up to , an overall length of up to and a draft of up to , equivalent to a container ship carrying around 12,000 containers, each in length (TEU). The new locks are supported by new approach channels, including a channel at Miraflores from the locks to the Gaillard Cut, skirting Miraflores Lake. Each of these channels are wide, which will require post-Panamax vessels to navigate the channels in one direction at a time. The Gaillard Cut and the channel through Gatun Lake were widened to at least on the straight portions and at least on the bends. The maximum level of Gatun Lake was raised from to . Each flight of locks is accompanied by nine water reutilization basins (three per lock chamber), each basin being about wide, long and deep. These gravity-fed basins allow 60 percent of the water used in each transit to be reused; the new locks consequently use 7 percent less water per transit than each of the existing lock lanes. The deepening of Gatun Lake and the raising of its maximum water level also provide capacity for significantly more water storage. These measures are intended to allow the expanded canal to operate without constructing new reservoirs. The estimated cost of the project is . The project was designed to allow for an anticipated growth in traffic from 280 million PC/UMS tons in 2005 to nearly 510 million PC/UMS tons in 2025. The expanded canal will have a maximum sustainable capacity of about 600 million PC/UMS tons per year. Tolls will continue to be calculated based on vessel tonnage, and in some cases depend on the locks used. An article in the February 2007 issue of "Popular Mechanics" magazine described the plans for the canal expansion, focusing on the engineering aspects of the expansion project. There is also a follow-up article in the February 2010 issue of "Popular Mechanics". On September 3, 2007, thousands of Panamanians stood across from Paraíso Hill in Panama to witness a huge initial explosion and launch of the Expansion Program. The first phase of the project was the dry excavations of the wide trench connecting the Gaillard Cut with the Pacific coast, removing 47 million cubic meters of earth and rock. By June 2012, a 30 m reinforced concrete monolith had been completed, the first of 46 such monoliths which will line the new Pacific-side lock walls. By early July 2012, however, it was announced that the canal expansion project had fallen six months behind schedule, leading expectations for the expansion to open in April 2015 rather than October 2014, as originally planned. By September 2014, the new gates were projected to be open for transit at the "beginning of 2016." It was announced in July 2009 that the Belgian dredging company Jan De Nul, together with a consortium of contractors consisting of the Spanish Sacyr Vallehermoso, the Italian Impregilo, and the Panamanian company Grupo Cusa, had been awarded the contract to build the six new locks for US$3.1 billion, which was one billion less than the next highest competing bid due to having a concrete budget 71 percent smaller than that of the next bidder and allotted roughly 25 percent less for steel to reinforce that concrete. The contract resulted in $100 million in dredging works over the next few years for the Belgian company and a great deal of work for its construction division. The design of the locks is a carbon copy of the Berendrecht Lock, which is 68 m wide and 500 m long, making it the second largest lock in the world after the Kieldrecht lock in the port of Antwerp, Belgium. Completed in 1989 by the Port of Antwerp, which De Nul helped build, the company still has engineers and specialists who were part of that project. In January 2014, a contract dispute threatened the progress of the project. There was a delay of less than two months however, with work by the consortium members reaching goals by June 2014. In June 2015, flooding of the new locks began: first on the Atlantic side, then on the Pacific; by then, the canal's re-inauguration was slated for April 2016. On March 23, 2016, the expansion inauguration was set for June 26, 2016. The new locks opened for commercial traffic on 26 June 2016, and the first ship to cross the canal using the third set of locks was a modern New Panamax vessel, the Chinese-owned container ship "Cosco Shipping Panama". The original locks, now over 100 years old, allow engineers greater access for maintenance, and are projected to continue operating indefinitely. The total cost is unknown since the expansion's contractors are seeking at least an addition from the canal authority due to excess expenses. On July 7, 2014, Wang Jing, chairman of the HK Nicaragua Canal Development Investment Co. Ltd. (HKND Group) advised that a route for Nicaragua's proposed canal had been approved. The construction work was projected by HKND to begin in 2014 and take 5 years, although there has been little progress whilst a series of environmental impact assessments are being made. The Nicaraguan parliament has approved plans for the canal through Nicaragua, and according to the deal, the company will be responsible for operating and maintaining the canal for a 50-year period. The government of Nicaragua hopes this will boost the economy; the opposition is concerned with its environmental impact. According to the independent impact assessment by British firm ERM, some 30,000 local residents will be displaced by the canal, although opposition leaders and Amnesty International claim the figure will be in the hundreds of thousands. Supporters and the environmental impact study claim there will be net environmental benefits, but critics argue that nearly of delicate ecosystems will be destroyed by the time construction is completed. China is investigating a proposal to construct a railway between Colombia's Pacific and Caribbean coasts. Individuals, companies, and governments have explored the possibility of constructing deep water ports and rail links connecting coasts as a "dry canal" in Guatemala, Costa Rica, and El Salvador/Honduras. However, plans to construct these sea-rail-sea links have yet to materialize. During the last one hundred years, the Panama Canal Authority has appointed a few "Panama Canal Honorary Pilots". The most recent of these were Commodore Ronald Warwick, a former Master of the Cunard Liners "Queen Elizabeth 2" and "RMS Queen Mary 2", who has traversed the Canal more than 50 times, and Captain Raffaele Minotauro, an Unlimited Master Senior Grade, of the former Italian governmental navigation company known as the "Italian Line". This service can be requested from the Panama canal authority at any time who will consider each request.
https://en.wikipedia.org/wiki?curid=24849
Political fiction Political fiction employs narrative to comment on political events, systems and theories. Works of political fiction, such as political novels, often "directly criticize an existing society or present an alternative, even fantastic, reality". The political novel overlaps with the social novel, proletarian novel, and social science fiction. Plato's "Republic", a Socratic dialogue written around 380 BC, has been one of the world's most influential works of philosophy and political theory, both intellectually and historically. The "Republic" is concerned with justice (), the order and character of the just city-state, and the just man. Other influential politically-themed works include Thomas More's "Utopia" (1516), Jonathan Swift's "Gulliver's Travels" (1726), Voltaire's "Candide" (1759), and Harriet Beecher Stowe's "Uncle Tom's Cabin" (1852). Political fiction frequently employs satire, often in the utopian and dystopian genres. This includes totalitarian dystopias of the early 20th century such as Jack London's "The Iron Heel", Sinclair Lewis' "It Can't Happen Here", and George Orwell's "Nineteen Eighty-Four". The Greek playwright Aristophanes' plays are known for their political and social satire, particularly in his criticism of the powerful Athenian general, Cleon, in plays such as "The Knights". Aristophanes is also notable for the persecution he underwent. Aristophanes' plays turned upon images of filth and disease. His bawdy style was adopted by Greek dramatist-comedian Menander, whose early play, "Drunkenness", contains an attack on the politician, Callimedon. Jonathan Swift's "A Modest Proposal" (1729) is an 18th-century Juvenalian satirical essay in which he suggests that the impoverished Irish might ease their economic troubles by selling their children as food for rich gentlemen and ladies. The satirical hyperbole mocks heartless attitudes towards the poor, as well as British policy toward the Irish in general. George Orwell's "Animal Farm" (1945) is an allegorical and dystopian novella which satirises the Russian Revolution of 1917 and the Soviet Union's Stalinist era. Orwell, a democratic socialist, was a critic of Joseph Stalin and was hostile to Moscow-directed Stalinism—an attitude that had been shaped by his experiences during the Spanish Civil War. The Soviet Union, he believed, had become a brutal dictatorship, built upon a cult of personality and enforced by a reign of terror. Orwell described his "Animal Farm" as "a satirical tale against Stalin", and in his essay "Why I Write" (1946) he wrote that "Animal Farm" was the first book in which he tried, with full consciousness of what he was doing, "to fuse political purpose and artistic purpose into one whole." Orwell's most famous work, however, is "Nineteen Eighty-Four" (published in 1949), many of whose terms and concepts, such as "Big Brother", "doublethink", "thoughtcrime", "Newspeak", "Room 101", "telescreen", "2 + 2 = 5", and "memory hole", have entered into common use. "Nineteen Eighty-Four" popularised the adjective "Orwellian", which describes official deception, secret surveillance, and manipulation of recorded history by a totalitarian or authoritarian state. The poet Jan Kochanowski's play, "The Dismissal of the Greek Envoys" (1578), the first tragedy written in the Polish language, recounts an incident leading up to the Trojan War. Its theme of the responsibilities of statesmanship resonates to the present day. The political comedy, "The Return of the Deputy" (1790), by Julian Ursyn Niemcewicz—Polish poet, playwright, statesman, and comrade-in-arms of Tadeusz Kościuszko—was written in about two weeks' time while Niemcewicz was serving as a deputy to the historic Four-Year Sejm of 1788–92. The comedy's premiere in January 1791 was an enormous success, sparking widespread debate, royal communiques, and diplomatic correspondence. As Niemcewicz had hoped, it set the stage for passage of Poland's epochal Constitution of 3 May 1791, which is regarded as Europe's first, and the world's second, modern written national constitution, after the United States Constitution implemented in 1789. The comedy pits proponents against opponents of political reforms: of abolishing the destabilizing free election of Poland's kings; of abolishing the legislatively destructive "liberum veto"; of granting greater rights to peasants and townspeople; of curbing the privileges of the mostly self-interested noble class; and of promoting a more active Polish role in international affairs, in the interest of stopping the depredations of Poland's neighbors, Russia, Prussia, and Austria (who will in 1795 complete the dismemberment of the Polish-Lithuanian Commonwealth). Romantic interest is provided by a rivalry between a reformer and a conservative for a young lady's hand—which is won by the proponent of reforms. An early example of the political novel is "The Betrothed" (1827) by Alessandro Manzoni, an Italian historical novel. Set in northern Italy in 1628, during the oppressive years of direct Spanish rule, it has been seen sometimes as a veiled attack on the Austrian Empire, which controlled Italy at the time the novel was written. It has been called the most famous and widely read novel in the Italian language. In the 1840s British politician Benjamin Disraeli wrote a trilogy of novels with political themes. With "Coningsby; or, The New Generation" (1844), Disraeli, in historian Robert Blake's view, "infused the novel genre with political sensibility, espousing the belief that England's future as a world power depended not on the complacent old guard, but on youthful, idealistic politicians." "Coningsby" was followed by "Sybil; or, The Two Nations" (1845), another political novel, which was less idealistic and more clear-eyed than "Coningsby"; the "two nations" of its subtitle referred to the huge economic and social gap between the privileged few and the deprived working classes. The last of Disraeli's political-novel trilogy, "Tancred; or, The New Crusade" (1847), promoted the Church of England's role in reviving Britain's flagging spirituality. Ivan Turgenev wrote "Fathers and Sons" (1862) as a response to the growing cultural schism that he saw between Russia's liberals of the 1830s and 1840s, and the growing Russian nihilist movement among their sons. Both the nihilists and the 1830s liberals sought Western-based social change in Russia. Additionally, these two modes of thought were contrasted with the Slavophiles, who believed that Russia's path lay in its traditional spirituality. Turgenev's novel was responsible for popularizing the use of the term "nihilism", which became widely used after the novel was published. The Polish writer Bolesław Prus' novel, "Pharaoh" (1895), is set in the Egypt of 1087–85 BCE as that country experiences internal stresses and external threats that will culminate in the fall of its Twentieth Dynasty and New Kingdom. The young protagonist Ramses learns that those who would challenge the powers that be are vulnerable to co-option, seduction, subornation, defamation, intimidation, and assassination. Perhaps the chief lesson, belatedly absorbed by Ramses as pharaoh, is the importance, to power, of knowledge. Prus' vision of the fall of an ancient civilization derives some of its power from the author's intimate awareness of the final demise of the Polish-Lithuanian Commonwealth in 1795, a century before he completed "Pharaoh". This is a political awareness that Prus shared with his 10-years-junior novelist compatriot, Joseph Conrad, who was an admirer of Prus' writings. "Pharaoh" has been translated into 20 languages and adapted as a 1966 Polish feature film. It is also known to have been Joseph Stalin's favourite book. Joseph Conrad wrote several novels with political themes: "Nostromo (1904)", "The Secret Agent" (1907), and "Under Western Eyes" (1911). "Nostromo" (1904) is set amid political upheaval in the fictitious South American country of Costaguana, where a trusted Italian-descended longshoreman, Giovanni Battista Fidanza—the novel's eponymous "Nostromo" (Italian for "our man")—is instructed by English-descended silver-mine owner Charles Gould to take Gould's silver abroad so that it will not fall into the hands of revolutionaries. The role of politics is paramount in "The Secret Agent", as the main character, Verloc, works for a quasi-political organisation. The plot to destroy Greenwich Observatory is in itself anarchistic. Vladimir asserts that the bombing "must be purely destructive" and that the anarchists who will be implicated as the architects of the explosion "should make it clear that [they] are perfectly determined to make a clean sweep of the whole social creation." However, the political form of anarchism is ultimately controlled in the novel: the only supposed politically motivated act is orchestrated by a secret government agency. Conrad's third political novel, "Under Western Eyes", is connected to Russian history. Its first audience read it against the backdrop of the failed Revolution of 1905 and in the shadow of the movements and impulses that would take shape as the revolutions of 1917. Conrad's earlier novella, "Heart of Darkness" (1899), also had political implications, in its depiction of European colonial depredations in Africa, which Conrad witnessed during his employ in the Belgian Congo. John Steinbeck's novel "The Grapes of Wrath" (1939) is a passionate depiction of the plight of the poor. However, many of Steinbeck's contemporaries attacked his social and political views. Bryan Cordyack writes: "Steinbeck was attacked as a propagandist and a socialist from both the left and the right of the political spectrum. The most fervent of these attacks came from the Associated Farmers of California; they were displeased with the book's depiction of California farmers' attitudes and conduct toward the migrants. They denounced the book as a 'pack of lies' and labeled it 'communist propaganda'". Some accused Steinbeck of exaggerating camp conditions to make a political point. Steinbeck had visited the camps well before publication of the novel and argued that their inhumane nature destroyed the settlers' spirit. "The Quiet American" (1955) by English novelist Graham Greene questions the foundations of growing American involvement in Vietnam in the 1950s. The novel has received much attention due to its prediction of the outcome of the Vietnam War and subsequent American foreign policy since the 1950s. Graham Greene portrays a U.S. official named Pyle as so blinded by American exceptionalism that he cannot see the calamities he brings upon the Vietnamese. The book uses Greene's experiences as a war correspondent for "The Times" and "Le Figaro" in French Indochina in 1951–54. "The Gay Place" (1961) is a set of politically-themed novellas with interlocking plots and characters by American author Billy Lee Brammer. Set in an unnamed state identical to Texas, each novella has a different protagonist: Roy Sherwood, a member of the state legislature; Neil Christiansen, the state's junior senator; and Jay McGown, the governor's speech-writer. The governor himself, Arthur Fenstemaker, a master politician (said to have been based on Brammer's mentor Lyndon Johnson) serves as the dominant figure throughout. The book also includes characters based on Brammer, his wife Nadine, Johnson's wife Ladybird, and his brother Sam Houston Johnson. The book has been widely acclaimed one of the best American political novels ever written. The proletarian novel is written by workers, mainly for other workers. It overlaps and sometimes is synonymous with the working-class novel, socialist novel, social-problem novel (also problem novel, sociological novel, or social novel), propaganda or thesis novel, and socialist-realism novel. The intention of the writers of proletarian literature is to lift the workers from the slums by inspiring them to embrace the possibilities of social change or of a political revolution. As such, it is a form of political fiction. The proletarian novel may comment on political events, systems, and theories, and is frequently seen as an instrument to promote social reform or political revolution among the working classes. Proletarian literature is created especially by communist, socialist, and anarchist authors. It is about the lives of the poor, and the period from 1930 to 1945, in particular, produced many such novels. However, proletarian works were also produced before and after those dates. In Britain, the terms "working-class" literature, novel, etc., are more generally used. A closely related type of novel, which frequently has a political dimension, is the social novel – also known as the "social-problem" or "social-protest" novel – a "work of fiction in which a prevailing social problem, such as gender, race, or class prejudice, is dramatized through its effect on the characters of a novel". More specific examples of social problems that are addressed in such works include poverty, conditions in factories and mines, the plight of child labor, violence against women, rising criminality, and epidemics caused by overcrowding and poor sanitation in cities. Charles Dickens was a fierce critic of the poverty and social stratification of Victorian society. Karl Marx asserted that Dickens "issued to the world more political and social truths than have been uttered by all the professional politicians, publicists and moralists put together". On the other hand, George Orwell, in his essay on Dickens, wrote: "There is no clear sign that he wants the existing order to be overthrown, or that he believes it would make very much difference if it were overthrown. For in reality his target is not so much society as 'human nature'." Dickens's second novel, "Oliver Twist" (1839), shocked readers with its images of poverty and crime: it destroyed middle-class polemics about criminals, making any pretence to ignorance about what poverty entailed impossible. Charles Dickens's "Hard Times" (1854) is set in a small Midlands industrial town and particularly criticizes the effect of Utilitarianism on the lives of cities' working classes. John Ruskin declared "Hard Times" his favourite Dickens work due to its exploration of important social questions. Walter Allen characterised "Hard Times" as an unsurpassed "critique of industrial society", "This is a list of a few of the early or notable examples; others belong on the main list"
https://en.wikipedia.org/wiki?curid=24850
Potato chip Potato chips (often just chips), or crisps (in British English), are thin slices of potato that have been either deep fried or baked until crunchy. They are commonly served as a snack, side dish, or appetizer. The basic chips are cooked and salted; additional varieties are manufactured using various flavorings and ingredients including herbs, spices, cheeses, other natural flavors, artificial flavors, and additives. Potato chips form a large part of the snack food and convenience food market in Western countries. The global potato chip market generated total revenue of US$16.49 billion in 2005. This accounted for 35.5% of the total savory snacks market in that year ($46.1 billion). The earliest known recipe for something similar to today's potato chips is in William Kitchiner's book "The Cook's Oracle" published in 1817, which was a bestseller in the United Kingdom and the United States. The 1822 edition's recipe for "Potatoes fried in Slices or Shavings" reads "peel large potatoes… cut them in shavings round and round, as you would peel a lemon; dry them well in a clean cloth, and fry them in lard or dripping". An 1825 British book about French cookery calls them "Pommes de Terre frites" (second recipe) and calls for thin slices of potato fried in "clarified butter or goose dripping", drained and sprinkled with salt. Early recipes for potato chips in the US are found in Mary Randolph's "Virginia House-Wife" (1824) and in N.K.M. Lee's "Cook's Own Book" (1832), both of which explicitly cite Kitchiner. A legend associates the creation of potato chips with Saratoga Springs, New York decades later than the first recorded recipe. By the late nineteenth century, a popular version of the story attributed the dish to George Crum, a cook at Moon's Lake House who was trying to appease an unhappy customer on 24 August 1853. The customer kept sending back his French-fried potatoes, complaining that they were too thick, too "soggy", or not salted enough. Frustrated, Crum sliced several potatoes extremely thin, fried them to a crisp, and seasoned them with extra salt. To his surprise, the customer loved them. They soon came to be called "Saratoga Chips", a name that persisted into the mid-twentieth century. A version of this story was popularized in a 1973 national advertising campaign by St. Regis Paper Company which manufactured packaging for chips, claiming that Crum's customer was Cornelius Vanderbilt. Crum was already renowned as a chef at the time, and he owned a lakeside restaurant by 1860 which he called Crum's House. The "Saratoga Chips" brand name still exists today. In the 20th century, potato chips spread beyond chef-cooked restaurant fare and began to be mass-produced for home consumption. The Dayton, Ohio-based Mikesell's Potato Chip Company, founded in 1910, identifies as the "oldest potato chip company in the United States". New England-based Tri-Sum Potato Chips, founded in 1908 as the Leominster Potato Chip Company, in Leominster, Massachusetts, claims to be America's first potato chip manufacturer. In an idea originated by the Smiths Potato Crisps Company Ltd, formed in 1920, Frank Smith packaged a twist of salt with his chips in greaseproof paper bags, which were sold around London. The potato chip remained otherwise unseasoned until an innovation by Joe "Spud" Murphy, the owner of the Irish crisps company Tayto, who in the 1950s developed a technology to add seasoning during manufacture. After some trial and error, Murphy and his employee Seamus Burke produced the world's first seasoned chips: Cheese & Onion and Salt & Vinegar. Companies worldwide sought to buy the rights to Tayto's technique. The first flavored chips in the United States, barbecue flavor, were being manufactured and sold by 1954. In 1958, Herr's was the first company to introduce barbecue-flavored potato chips in Pennsylvania. Chips sold in markets were usually sold in tins or scooped out of storefront glass bins and delivered by horse and wagon. Early potato chip bags were wax paper with the ends ironed or stapled together. At first, potato chips were packaged in barrels or tins, which left chips at the bottom stale and crumbled. In the 1920s, Laura Scudder, an entrepreneur in Monterey Park, California, started having her workers take home sheets of wax paper to iron into the form of bags, which were filled with chips at her factory the next day. This pioneering method reduced crumbling and kept the chips fresh and crisp longer. This innovation, along with the invention of cellophane, allowed potato chips to become a mass-market product. Today, chips are packaged in plastic bags, with nitrogen gas blown in prior to sealing to lengthen shelf life, and provide protection against crushing. Chips were long made in a batch process, where the potato slices are rinsed with cold water to release starch, fried at a low temperature of , and continuously raked to prevent them from sticking together. Industrial advances resulted in a shift to production by a continuous process, running the chips through a vat of hot oil and drying them in a conveyor process. Some small producers continued to use a batch process, notably in Maui. In 1980, inspired by the Maui Chip, an entrepreneur started Cape Cod Potato Chips to produce thicker, batch-cooked "Hawaiian style" potato chips, which came to be known as kettle-style (US) or hand-cooked (UK) chips and became a premium, "gourmet" item. Kettle chips are thicker and the surface starch is not rinsed off, resulting in a style of chip called "hard bite". Little consistency exists in the English-speaking world for the name of this food. North American English uses "chips", though Canadians may also call French fries, especially thick ones, "chips" as well. "Crisps" may be used for thin fried potatoes made from batter. In the United Kingdom and Ireland, "crisps" are potato chips which are eaten at room temperature, whilst "chips" are similar to french fries (as in "fish and chips") and are served hot. In Australia, some parts of South Africa, New Zealand, India, and the West Indies, especially in Barbados, both forms of potato product are simply known as "chips", as are the larger "home-style" variety. In the north of New Zealand, they are sometimes known as "chippies", but they are marketed as "chips" throughout the country. In Australia and New Zealand, a distinction is sometimes made between "hot chips" (fried potatoes) and "chips" or "potato chips". In Bangladesh, they are generally known as "chip" or "chips", and much less frequently as "crisps" (pronounced "kirisp") and locally, "alu bhaja" (for their similarity to the fried potato dish, "bhajji"). In German-speaking countries (Austria, Germany: ""Kartoffelchips"", or colloquially "Pommes" or "Fritten"; Switzerland: ""Pommes Chips"") and in countries of the former SFR Yugoslavia, fried thin potato slices are known as "chips" (locally pronounced very similarly to the English pronunciation), with a clear distinction from French fries. In Brazil, "home-style" potato chips are known as ("Portuguese potatoes") if their sides are relatively smooth and ("Prussian potatoes") if their sides show a wafer biscuit-like pattern, whilst American-like industrial uniform potato chips made from a fried potato purée-based dough are known as "batata chips" ("potato chips"), or just . Most potato chips contain high levels of sodium, from salt. This has been linked to health issues such as high blood pressure. However, researchers at Queen Mary University of London in 2004 have noted that a small "bag of ready-salted crisps" contains less salt than a serving of many breakfast cereals, including "every brand of cornflakes on sale in the UK." Some potato chip companies have responded to the long-standing concerns by investing in research and development to modify existing recipes and create health-conscious products. PepsiCo research shows that about 80% of salt on chips is not sensed by the tongue before being swallowed. Frito-Lay spent $414 million in 2009 on product development, including development of salt crystals that would reduce the salt content of Lay's potato chips without adversely affecting flavor. Unsalted chips are available, e.g. the longstanding British brand Salt 'n' Shake, whose chips are not seasoned, but instead include a small salt sachet in the bag for seasoning to taste. Many other popular brands in the United States, such as Frito-Lay, also offer such a product. Another possible health concern related to potato chips is acrylamide, which is produced when potatoes are fried or baked at high temperatures. Studies show that laboratory animals exposed to high levels of acrylamide develop cancer; however, it is currently unclear whether a similar risk exists in humans. In August 2008, California Attorney General Jerry Brown announced a settlement with Frito-Lay, Kettle Foods, and Lance Inc. the makers of Cape Cod Potato Chips for violating the state's Safe Drinking Water and Toxic Enforcement Act. The state had alleged in 2005 that potato chips from these companies failed to document that they contained high levels of acrylamide, which is listed by California since the 1990s as a carcinogen. These companies paid fines and agreed to reduce acrylamide levels to be under 275 parts per billion. Many potato chip manufacturers attempt to remove burned and thus potentially acrylamide-rich chips before the packaging process. Large scanners are used to eliminate chips worst affected by heat. In Canada, seasonings include the unique all dressed, as well as dill pickle, jalapeño, ketchup, barbecue, sour cream and onion, and salt and vinegar. In 2006, Lay's introduced wasabi chips in Toronto and Vancouver, but no longer offers them. Lays has recently released new Lay's Poppables which come in 3 flavours: Sea Salt and Vinegar, Honey BBQ, Sea Salt and White cheddar. Loblaw, Canada's largest food retailer, offers several unusual flavors under its President's Choice brand, including poutine, maple bacon, Jamaican jerk chicken, Greek feta and olive, ballpark hot dog, and barbecue baby back ribs. In Hong Kong, the two prominent potato chips are the spicy "Ethnican" variety by Calbee, and barbecue by Jack 'n Jill. Lay's are also popular in Hong Kong. In Indonesia, potato chips are commonly called as "kripik kentang", traditionally fell under "kripik" category. Indonesian potato chips market is mainly ruled by two brands; Indofood's Chitato (since 1990s) and Lay's (Frito-Lay). In 2014, Japan's Calbee and Indonesia's Wings Food formed Calbeewings, a joint venture and marketed Potabee potato chips in Indonesia. Common potato chips flavors marketed in Indonesia include beef barbecue, spicy chicken, cheese and plain salted. Lay's potato chips sold in Indonesia is available in 6 flavors; honey butter, sour cream and onion, nori seaweed, beef barbecue, classic salty, and salmon teriyaki flavors. Potabee sold in Indonesia offers two flavors; beef BBQ and grilled seaweed. In 2018 Chitato launched "Do Us A Flavor" campaign that sell three unusual flavors; beef rendang, fried crab golden egg yolk, and mango sticky rice. In Ireland, the two most popular flavors are cheese and onion, and salt and vinegar. However in Ireland, the word "Tayto" is synonymous with potato chips after the Tayto brand and can be used to describe all varieties of chips, including those not produced by Tayto. Hunky Dorys and King are other popular Irish brands. In November 2010, the Tayto company opened a theme park called "Tayto Park". In Germany, Belgium and the Netherlands only two flavors were traditionally available, red paprika ("Paprika", sometimes also called "ungarisch" (from "Hungarian")) and salted ("gesalzen"). These are still by far the most common and popular types, but some vendors started to offer a number of other flavors such as sour cream and onion, cheese, oriental, or more exotic seasonings like "Chakalaka", "Currywurst", "Pommes" (french fries), ""Rot-weiss"" (red and white - french fries with tomato ketchup and mayonnaise). Potato chips made from ground potatoes are called "Stapelchips" rather than "Kartoffelchips" for legal reasons. In Colombia, lemon, chicken, chorizo, and sirloin steak with mushroom sauce flavored potato chips are sold. In Japan, flavors include norishio (nori and salt), "consommé", wasabi, soy sauce and butter, garlic, plum, barbecue, pizza, mayonnaise, and black pepper. Chili, scallop with butter, teriyaki, takoyaki, and yakitori chip flavors are also available. Major manufacturers are Calbee, Koikeya and Yamayoshi. The market in the United Kingdom is led by Walkers, which held 56% of the British crisp market in 2013. Walkers is known for its wide variety of crisps with the most popular flavors being Cheese & Onion, Salt & Vinegar, Prawn Cocktail, Beef and Onion, Roast Chicken, Smoky Bacon, Worcester Sauce, Pickled Onion, Tomato Ketchup, and Salt & Shake - Original. More exotic flavors are Thai sweet chili, roast pork and creamy mustard sauce, lime and Thai spices, chicken with Italian herbs, Spicy Sriracha, BBQ Pulled Pork, sea salt and cracked black pepper, sea salt and cider vinegar, spicy and aromatic curry, turkey and bacon, caramelized onion and sweet balsamic vinegar, Stilton and cranberry. Since 2008, Walkers has launched its "Do Us a Flavour" campaign, challenging the British public to think up unique flavors for their crisps. Six flavors were chosen from among the entries and released as special editions. In 2014 the public had to pick one of Walkers' base ingredients, which was made up of six flavors from around the UK – Somerset Cheddar, Devonshire Chicken, Norfolk Pork, Dorset Sour Cream, Vale of Evesham Tomatoes and Aberdeen Angus Beef – then add their own unique flavor. In 2018 Walkers launched six new flavors to celebrate the brand's 70th birthday, with each flavor representing a different decade. In 1981, hedgehog flavored crisps were produced by the landlord of "The Vaults" in Welshpool, Philip Lewis, for his customers who kept asking for them as a joke. They became unexpectedly popular, but this led to controversy. There was concern that real hedgehogs were being slaughtered, but it proved that they were actually flavored with pork fat. This then caused concern about false advertising, and so a compromise was agreed with the trading standards authorities: the labelling was amended to "Hedgehog® Flavoured Crisps". In the United States, potato chips are made by national chains like Frito-Lay, Pringles and Kettle Brand; major regional brands like Jay's of Chicago, Better Made of Detroit and Old Dutch of Minneapolis; and specialty brands with local or uneven distribution. Potato chip flavorings include variations of barbecue, as well as sour cream and onion, sour cream and cheddar, salt and vinegar, ranch, jalapeno and cheese. "Hot" is a common flavor such as Jay's Hot Stuff and Better Made Red Hot. Better Made also make Rainbow chips, heavily cooked dark-colored chips which would be rejected by most processors. In the Gulf South, Zapp's of Gramercy, Louisiana, makes kettle-cooked chips using regional flavors such as Crawtator, Cajun dill, Voodoo, and Creole onion. Pennsylvania leads the United States in potato chip production, and has been dubbed "the Potato Chip Capital" by several sources. Pennsylvania-based companies that produce potato chips include Utz Quality Foods, Herr's Snacks, Snyder's of Berlin, Snyder's of Hanover, Martin's Potato Chips, Wise Foods, Middleswarth Potato Chips, Dieffenbach's Potato Chips, Hartley's Potato Chips, Gibbles Foods, Stehman's Potato Chips, and Charles Chips. Many of these companies are members of SNAC International, the global snack industry association. Another type of potato chip, notably the Pringles and Lay's Stax brands, is made by extruding or pressing a dough made from dehydrated potato flour into the desired shape before frying. This makes chips that are uniform in size and shape, which allows them to be stacked and packaged in rigid cardboard or plastic canisters. Pringles are officially branded as "potato crisps" in the US. Pringles may be termed "potato chips" in Britain, to distinguish them from traditional "crisps". Munchos, another brand that uses the term "potato crisps", has deep air pockets in its chips that give it a curved shape, though the chips themselves resemble regular bagged chips. An additional variant of potato chips exists in the form of "potato sticks", also called "shoestring potatoes". These are made as extremely thin (2 to 3 mm) versions of the popular French fry but are fried in the manner of regular salted potato chips. A hickory-smoke-flavored version is popular in Canada, going by the vending machine name "Hickory Sticks". Potato sticks are typically packaged in rigid containers, although some manufacturers use flexible pouches, similar to potato chip bags. Potato sticks were originally packed in hermetically sealed steel cans. In the 1960s, manufacturers switched to the less expensive composite canister (similar to the Pringles container). Reckitt Benckiser was a market leader in this category under the Durkee Potato Stix and French's Potato Sticks names but exited the business in 2008. In 2014, French's reentered the market. A larger variant (about 1 cm thick) made with dehydrated potatoes is marketed as Andy Capp's Pub Fries, using the theme of a long-running British comic strip, which are baked and sold in a variety of flavors. Walkers make a similar product (using the Smiths brand) called "Chipsticks" which are sold in ready-salted and salt and vinegar flavors. Some companies have also marketed baked potato chips as an alternative with lower fat content. Additionally, some varieties of fat-free chips have been made using artificial, and indigestible, fat substitutes. These became well known in the media when an ingredient many contained, Olestra, was linked in some individuals to abdominal discomfort and loose stools. Many other products might be called "crisps" in Britain, but would not be classed as "potato chips" because they are not made with potato or are not chipped (for example, Wotsits, Quavers, Skips, Hula Hoops, and Monster Munch). Sweet potato chips are eaten in Korea, New Zealand, and Japan; parsnip, beetroot, and carrot crisps are available in the United Kingdom. India is famous for a large number of localized 'chips shops', selling not only potato chips, but also other varieties such as plantain chips, tapioca chips, yam chips, and even carrot chips. Plantain chips, also known as chifles or tostones, are also sold in the Western Hemisphere from Canada to Chile. In the Philippines, banana chips can be found sold at local stores. In Kenya, chips are made from arrowroot and cassava. In the United Kingdom, Sweden, Finland, and Australia, a new variety of Pringles made from rice has been released and marketed as lower in fat than its potato counterparts.
https://en.wikipedia.org/wiki?curid=24851
Prohibition Prohibition is the act or practice of forbidding something by law; more particularly the term refers to the banning of the manufacture, storage (whether in barrels or in bottles), transportation, sale, possession, and consumption of alcoholic beverages. The word is also used to refer to a period of time during which such bans are enforced. Some kind of limitation on the trade in alcohol can be seen in the Code of Hammurabi (ca.1772 BCE) specifically banning the selling of beer for money. It could only be bartered for barley: "If a beer seller do not receive barley as the price for beer, but if she receive money or make the beer a measure smaller than the barley measure received, they shall throw her into the water." In the early twentieth century, much of the impetus for the prohibition movement in the Nordic countries and North America came from moralistic convictions of pietistic Protestants. Prohibition movements in the West coincided with the advent of women's suffrage, with newly empowered women as part of the political process strongly supporting policies that curbed alcohol consumption. The first half of the 20th century saw periods of prohibition of alcoholic beverages in several countries: After several years, prohibition failed in North America and elsewhere. Rum-running or bootlegging became widespread, and organized crime took control of the distribution of alcohol. Distilleries and breweries in Canada, Mexico and the Caribbean flourished as their products were either consumed by visiting Americans or illegally exported to the United States. Chicago became notorious as a haven for prohibition dodgers during the time known as the Roaring Twenties. Prohibition generally came to an end in the late 1920s or early 1930s in most of North America and Europe, although a few locations continued prohibition for many more years. In some countries where the dominant religion forbids the use of alcohol, the production, sale, and consumption of alcoholic beverages is prohibited or restricted today. For example, in Saudi Arabia and Libya alcohol is banned; in Pakistan and Iran it is illegal with exceptions. In the British colony of Nigeria, missionary forces demanded prohibition of liquor, which proved highly unpopular. Both Africans and Europeans found illegal supplies such as secret stills, obtaining colonial liquor permits, and smuggling. The experiment began in 1890 and was repealed in 1939, During the coronavirus outbreak of 2020 alcohol sales, and even the transportation of alcohol outside of one's home, was made illegal. This order came into effect during the nation-wide lockdown on 27 March 2020. The purpose of the ban was intended to prevent drunken fights, reduce domestic violence, stop drunk driving, and eliminate the weekend binge-drinking so prevalent across South Africa. Police, medics, and analysts estimate - conservatively - that alcohol is involved in, or responsible for, at least 40% of all emergency hospital admissions. By reducing the number of people within hospitals, and of course within social gartherings, the goal of prohibition was to reduce the rate of transmission, and thus slow the spread of the virus. Sale of alcohol is banned in Afghanistan. In Bangladesh, alcohol is somewhat prohibited due to its proscription in the Islamic faith. However, the purchase and consumption is allowed in the country. The Garo tribe consume a type of rice beer, and Christians in this country drink and purchase wine for their holy communion. In India alcohol is a state subject and individual states can legislate prohibition, but currently most states do not have prohibition and sale/consumption is freely available in 25 out of 29 states. Prohibition is in force in the states of Gujarat, Bihar and Nagaland, parts of Manipur, and the union territory of Lakshadweep. All other States and union territories of India permit the sale of alcohol. Election days and certain national holidays such as "Independence Day" are meant to be "dry days" when liquor sale is not permitted but consumption is allowed. Some Indian states observe dry days on major religious festivals/occasions depending on the popularity of the festival in that region. The Maldives ban the import of alcohol, x-raying all baggage on arrival. Alcoholic beverages are available only to foreign tourists on resort islands and may not be taken off the resort. Pakistan allowed the free sale and consumption of alcohol for three decades from 1947, but restrictions were introduced by Zulfikar Ali Bhutto just weeks before he was removed as prime minister in 1977. Since then, only members of non-Muslim minorities such as Hindus, Christians and Zoroastrians are allowed to apply for alcohol permits. The monthly quota is dependent upon one's income, but is actually about five bottles of liquor or 100 bottles of beer. In a country of 180 million, only about 60 outlets are allowed to sell alcohol. The Murree Brewery in Rawalpindi was once the only legal brewery, but today there are more. The ban officially is enforced by the country's Islamic Ideology Council, but it is not strictly policed. Members of religious minorities, however, often sell their liquor permits to Muslims as part of a continuing black market trade in alcohol. In 1955 Sri Lanka passed a law prohibiting adult women from buying alcohol. In January 2018, Finance Minister Mangala Samaraweera announced that the law would be amended, allowing women to legally consume alcohol and work in venues that sell alcohol. The legalization was overruled by President Maithripala Sirisena several days later. Since the 1979 Islamic Revolution, the sale and consumption of alcohol is banned in Iran. All people are banned from drinking alcohol but some people trade and sell it illegally. The consumption, importation and brewing of, and trafficking in liquor is strictly against the law. The sale, consumption, importation and brewing of, and trafficking in liquor is strictly against the law. Alcohol is banned in Yemen. In Brunei, alcohol consumption and sale is banned in public. Non-Muslims are allowed to purchase a limited amount of alcohol from their point of embarcation overseas for their own private consumption, and non-Muslims who are at least the age of 18 are allowed to bring in not more than two bottles of liquor (about two litres) and twelve cans of beer per person into the country. Alcohol sales are banned in small shops and convenience stores. Alcohol is banned only for Muslims in Malaysia due to its Islamic faith and sharia law. Nevertheless, alcoholic products can easily be found in supermarkets, specialty shops, and convenience stores all over the country. Non-halal restaurants also typically sell alcohol. There are only restrictions during elections in the Philippines. Alcohol is prohibited from purchase two days prior to an election. The Philippine Commission on Elections may opt to extend the liquor ban. In the 2010 elections, the liquor ban was a minimum two days; in the 2013 elections, there was a proposal that it be extended to five days. This was overturned by the Supreme Court. Other than election-related prohibition, alcohol is freely sold to anyone above the legal drinking age. Alcohol is prohibited from being sold during election time, from 6 pm the day prior to voting, until the end of the day of voting itself. Alcohol is also prohibited on major Buddhist holy days, and sometimes on Royal Commemoration days, such as birthdays. Thailand also enforces time-limited bans on alcohol on a daily basis. Alcohol can only be legally purchased in stores or restaurants between 11 am–2 pm and 5 pm–midnight. This law is enforced by all major retailers (most notably 7-Eleven) and restaurants but is frequently ignored by the smaller 'mom and pop' stores. Hotels and resorts are exempt from the rules. The consumption of alcohol is also banned at any time within 200 meters of a filling station (where sale of alcohol is also illegal), schools, temples or hospitals as well as on board any type of road vehicle regardless of whether it is being consumed by the driver or passenger. At certain times of the year – Thai New Year (Songkran) as an example – the government may also enforce arbitrary bans on the sale and consumption of alcohol in specific public areas where large scale festivities are due to take place and large crowds are expected. On 14 September 2012, the government of the Czech Republic banned all sales of liquor with more than 20% alcohol. From this date on it was illegal to sell such alcoholic beverages in shops, supermarkets, bars, restaurants, gas stations, e-shops etc. This measure was taken in response to the wave of methanol poisoning cases resulting in the deaths of 18 people in the Czech Republic. Since the beginning of the "methanol affair" the total number of deaths has increased to 25. The ban was to be valid until further notice, though restrictions were eased towards the end of September. The last bans on Czech alcohol with regard to the poisoning cases were lifted on 10 October 2012, when neighbouring Slovakia and Poland allowed its import once again. The Nordic countries, with the exception of Denmark, have had a strong temperance movement since the late 1800s, closely linked to the Christian revival movement of the late 19th century, but also to several worker organisations. As an example, in 1910 the temperance organisations in Sweden had some 330,000 members, which was about 6% of a population of 5.5 million. This heavily influenced the decisions of Nordic politicians in the early 20th century. In 1907, the Faroe Islands passed a law prohibiting all sale of alcohol, which was in force until 1992. Very restricted private importation from Denmark was allowed from 1928 on. In 1914, Sweden put in place a rationing system, the Bratt System, in force until 1955. A referendum in 1922 rejected an attempt to enforce total prohibition. In 1915, Iceland instituted total prohibition. The ban for wine was lifted in 1922 and spirits in 1935, but beer remained prohibited until 1989 (circumvented by mixing light beer and spirits). In 1916, Norway prohibited distilled beverages, and in 1917 the prohibition was extended to also include fortified wine and beer. The wine and beer ban was lifted in 1923, and in 1927 the ban of distilled beverages was also lifted. In 1919, Finland enacted prohibition, as one of the first acts after independence from the Russian Empire. Four previous attempts to institute prohibition in the early 20th century had failed due to opposition from the tsar. After a development similar to the one in the United States during its prohibition, with large-scale smuggling and increasing violence and crime rates, public opinion turned against the prohibition, and after a national referendum where 70% voted for a repeal of the law, prohibition was ended in early 1932. Today, all Nordic countries (with the exception of Denmark) continue to have strict controls on the sale of alcohol which is highly taxed (dutied) to the public. There are government monopolies in place for selling spirits, wine, and stronger beers in Norway (Vinmonopolet), Finland (Alko), Sweden (Systembolaget), Iceland (Vínbúðin), and the Faroe Islands (Rúsdrekkasøla Landsins). Bars and restaurants may, however, import alcoholic beverages directly or through other companies. Greenland, which is part of the Kingdom of Denmark, does not share its easier controls on the sale of alcohol. Greenland has (like Denmark) sales in food shops, but prices are high. Private import when traveling from Denmark is only allowed in small quantities. In the Russian Empire, a limited version of a Dry Law was introduced in 1914. It continued through the turmoil of the Russian Revolution of 1917 and the Russian Civil War into the period of Soviet Russia and the Soviet Union until 1925. Although the sale or consumption of commercial alcohol has never been prohibited by law in the United Kingdom, historically, various groups in the UK have campaigned for the prohibition of alcohol; including the Society of Friends (Quakers), The Methodist Church and other non-conformists, as well as temperance movements such as Band of Hope and temperance Chartist movements of the nineteenth century. Formed in 1853 and inspired by the Maine law in the United States, the United Kingdom Alliance aimed at promoting a similar law prohibiting the sale of alcohol in the UK. This hard-line group of prohibitionists was opposed by other temperance organisations who preferred moral persuasion to a legal ban. This division in the ranks limited the effectiveness of the temperance movement as a whole. The impotence of legislation in this field was demonstrated when the Sale of Beer Act 1854 which restricted Sunday opening hours had to be repealed, following widespread rioting. In 1859, a prototype prohibition bill was overwhelmingly defeated in the House of Commons. On 22 March 1917, during the First World War at a crowded meeting in the Queen's Hall in London (chaired by Alfred Booth) many influential people including Agnes Weston spoke, or letters from them were read out, against alcohol consumption, calling for prohibition; General Sir Reginald Hart wrote to the meeting that "Every experienced officer knew that practically all unhappiness and crime in the Army is due to drink". At the meeting, Lord Channing said that it was a pity that the whole Cabinet did not follow the example of King George V and Lord Kitchener when in 1914 those two spoke calling for complete prohibition for the duration of the war. Edwin Scrymgeour served as Member of Parliament for Dundee between 15 November 1922 and 8 October 1931. He remains the only person to have been elected to the House of Commons on a prohibitionist ticket. In 1922 he defeated incumbent Liberal member Winston Churchill; winning the seat for the Scottish Prohibition Party, which he had founded in 1901, and for which he had stood for election successfully as a Dundee Burgh Councillor in 1905 and unsuccessfully as a parliamentary candidate between 1908 and 1922. Indigenous peoples in Canada were subject to prohibitory alcohol laws under the "Indian Act" of 1876. Sections of the "Indian Act" regarding liquor were not repealed for over a hundred years, until 1985. An official, but non-binding, federal referendum on prohibition was held in 1898. Prime Minister Wilfrid Laurier's government chose not to introduce a federal bill on prohibition, mindful of the strong antipathy in Quebec. As a result, Canadian prohibition was instead enacted through laws passed by the provinces during the first twenty years of the 20th century, especially during the 1910s. However, Canada did enact a national prohibition from 1918 to 1920, as a temporary wartime measure. Much of the rum-running during prohibition took place in Windsor, Ontario. The provinces later repealed their prohibition laws, mostly during the 1920s, although some local municipalities remain dry. Some communities in the Chiapas state of southern Mexico are under the control of the radical leftist Zapatista Army of National Liberation, and often ban alcohol as part of what was described as "a collective decision". This prohibition has been used by many villages as a way to decrease domestic violence and has generally been favored by women. However, this prohibition is not recognized by federal Mexican law as the Zapatista movement is strongly opposed by the federal government. The sale and purchase of alcohol is prohibited on and the night before certain national holidays, such as "Natalicio de Benito Juárez" (birthdate of Benito Juárez) and "Día de la Revolución", which are meant to be dry nationally. The same "dry law" applies to the days before presidential elections every six years. Prohibition in the United States focused on the manufacture, transportation, and sale of alcoholic beverages; however, exceptions were made for medicinal and religious uses. Alcohol consumption was never illegal under federal law. Nationwide Prohibition did not begin in the United States until January 1920, when the Eighteenth Amendment to the U.S. Constitution went into effect. The 18th amendment was ratified in 1919, and was repealed in December 1933 with the ratification of the Twenty-first Amendment. Concern over excessive alcohol consumption began during the American colonial era, when fines were imposed for drunken behavior and for selling liquor without a license. In the eighteenth century, when drinking was a part of everyday American life, Protestant religious groups, especially the Methodists, and health reformers, including Benjamin Rush and others, urged Americans to curb their drinking habits for moral and health reasons. In particular, Benjamin Rush believed Americans were drinking hard spirits in excess, so he created "A Moral and Physical Thermometer," displaying the progression of behaviors caused by the consumption of various alcohols. By the 1840s the temperance movement was actively encouraging individuals to immediately stop drinking. Music (a completely new genre) was composed and performed in support of the efforts, both in social contexts and in response to state legislation attempts to regulate alcohol. Many took a pledge of total abstinence (teetotalism) from drinking distilled liquor as well as beer and wine. However, the issue of slavery, and then the Civil War, overshadowed the temperance movement, and temperance groups petered out until they found new life in the 1870s. Prohibition was a major reform movement from the 1870s until the 1920s, when nationwide prohibition went into effect. Prohibition was supported by evangelical Protestant churches, especially the Methodists, Baptists, Presbyterians, Disciples of Christ, Congregationalists, Quakers, and Scandinavian Lutherans. Opposition came from Catholics, Episcopalians, and German Lutherans. Kansas and Maine were early adopters of statewide prohibition. Following passage of the Maine law, Delaware, Ohio, Illinois, Rhode Island, Minnesota, Massachusetts, Connecticut, Pennsylvania, and New York, among others, soon passed statewide prohibition legislation; however, a number of these laws were overturned. Women served a special role in the temperance movement. Along with prostitution, alcohol was seen as a vice that kept men out of their homes and caused them to oppress their wives. Carrie Nation, a middle-aged woman living in Kansas in the early 1900s, grew tired of the moral protesting, and began a campaign destroying bars first in Kansas and later across the entire United States. She said once that "almost everyone who was in jail was directly or indirectly there from the influence of intoxicating drinks," which encapsulated people's negative attitudes towards alcohol at the time. Nation also said, after she destroyed a painting of a nude woman, "It is very significant that the pictures of naked women are in saloons. Women are stripped of everything by them. Her husband is torn from her, she is robbed of her sons, her home, her food, and her virtue, and then they strip her clothes off and hang her up bare of all things!" As temperance groups continued to promote prohibition, other groups opposed increased alcohol restrictions. For example, Chicago's citizens fought against enforcing Sunday closings laws in the 1850s, which included mob violence. It was also during this time when patent medicines, many of which contained alcohol, gained popularity. During the American Civil War efforts at increasing federal revenue included imposition of taxes on liquor and beer. The liquor industry responded to the taxes by forming an industry lobby, the United States Brewers Association, that succeeded in reducing the tax rate on beer from $1 to 60 cents. The Women's Crusade of 1873 and the Woman's Christian Temperance Union (WCTU), founded in 1874, "marked the formal entrance of women into the temperance movement." Organizations such as the Women's Christian Temperance Movement were a venue through which certain women organized and demanded political action, well before they were granted the vote. The WCTU and the Prohibition Party, organized in 1869, remained major players in the temperance movement until the early twentieth century, when the Anti-Saloon League, formed in 1895, emerged as the movement's leader. Between 1880 and 1890, although several states enacted local option laws that allowed counties or towns to go dry by referendum, only six states had statewide prohibition by state statute or constitutional amendment. The League, with the support of evangelical Protestant churches and other Progressive-era reformers continued to press for prohibition legislation. Opposition to prohibition was strong in America's urban industrial centers, where a large, immigrant, working-class population generally opposed it, as did Jewish and Catholic religious groups. In the years leading up to World War I, nativism, American patriotism, distrust of immigrants, and anti-German sentiment became associated with the prohibition movement. Through the use of pressure politics on legislators, the League and other temperance reformers achieved the goal of nationwide prohibition by emphasizing the need to destroy the moral corruption of the saloons and the political power of the brewing industry, and to reduce domestic violence in the home. By 1913, nine states had statewide prohibition and thirty-one others had local option laws in effect, which included nearly 50% of the U.S. population. At that time the League and other reformers turned their efforts toward attaining a constitutional amendment and grassroots support for nationwide prohibition. In December 1917, Congress submitted a constitutional amendment on nationwide prohibition to the states for ratification. The new constitutional amendment prohibited "the manufacture, sale, or transportation of intoxicating liquors within, the importation thereof into, or the exportation thereof from the United States and all territory subject to the jurisdiction thereof for beverage purposes". It was ratified and became law on January 16, 1919, assuring its passage into law. On October 28, 1919, Congress passed the National Prohibition Act, also known as the Volstead Act, which provided enabling legislation to implement the Eighteenth Amendment. When the National Prohibition Act was passed on October 28, 1919, thirty-three of the forty-eight states were already dry. After a year's required delay, national prohibition began on January 16, 1920. During the first years of Prohibition, the new law was enforced in regions such as the rural South and western states, where it had popular support; however, in large urban cities and small industrial or mining towns, residents defied or ignored the law. The Ku Klux Klan was a major supporter of Prohibition; once it was passed they helped with its enforcement. For example, in 1923, Klansmen traded pistol shots with bootleggers, burned down roadhouses, and whipped liquor sellers, and anybody else who broke the moral code. The Prohibition was effective in reducing per-capita consumption, and consumption remained lower for a quarter-century after Prohibition had been repealed. Prohibition reduced alcohol consumption but did not stop it. Drinking itself was never illegal, although manufacturing and sale of alcoholic beverages was outlawed, so people who had bought alcohol before January 16, 1920, could and did continue to serve it privately. In addition, the illicit market soon grew to about two-thirds its pre-Prohibition levels. Illegal stills flourished in remote rural areas as well as city slums, and large quantities were smuggled from Canada. Bootlegging – and the related speakeasies – became a major business activity for organized crime groups, under leaders such as Al Capone in Chicago and Lucky Luciano in New York City. Indeed, Capone became a national symbol of Prohibition's violent side and was a top target for President Hoover. Prohibition lost support during the Great Depression, which started in 1929. So-called "wets" – people in favor of repeal – argued that legal sales would reduce violent gang crime, increase employment and raise tax revenues. The repeal movement was initiated and financed by the Association Against the Prohibition Amendment, who worked to elect Congressmen who agreed to support repeal. The group's wealthy supporters included John D. Rockefeller, Jr., S. S. Kresge, and the Du Pont family, among others, who had abandoned the dry cause. Pauline Sabin, a wealthy Republican who founded the Women's Organization for National Prohibition Reform (WONPR), argued that Prohibition should be repealed because it made the United States a nation of hypocrites and undermined its respect for the rule of law. This hypocrisy and the fact that women had initially led the prohibition movement convinced Sabin to establish the WONPR. Their efforts eventually led to the repeal of prohibition. When Sabin's fellow Republicans would not support her efforts, she went to the Democrats, who switched their support of the dry cause to endorse repeal under the leadership of liberal politicians such as Fiorello La Guardia and Franklin D. Roosevelt. Sabin and her supporters emphasized that repeal would generate enormous sums of much-needed tax revenue, and weaken the base of organized crime. Repeal of Prohibition was accomplished with the ratification of the Twenty-first Amendment on December 5, 1933. Under its terms, states were allowed to set their own laws for the control of alcohol. Following repeal, public interest in an organized prohibition movement dwindled. However, it survived for a while in a few southern and border states. To this day, there are still counties and parishes within the US known as "dry," where the sale of alcohol – liquor, and sometimes wine and beer – is prohibited. Some such counties/parishes/municipalities have adopted "Moist county" policies, however, in order to expand tax revenue. Some municipalities regulate when alcohol can be sold; an example is restricting or banning sales on Sunday, under the so-called "blue laws." Between 1832 and 1953, federal legislation prohibited the sale of alcohol to aboriginal Americans, with very limited success. After 1953, aboriginal American communities and reservations were permitted to pass their own local ordinances governing the sale of alcoholic beverages. In Venezuela, twenty-four hours before every election, the government prohibits the sale and distribution of alcoholic beverages throughout the national territory, including the restriction to all dealers, liquor stores, supermarkets, restaurants, wineries, pubs, bars, public entertainment, clubs and any establishment that markets alcoholic beverages. The same is done during Holy Week as a measure to reduce the alarming rate of road traffic accidents during these holidays. The Australian Capital Territory (then the Federal Capital Territory) was the first jurisdiction in Australia to have prohibition laws. In 1911, King O'Malley, then Minister of Home Affairs, shepherded laws through Parliament preventing new issue or transfer of licences to sell alcohol, to address unruly behaviour among workers building the new capital city. Prohibition was partial, since possession of alcohol purchased outside of the Territory remained legal and the few pubs that had existing licences could continue to operate. The Federal Parliament repealed the laws after residents of the Federal Capital Territory voted for the end of them in a 1928 plebiscite. Since then, some state governments and local councils have enacted dry areas. This is where the purchase or consumption of alcohol is only permitted in licensed areas such as liquor stores, clubs, cafes, bars, hotels, restaurants, and also private homes. In public places such as streets, parks, and squares, consumption is not permitted, but carrying bottles that were purchased at licensed venues is allowed. Almost all dry areas are small defined districts within larger urban or rural communities. More recently, alcohol has been prohibited in many remote indigenous communities. Penalties for transporting alcohol into these "dry" communities are severe and can result in confiscation of any vehicles involved; in dry areas within the Northern Territory, all vehicles used to transport alcohol are seized. In New Zealand, prohibition was a moralistic reform movement begun in the mid-1880s by the Protestant evangelical and Nonconformist churches and the Woman's Christian Temperance Union and after 1890 by the Prohibition League. It assumed that individual virtue was all that was needed to carry the colony forward from a pioneering society to a more mature one, but it never achieved its goal of national prohibition. However, both the Church of England and the largely Irish Catholic Church rejected prohibition as an intrusion of government into the church's domain, while the growing labor movement saw capitalism rather than alcohol as the enemy. Reformers hoped that the women's vote, in which New Zealand was a pioneer, would swing the balance, but the women were not as well organized as in other countries. Prohibition had a majority in a national referendum in 1911, but needed a 60% vote to pass. The movement kept trying in the 1920s, losing three more referenda by close votes; it managed to keep in place a 6 p.m. closing hour for pubs and Sunday closing. The Depression and war years effectively ended the movement. but their 6 p.m. closing hour remained until October 1967 when it was extended to 10 p.m. For many years, referenda were held for individual towns or electorates, often coincident with general elections. The ballots determined whether these individual areas would be "dry" – that is, alcohol could not be purchased or consumed in public in these areas. One notable example was the southern city of Invercargill, which was dry from 1907 to 1943. People wanting alcohol usually travelled to places outside the city (such as the nearby township of Lorneville or the town of Winton) to drink in the local pubs or purchase alcohol to take back home. The last bastion of this 'dry' area remains in force in the form of a licensing trust which still to this day governs the sale of liquor in Invercargill. The city does not allow the sale of alcohol (beer and wine included) in supermarkets unlike the remainder of New Zealand, and all form of alcohol regardless of the sort can only be sold in bars and liquor stores. Prohibition was of limited success in New Zealand as—like in other countries—it led to organised bootlegging. The most famous bootlegged alcohol in New Zealand was that produced in the Hokonui Hills close to the town of Gore (not coincidentally, the nearest large town to Invercargill). Even today, the term "Hokonui" conjures up images of illicit whisky to many New Zealanders. In many countries in Latin America, the Philippines, Thailand, Turkey and several US states, the sale but not the consumption of alcohol is prohibited before and during elections.
https://en.wikipedia.org/wiki?curid=24856
Phenothiazine Phenothiazine, abbreviated PTZ, is an organic compound that has the formula S(C6H4)2NH and is related to the thiazine-class of heterocyclic compounds. Derivatives of phenothiazine are highly bioactive and have widespread use and rich history. The derivatives chlorpromazine and promethazine revolutionized the field of psychiatry and allergy treatment, respectively. An earlier derivative, methylene blue, was one of the first antimalarial drugs, and derivatives are under investigation as possible anti-infective drugs. Phenothiazine is a prototypical pharmaceutical lead structure in medicinal chemistry. Phenothiazine itself is only of theoretical interest, but its derivatives revolutionized psychiatry, other fields of medicine, and pest management. Other derivatives have been studied for possible use in advanced batteries and fuel cells. In 1876, methylene blue, a derivative of phenothiazine, was synthesized by Heinrich Caro at BASF. The structure was deduced in 1885 by Heinrich August Bernthsen. Bernthsen synthesized phenothiazine in 1883. In the mid 1880s, Paul Ehrlich began to use methylene blue in his cell staining experiments that led to pioneering discoveries about different cell types. He was awarded a Nobel Prize based in part on that work. He became particularly interested in its use to stain bacteria and parasites such as "Plasmodiidae" – the genus that includes the malaria pathogen – and found that it could be stained with methylene blue. He thought methylene blue could possibly be used in the treatment of malaria, tested it clinically, and by the 1890s methylene blue was being used for that purpose. For the next several decades, research on derivatives lapsed until phenothiazine itself came to market as an insecticide and deworming drug. In the 1940s, chemists working with Paul Charpentier at Rhone-Poulenc Laboratories in Paris (a precursor company to Sanofi), began making derivatives. This work led to promethazine which had no activity against infective organisms, but did have good antihistamine activity, with a strong sedative effect. It went to market as a drug for allergies and for anesthesia. As of 2012 it was still on the market. At the end of the 1940s the same lab produced chlorpromazine which had an even stronger sedative and soothing effect, and Jean Delay and Pierre Deniker attempted to use it on their psychiatric patients, publishing their results in the early 1950s. The strong effects they found opened the door of the modern field of psychiatry and led to a proliferation of work on phenothiazine derivatives. The systematic research conducted by chemists to explore phenothiazine derivatives and their activity was a pioneering example of medicinal chemistry; phenothiazine is often discussed as a prototypical example of a pharmaceutical lead structure. The term "phenothiazines" describes the largest of the five main classes of antipsychotic drugs. These drugs have antipsychotic and, often, antiemetic properties, although they may also cause severe side effects such as extrapyramidal symptoms (including akathisia and tardive dyskinesia), hyperprolactinaemia, and the rare but potentially fatal neuroleptic malignant syndrome, as well as substantial weight gain. Use of phenothiazines has been associated with antiphospholipid syndrome, but no causal relationship has been established. Phenothiazine antipsychotics are classified into three groups that differ with respect to the substituent on nitrogen: the aliphatic compounds (bearing acyclic groups), the "piperidines" (bearing piperidine-derived groups), and the piperazine (bearing piperazine-derived substituents). The synthetic dye methylene blue, containing the structure, was described in 1876. Many water-soluble phenothiazine derivatives, such as methylene blue, methylene green, thionine, and others, can be electropolymerized into conductive polymers used as electrocatalysts for NADH oxidation in enzymatic biosensors and biofuel cells. Phenothiazine is used as an anaerobic inhibitor for acrylic acid polymerization, often used as an in-process inhibitor during the purification of acrylic acid. Like many commercially significant compounds, phenothiazine has numerous trade names, including AFI-Tiazin, Agrazine, Antiverm, Biverm, Dibenzothiazine, Orimon, Lethelmin, Souframine, Nemazene, Vermitin, Padophene, Fenoverm, Fentiazine, Contaverm, Fenothiazine, Phenovarm, Ieeno, ENT 38, Helmetina, Helmetine, Penthazine, XL-50, Wurm-thional, Phenegic, Phenovis, Phenoxur, and Reconox. Phenothiazine was formerly used as an insecticide and as a drug to treat infections with parasitic worms (anthelminthic) in livestock and people, but its use for those purposes has been superseded by other chemicals. Phenothiazine was introduced by DuPont as an insecticide in 1935. About 3,500,000 pounds were sold in the US in 1944. However, because it was degraded by sunlight and air, it was difficult to determine how much to use in the field, and its use waned in the 1940s with the arrival of new pesticides like DDT that were more durable. As of July 2015 it is not registered for pesticide use in the US, Europe, or Australia. It was introduced as anthelminthic in livestock in 1940 and is considered, with thiabendazole, to be the first modern anthelminthic. The first instances of resistance were noted in 1961. Uses for this purpose in the US are still described but it has "virtually disappeared from the market." In the 1940s it also was introduced as antihelminthic for humans; since it was often given to children, the drug was often sold in chocolate, leading to the popular name, "worm chocolate." Phenothiazine was superseded by other drugs in the 1950s. The central C4SN ring is folded in phenothiazines. The compound was originally prepared by Bernthsen in 1883 via the reaction of diphenylamine with sulfur, but more recent syntheses rely on the cyclization of 2-substituted diphenyl sulfides. Few pharmaceutically significant phenothiazines are prepared from phenothiazine, although some of them are. Phenothiazines are electron donors, forming charge-transfer salts with many acceptors.
https://en.wikipedia.org/wiki?curid=24857
Pale Fire Pale Fire is a 1962 novel by Vladimir Nabokov. The novel is presented as a 999-line poem titled "Pale Fire", written by the fictional poet John Shade, with a foreword, lengthy commentary and index written by Shade's neighbor and academic colleague, Charles Kinbote. Together these elements form a narrative in which both fictional authors are central characters. "Pale Fire" has spawned a wide variety of interpretations and a large body of written criticism, which Finnish literary scholar estimated in 1995 as more than 80 studies. The Nabokov authority Brian Boyd has called it "Nabokov's most perfect novel", and the critic Harold Bloom called it "the surest demonstration of his own genius ... that remarkable tour de force". It was ranked 53rd on the list of the Modern Library 100 Best Novels and 1st on the American literary critic Larry McCaffery's "." Starting with the epigraph and table of contents, "Pale Fire" looks like the publication of a 999-line poem in four cantos ("Pale Fire") by the fictional John Shade with a foreword, extensive commentary, and index by his self-appointed editor, Charles Kinbote. Kinbote's commentary takes the form of notes to various numbered lines of the poem. Here and in the rest of his critical apparatus, Kinbote explicates the poem very little. Focusing instead on his own concerns, he divulges what proves to be the plot piece by piece, some of which can be connected by following the many cross-references. Espen Aarseth noted that "Pale Fire" "can be read either unicursally, straight through, or multicursally, jumping between the comments and the poem." Thus, although the narration is non-linear and multidimensional, the reader can still choose to read the novel in a linear manner without risking misinterpretation. The novel's unusual structure has attracted much attention, and it is often cited as an important example of metafiction; it has also been called a poioumenon. The connection between "Pale Fire" and hypertext was stated soon after its publication; in 1969, the information-technology researcher Ted Nelson obtained permission from the novel's publishers to use it for a hypertext demonstration at Brown University. A 2009 paper also compares "Pale Fire" to hypertext. The interaction between Kinbote and Shade takes place in the fictitious small college town of New Wye, Appalachia, where they live across a lane from each other, from February to July 1959. Kinbote writes his commentary from then to October 1959 in a tourist cabin in the equally fictitious western town of Cedarn, Utana. Both authors recount many earlier events, Shade mostly in New Wye and Kinbote in New Wye and in Europe, especially the "distant northern land" of Zembla. Shade's poem digressively describes many aspects of his life. Canto 1 includes his early encounters with death and glimpses of what he takes to be the supernatural. Canto 2 is about his family and the apparent suicide of his daughter, Hazel Shade. Canto 3 focuses on Shade's search for knowledge about an afterlife, culminating in a "faint hope" in higher powers "playing a game of worlds" as indicated by apparent coincidences. Canto 4 offers details on Shade's daily life and creative process, as well as thoughts on his poetry, which he finds to be a means of somehow understanding the universe. In Kinbote's editorial contributions he tells three stories intermixed with each other. One is his own story, notably including what he thinks of as his friendship with Shade. After Shade was murdered, Kinbote acquired the manuscript, including some variants, and has taken it upon himself to oversee the poem's publication, telling readers that it lacks only line 1000. Kinbote's second story deals with King Charles II, "The Beloved", the deposed king of Zembla. King Charles escaped imprisonment by Soviet-backed revolutionaries, making use of a secret passage and brave adherents in disguise. Kinbote repeatedly claims that he inspired Shade to write the poem by recounting King Charles's escape to him and that possible allusions to the king, and to Zembla, appear in Shade's poem, especially in rejected drafts. However, no explicit reference to King Charles is to be found in the poem. Kinbote's third story is that of Gradus, an assassin dispatched by the new rulers of Zembla to kill the exiled King Charles. Gradus makes his way from Zembla through Europe and America to New Wye, suffering comic mishaps. In the last note, to the missing line 1000, Kinbote narrates how Gradus killed Shade by mistake. Towards the end of the narrative, Kinbote all but states that he is in fact the exiled King Charles, living incognito; however, enough details throughout the story, as well as direct statements of ambiguous sincerity by Kinbote towards the novel's end, suggest that King Charles and Zembla are both fictitious. In the latter interpretation, Kinbote is delusional and has built an elaborate picture of Zembla complete with samples of a constructed language as a by-product of insanity; similarly, Gradus was simply an unhinged man trying to kill Shade, and his backstory as a revolutionary assassin is also made up. In an interview, Nabokov later claimed that Kinbote committed suicide after finishing the book. The critic Michael Wood has stated, "This is authorial trespassing, and we don't have to pay attention to it", but Brian Boyd has argued that internal evidence points to Kinbote's suicide. One of Kinbote's annotations to Shade's poem (corresponding to line 493) addresses the subject of suicide at some length. As Nabokov pointed out himself, the title of John Shade's poem is from Shakespeare's "Timon of Athens:" "The moon's an arrant thief, / And her pale fire she snatches from the sun" (Act IV, scene 3), a line often taken as a metaphor about creativity and inspiration. Kinbote quotes the passage but does not recognize it, as he says he has access only to an inaccurate Zemblan translation of the play "in his Timonian cave", and in a separate note he even rails against the common practice of using quotations as titles. Some critics have noted a secondary reference in the book's title to "Hamlet", where the Ghost remarks how the glow-worm "'gins to pale his uneffectual fire" (Act I, scene 5). The title is first mentioned in the foreword: "I recall seeing him from my porch, on a brilliant morning, burning a whole stack of [index cards of drafts of the poem] in the pale fire of the incinerator...". According to Norman Page, "Pale Fire" excited as diverse criticism as any of Nabokov's novels. Mary McCarthy's review was extremely laudatory; the Vintage edition excerpts it on the front cover. She tried to explicate hidden references and connections. Dwight Macdonald responded by saying the book was "unreadable" and both it and McCarthy's review were as pedantic as Kinbote. Anthony Burgess, like McCarthy, extolled the book, while Alfred Chester condemned it as "a total wreck". Some other early reviews were less decided, praising the book's satire and comedy but noting its difficulty and finding its subject slight or saying that its artistry offers "only a kibitzer's pleasure". Macdonald called the reviews he had seen, other than McCarthy's, "cautiously unfavorable". "TIME" magazine's 1962 review stated that "Pale Fire does not really cohere as a satire; good as it is, the novel in the end seems to be mostly an exercise in agility – or perhaps in bewilderment", though this did not prevent "TIME" from including the book in its 2005 list of the 100 best English-language novels published since 1923. The first Russian translation of the novel, one created by Véra Nabokov, its dedicatee, was published in 1983 by Ardis in Ann Arbor, Michigan (Alexei Tsvetkov initially played an important role in this translation). After Nabokov's reputation was rehabilitated in the Soviet Union (his novels started being published there in 1986 and the first book composed entirely of Nabokov's works was printed in 1988), "Pale Fire" was published in 1991 in Sverdlovsk (in Sergei Ilyin's Russian translation). Some readers concentrate on the apparent story, focusing on traditional aspects of fiction such as the relationship among the characters. In 1997, Brian Boyd published a much-discussed study arguing that the ghost of John Shade influenced Kinbote's contributions. He expanded this essay into a book in which he also argues that, in order to trigger Shade's poem, Hazel Shade's ghost induced Kinbote to recount his Zemblan delusions to Shade. Some readers, starting with Mary McCarthy and including Boyd, Nabokov's annotator Alfred Appel, and D. Barton Johnson, see Charles Kinbote as an alter-ego of the insane Professor V. Botkin, to whose delusions John Shade and the rest of the faculty of Wordsmith College generally condescend. Nabokov himself endorsed this reading, stating in an interview in 1962 (the novel's year of publication) that "Pale Fire" "is full of plums that I keep hoping somebody will find. For instance, the nasty commentator is not an ex-King of Zembla nor is he professor Kinbote. He is professor Botkin, or Botkine, a Russian and a madman." The novel's intricate structure of teasing cross-references leads readers to this "plum". The Index, supposedly created by Kinbote, features an entry for a "Botkin, V.," describing this Botkin as an "American scholar of Russian descent"—and referring to a note in the Commentary on line 894 of Shade's poem, in which no such person is directly mentioned but a character suggests that "Kinbote" is "a kind of anagram of Botkin or Botkine". In this interpretation, "Gradus" the murderer is an American named Jack Grey who wanted to kill Judge Goldsworth, whose house "Pale Fire's" commentator—whatever his "true" name is—is renting. Goldsworth had condemned Grey to an asylum from which he escaped shortly before mistakenly killing Shade, who resembled Goldsworth. Other readers see a story quite different from the apparent narrative. "Shadeans" maintain that John Shade wrote not only the poem, but the commentary as well, having invented his own death and the character of Kinbote as a literary device. According to Boyd, Andrew Field invented the Shadean theory and Julia Bader expanded it; Boyd himself espoused the theory for a time. In an alternative version of the Shadean theory, Tiffany DeRewal and Matthew Roth argued that Kinbote is not a separate person but is a dissociated, alternative personality of John Shade. (An early reviewer had mentioned that "a case might be made" for such a reading.) "Kinboteans", a decidedly smaller group, believe that Kinbote invented the existence of John Shade. Boyd credits the Kinbotean theory to Page Stegner and adds that most of its adherents are newcomers to the book. Some readers see the book as oscillating undecidably between these alternatives, like the Rubin vase (a drawing that may be two profiles or a goblet). Though a minority of commentators believe or at least accept the possibility that Zembla is as "real" as New Wye, most assume that Zembla, or at least the operetta-quaint and homosexually gratified palace life enjoyed by Charles Kinbote before he is overthrown, is imaginary in the context of the story. The name "Zembla" (taken from "Nova Zembla", a former latinization of Novaya Zemlya) may evoke popular fantasy literature about royalty such as "The Prisoner of Zenda". As in other Nabokov books, however, the fiction is an exaggerated or comically distorted version of his own life as a son of privilege before the Russian Revolution and an exile afterwards, and the central murder has resemblances (emphasized by Priscilla Meyer) to Nabokov's father's murder by an assassin who was trying to kill someone else. Still other readers de-emphasize any sort of "real story" and may doubt the existence of such a thing. In the interplay of allusions and thematic links, they find a multifaceted image of English literature, criticism, or glimpses of a higher world and an afterlife. The first two lines of John Shade's 999-line poem, "Pale Fire", have become Nabokov's most quoted couplet: I was the shadow of the waxwing slain By the false azure in the window pane Like many of Nabokov's fictions, "Pale Fire" alludes to others of his. "Hurricane Lolita" is mentioned, and Pnin appears as a minor character. There are many resemblances to "Ultima Thule" and "Solus Rex", two short stories by Nabokov intended to be the first two chapters of a novel in Russian that he never continued. The placename Thule appears in "Pale Fire", as does the phrase "solus rex" (a chess problem in which one player has no pieces but the king). The book is also full of references to culture, nature, and literature. They include: See also "The Ambidextrous Universe", a later book referencing "Pale Fire" which in turn triggered a reciprocal response in a subsequent Nabokov novel ("", 1969).
https://en.wikipedia.org/wiki?curid=24861
Preservative A preservative is a substance or a chemical that is added to products such as food products, beverages, pharmaceutical drugs, paints, biological samples, cosmetics, wood, and many other products to prevent decomposition by microbial growth or by undesirable chemical changes. In general, preservation is implemented in two modes, chemical and physical. Chemical preservation entails adding chemical compounds to the product. Physical preservation entails processes such as refrigeration or drying. Preservative food additives reduce the risk of foodborne infections, decrease microbial spoilage, and preserve fresh attributes and nutritional quality. Some physical techniques for food preservation include dehydration, UV-C radiation, freeze-drying, and refrigeration. Chemical preservation and physical preservation techniques are sometimes combined. Antimicrobial preservatives prevent degradation by bacteria. This method is the most traditional and ancient type of preserving—ancient methods such as pickling and adding honey prevent microorganism growth by modifying the pH level. The most commonly used antimicrobial preservative is lactic acid. Common antimicrobial preservatives are presented in the table. Nitrates and nitrites are also antimicrobial. The detailed mechanism of these chemical compounds range from inhibiting growth of the bacteria to the inhibition of specific enzymes. Water-based home and personal care products use broad-spectrum preservatives, such as isothiazolinones and formaldehyde releasers, which may cause sensitization, allergic skin reactions, and toxicity to aquatic life. The oxidation process spoils most food, especially those with a high fat content. Fats quickly turn rancid when exposed to oxygen. Antioxidants prevent or inhibit the oxidation process. The most common antioxidant additives are ascorbic acid (vitamin C) and ascorbates. Thus, antioxidants are commonly added to oils, cheese, and chips. Other antioxidants include the phenol derivatives BHA, BHT, TBHQ and propyl gallate. These agents suppress the formation of hydroperoxides. Other preservatives include ethanol and methylchloroisothiazolinone. A variety of agents are added to sequester (deactivate) metal ions that otherwise catalyze the oxidation of fats. Common sequestering agents are disodium EDTA, citric acid (and citrates), tartaric acid, and lecithin. Citric and ascorbic acids target enzymes that degrade fruits and vegetables, e.g., mono/polyphenol oxidase which turns surfaces of cut apples and potatoes brown. Ascorbic acid and tocopherol, which are vitamins, are common preservatives. Smoking entails exposing food to a variety of phenols, which are antioxidants. Natural preservatives include rosemary and oregano extract, hops, salt, sugar, vinegar, alcohol, diatomaceous earth and castor oil. Traditional preservatives, such as sodium benzoate have raised health concerns in the past. Benzoate was shown in a study to cause hypersensitivity in some asthma sufferers. This has caused reexamination of natural preservatives which occur in vegetables. Preservatives have been used since prehistoric times. Smoked meat for example has phenols and other chemicals that delay spoilage. The preservation of foods has evolved greatly over the centuries and has been instrumental in increasing food security. The use of preservatives other than traditional oils, salts, paints, etc. in food began in the late 19th century, but was not widespread until the 20th century. The use of food preservatives varies greatly depending on the country. Many developing countries that do not have strong governments to regulate food additives face either harmful levels of preservatives in foods or a complete avoidance of foods that are considered unnatural or foreign. These countries have also proven useful in case studies surrounding chemical preservatives, as they have been only recently introduced. In urban slums of highly populated countries, the knowledge about contents of food tends to be extremely low, despite consumption of these imported foods. In ancient times the sun and wind naturally dried out foods. Middle Eastern and Oriental cultures started drying foods in 1,200 B.C. in the sun. The Romans used a lot of dry fruit. In the Middle Ages, people made “still houses” where fruits, vegetables, and herbs could dry out in climates that did not have strong sunlight. Sometimes fires were made to create heat to dry foods. Drying prevents yeasts and bread molds ("Rhizopus") from growing by removing moisture so bacteria cannot grow. Cellars, caves, and cool streams were used for freezing. American estates had ice houses built to store ice and food on the ice. The icehouse was then converted to an “icebox”. The Icebox was converted in the 1800s to mechanical refrigeration. Clarence Birdseye found in the 1800s that freezing meats and vegetables at a low temperature made them taste better. Fermenting was discovered when a few grains of barley were left in the rain and turned into beer. Microorganisms ferment the starch-derived sugars into alcohols. This is also how fruits are fermented into wine and cabbage into Kimchi or sauerkraut. Anthropologists believe that as early as 10,000 B.C people began to settle and grow barley. They began to make beer and believed that it was a gift from gods. It was used to preserve foods and to create more nutritious foods from less desirable ingredients. Vitamins are produced through fermentation by microorganisms making the end product more nutritious. Pickling occurs when foods are placed in a container with vinegar or another acid. It is thought that pickling came about when people used to place food in wine or beer to preserve it due to them having a low pH. Containers had to be stoneware or glass (vinegar will dissolve metal from pots). After the food was eaten, the pickling brine had other uses. Romans would make a concentrated pickle sauce called “garum”. It was very concentrated and the dish that it would be used in would only need a few drops to get the fish taste. Due to new foods arriving from Europe in the 16th century, food preservation increased. Ketchup originated from Europe as an oriental fish brine and when it made it to America, sugar was added. Pickling sauces were soon part of many recipes such as chutneys, relish, piccalilli, mustard, and ketchup when different spices were added to them. The beginning of curing was done through dehydration. Salting was used by early cultures to help desiccate foods. Many different salts were used from different places such as rock salt, sea salt, spiced salt, etc.. People began to experiment and found in the 1800s that some salts gave meat an appealing red color instead of the grey that they were used to. During their experimenting in the 1920s they realized this mixture of salts were nitrates (saltpeter) that prevented "Clostridium botulinum" growth. Early cultures also used honey or sugar as a preservatives. Greece used a quince and honey mixture with a slight amount of drying and then tightly packed into jars. The Romans used the same technique but instead cooked the honey and quince mixture to make a solid texture. Indian and Oriental traders brought sugarcane to the northern climates where housewives were then able to make preservatives by heating fruit with the sugarcane. Canning started in 1790 from a French confectioner, Nicolas Appert, when he found that by applying heat to food in sealed glass bottles, the food is free from spoilage. Appert's ideas were tried by the French Navy with meat, vegetables, fruit, and milk in 1806. An Englishman, Peter Durand decided to use Appert's method on tin cans in 1810. Even though Appert found a method that worked, he did not understand why it worked because many believed that the lack of air caused the preservation. In 1864 Louis Pasteur linked food spoilage/illness to microorganisms. Different foods are placed into jars or cans and heated to a microorganism and enzyme inactivating temperature. They are then cooled forming a vacuum seal which prevents microorganisms from contaminating the foods. Public awareness of food preservatives is uneven. Americans have a perception that food-borne illnesses happen more often in other countries. This may be true, but the occurrence of illnesses, hospitalizations, and deaths are still high. It is estimated by the Center for Disease Control (CDC) that each year there are 76 million illnesses, 325,000 hospitalizations, and 5,000 deaths linked to food-borne illness. The increasing demand for ready-to-eat fresh food products has led to challenges for food distributors regarding the safety and quality of their foods. Artificial preservatives meet some of these challenges by preserving freshness for longer periods of time, but these preservatives can cause negative side-effects as well. Sodium nitrite is a preservative used in lunch meats, hams, sausages, hot dogs, and bacon to prevent botulism. It serves the important function of controlling the bacteria that cause botulism, but sodium nitrite can react with proteins, or during cooking at high heats, to form carcinogenic N-nitrosamines. It has also been linked to cancer in lab animals. The commonly used sodium benzoate has been found to extend the shelf life of bottled tomato paste to 40 weeks without loss of quality. However, it can form the carcinogen benzene when combined with vitamin C. Many food manufacturers have reformed their products to eliminate this combination, but a risk still exists. Consumption of sodium benzoate may also cause hyperactivity. For over 30 years, there has been a debate about whether or not preservatives and other food additives can cause hyperactivity. Studies have found that there may be increases in hyperactivity amongst children who consume artificial colorings and benzoate preservatives and who are already genetically predisposed to hyperactivity, but these studies were not entirely conclusive. Hyperactivity only increased moderately, and it was not determined if the preservatives, colorings, or a combination of the two were responsible for the increase.
https://en.wikipedia.org/wiki?curid=24862
Proteobacteria Proteobacteria is a major phylum of Gram-negative bacteria. They include a wide variety of pathogenic genera, such as "Escherichia", "Salmonella", "Vibrio", "Helicobacter", "Yersinia", "Legionellales", and many others. Others are free-living (nonparasitic) and include many of the bacteria responsible for nitrogen fixation. Carl Woese established this grouping in 1987, calling it informally the "purple bacteria and their relatives". Because of the great diversity of forms found in this group, it was named after Proteus, a Greek god of the sea capable of assuming many different shapes and is not named after the Proteobacteria genus "Proteus". Some Alphaproteobacteria can grow at very low levels of nutrients and have unusual morphology such as stalks and buds. Others include agriculturally important bacteria capable of inducing nitrogen fixation in symbiosis with plants. The type order is the Caulobacterales, comprising stalk-forming bacteria such as "Caulobacter". The mitochondria of eukaryotes are thought to be descendants of an alphaproteobacterium. The Betaproteobacteria are highly metabolically diverse and contain chemolithoautotrophs, photoautotrophs, and generalist heterotrophs. The type order is the Burkholderiales, comprising an enormous range of metabolic diversity, including opportunistic pathogens. The Gammaproteobacteria are the largest class in terms of species with validly published names. The type order is the Pseudomonadales, which include the genera "Pseudomonas" and the nitrogen-fixing "Azotobacter". The Deltaproteobacteria include bacteria that are predators on other bacteria and are important contributors to the anaerobic side of the sulfur cycle. The type order is the Myxococcales, which includes organisms with self-organising abilities such as "Myxococcus" spp. The Epsilonproteobacteria are often slender, Gram-negative rods that are helical or curved. The type order is the Campylobacterales, which includes important food pathogens such as "Campylobacter" spp. The Zetaproteobacteria are iron-oxidizing neutrophilic chemolithoautotrophs, distributed worldwide in estuaries and marine habitats. The type order is the Mariprofundales. The Hydrogenophilalia are obligate thermophiles and include heterotrophs and autotrophs. The type order is the Hydrogenophilales. The Acidithiobacillia contain only sulfur, iron, and uranium-oxidising autotrophs. The type order is the Acidithiobacillales, which includes economically important organisms used in the mining industry such as "Acidithiobacillus" spp. The Oligoflexia are filamentous aerobes. The type order is the Oligoflexales, which contains the genus "Oligoflexus". All "Proteobacteria" are Gram-negative (though some may stain Gram-positive or Gram-variable in practice), with an outer membrane mainly composed of lipopolysaccharides. Many move about using flagella, but some are nonmotile or rely on bacterial gliding. The latter include the myxobacteria, an order of bacteria that can aggregate to form multicellular fruiting bodies. Also, a wide variety in the types of metabolism exists. Most members are facultatively or obligately anaerobic, chemolithoautotrophic, and heterotrophic, but numerous exceptions occur. A variety of genera, which are not closely related to each other, convert energy from light through photosynthesis. "Proteobacteria" are associated with the imbalance of microbiota of the lower reproductive tract of women. These species are associated with inflammation. The group is defined primarily in terms of ribosomal RNA (rRNA) sequences. The "Proteobacteria" are divided into nine classes with validly published names, referred to by the Greek letters alpha through zeta, the Acidithiobacillia, Hydrogenophilalia, and Oligoflexia. These were previously regarded as subclasses of the phylum, but they are now treated as classes. These classes are monophyletic. The genus "Acidithiobacillus", part of the Gammaproteobacteria until it was transferred to class Acidithiobacillia in 2013, was previously regarded as paraphyletic to the "Betaproteobacteria" according to multigenome alignment studies. In 2017, the Betaproteobacteria was subject to major revisions and the class Hydrogenophilalia was created to contain the order Hydrogenophilales Proteobacterial classes with validly published names include some prominent genera: e.g.: Transformation, a process in which genetic material passes from bacterium to another, has been reported in at least 30 species of "Proteobacteria" distributed in the classes alpha, beta, gamma and epsilon. The best-studied "Proteobacteria" with respect to natural genetic transformation are the medically important human pathogens "Neisseria gonorrhoeae" (class beta), "Haemophilus influenzae" (class gamma) and "Helicobacter pylori" (class epsilon). Natural genetic transformation is a sexual process involving DNA transfer from one bacterial cell to another through the intervening medium and the integration of the donor sequence into the recipient genome. In pathogenic "Proteobacteria", transformation appears to serve as a DNA repair process that protects the pathogen's DNA from attack by their host's phagocytic defenses that employ oxidative free radicals.
https://en.wikipedia.org/wiki?curid=24863
Professional wrestling Professional wrestling is a form of theater that combines mock combat and drama. It is a parody of prizefighting tournaments, with the stage being a ring similar to the kind used in boxing. The mock combat itself is based on classical and "catch" wrestling, with modern additions of striking attacks, strength-based holds and throws and acrobatic maneuvers, and occasionally improvised weaponry. Much like some of the real prizefighters they parody, the characters in professional wrestling have large egos, flamboyant personalities, and turbulent rivalries and friendships with each other, often exaggerated to comical degrees. In the modern age of televised wrestling, some dramatic scenes are recorded outside the ring, such as backstage or in other locations, to supplement the drama that happens in the ring. All the "matches" have predetermined outcomes. The performances are all planned (if not rigidly scripted) so as to maximize the entertainment value and reduce the chance of actual injury. In the past, the performers and producers would not admit that their "sport" was just theater, but nowadays they openly acknowledge it outside of official venues. During performances and in promotional material, they still maintain the traditional facade for the sake of verisimilitude. Originating as a popular form of entertainment in 19th-century Europe and later as a sideshow exhibition in North American traveling carnivals and vaudeville halls, professional wrestling grew into a standalone genre of entertainment with many diverse variations in cultures around the globe, and has become a billion-dollar entertainment industry. Since the 1980s, local forms have greatly declined in Europe; wrestling from North America has experienced several different periods of prominent cultural popularity during its century-and-a-half of existence and has been exported back to Europe to fill the cultural gap left by the aforementioned decline of local versions. The advent of television gave professional wrestling a new outlet, and wrestling (along with boxing) became instrumental in making pay-per-view a viable method of content delivery. In light of the growth of online video-on-demand, native professional wrestling promotions in markets all over the world have been able to circumvent traditional content-delivery and reach customers directly via social media and word-of-mouth marketing. Show wrestling has become especially prominent in Central/North America, Japan and Europe (especially the United Kingdom). In Brazil, there was a very popular wrestling television program from the 1960s to the early 1980s called "Telecatch". High-profile figures in the sport have become celebrities or cultural icons in their native or adopted home countries. Although professional wrestling started out as small acts in sideshows, traveling circuses and carnivals, today it is a billion-dollar industry. Revenue is drawn from ticket sales, network television broadcasts, pay-per-view broadcasts, branded merchandise and home video. Pro wrestling was instrumental in making pay-per-view a viable method of content delivery. Annual shows such as WrestleMania, Wrestle Kingdom and formerly Starrcade are among the highest-selling pay-per-view programming each year. In modern day, internet programming has been utilized by a number of companies to air web shows, internet pay per views (IPPVs) or on-demand content, helping to generate internet-related revenue earnings from the evolving World Wide Web. Home video sales dominate the Billboard charts Recreational Sports DVD sales, with wrestling holding anywhere from 3 to 9 of the top 10 spots every week. Due to its persistent cultural presence and to its novelty within the performing arts, wrestling constitutes a recurring topic in both academia and the media. Several documentaries have been produced looking at professional wrestling, most notably, "Beyond the Mat" directed by Barry W. Blaustein, and "" featuring wrestler Bret Hart and directed by Paul Jay. There have also been many fictional depictions of wrestling; the 2008 film "The Wrestler" received several Oscar nominations and began a career revival for star Mickey Rourke. Currently, the largest professional wrestling company worldwide is the United States-based WWE, which bought out many smaller regional companies in the late 20th century, as well as its primary US competitors World Championship Wrestling (WCW) and Extreme Championship Wrestling (ECW) in early 2001. Other prominent professional wrestling companies worldwide include the US-based All Elite Wrestling (AEW), Impact Wrestling, formerly known as Total Nonstop Action Wrestling (TNA), and Ring of Honor (ROH); Consejo Mundial de Lucha Libre (CMLL), and Lucha Libre AAA Worldwide (AAA) in Mexico; and the Japanese New Japan Pro-Wrestling (NJPW), All Japan Pro Wrestling (AJPW), and Pro Wrestling Noah leagues. When talking about professional wrestling, there are two levels: the "in-show" happenings that are presented through the shows, and happenings which are outside the scope of performance (in other words, are real life) but have implications on the performance, such as performer contracts, legitimate injuries, etc. Because actual events are often co-opted by writers for incorporation into storylines for the performers, the lines are often blurred and become confused. Special care must be taken when talking about people who perform under their own name. The actions of the character should be considered fictional events, wholly separate from the life of the performer. This is similar to other entertainers who perform with a persona that shares their own name (such as Kurt Angle and his fictional persona). Some wrestlers would incorporate elements of their real-life personalities into their characters, even if they and their in-ring persona have different names. Historians are unsure at what point wrestling changed from competitive catch wrestling into worked entertainment. However, documented accounts do exist: WWE Superstar Bret "Hitman" Hart recalls "a long and fascinating talk" he had in the summer of 1981 with the great Lou Thesz who told him that: Those who participated felt that maintenance of a constant and complete illusion for all who were not involved was necessary to keep audience interest. For decades, wrestlers lived their public lives as though they were their characters. The practice of keeping the illusion, and the various methods used to do so, came to be known as "kayfabe" within wrestling circles, or "working the marks". An entire lexicon of slang jargon and euphemism developed to allow performers to communicate without outsiders' knowledge of what was being said. Occasionally a performer will deviate from the intended sequence of events. This is known as a shoot. Sometimes shoot-like elements are included in wrestling stories to blur the line between performance and reality. These are known as "worked shoots". However, the vast majority of events in professional wrestling are preplanned and improvised within accepted boundaries. Gradually, the predetermined nature of professional wrestling became an open secret, as prominent figures in the wrestling business (including World Wrestling Entertainment owner Vince McMahon) began to publicly admit that wrestling was entertainment, not competition. This public reveal has garnered mixed reactions from the wrestling community, as some feel that exposure ruins the experience to the spectators as does exposure in illusionism. Despite the public admission of the theatrical nature of professional wrestling, many U.S. states still regulate professional wrestling as they do other professional competitive sports. For example, New York State still regulates "professional wrestling" through the New York State Athletic Commission (SAC). However, some states are considering removing, or have removed, professional wrestling from the purview of the state's athletic commissioners. Professional wrestling shows can be considered a form of theater in the round, with the ring, ringside area, and entryway comprising a stage. However, there is a much more limited concept of a fourth wall than in most theatric performances, similar to pantomime involving audience participation. The audience is recognized and acknowledged by the performers as spectators to the sporting event being portrayed, and are encouraged to interact as such. This leads to a high level of audience participation; in fact, their reactions can dictate how the performance unfolds. Often, individual matches will be part of a longer story line conflict between "babyfaces" (often shortened to just "faces") and "heels". "Faces" (the "good guys") are those whose actions are intended to encourage the audience to cheer, while "heels" (the "bad guys") act to draw the spectators' ire. There is no governing authority for professional wrestling rules, although there is a general standard which has developed. Each promotion has its own variation, but all are similar enough to avoid confusion most of the time. Any rule described here is simply a standard, and may or may not correspond exactly with any given promotion's ruleset. Due to the staged nature of wrestling, these are not actual "rules" in the sense that they would be considered in similar articles about actual sports like freestyle wrestling. Instead, the "rules" in this article are implemented and supposedly enforced for the sake of suspension of disbelief (known as kayfabe in the jargon of the business). Matches are held between two or more sides ("corners"). Each corner may consist of one wrestler, or a team of two or more. Most team matches are governed by tag team rules (see below). Other matches are free-for-alls, with multiple combatants but no teams. In all variants, there can be only one winning team or wrestler. Matches are held within a wrestling ring, an elevated square canvas mat with posts on each corner. A cloth apron hangs over the edges of the ring. Three horizontal ropes or cables surround the ring, suspended with turnbuckles which are connected to the posts. For safety, the ropes are padded at the turnbuckles and cushioned mats surround the floor outside the ring. Guardrails or a similar barrier enclose this area from the audience. Wrestlers are generally expected to stay within the confines of the ring, though matches sometimes end up outside the ring, and even in the audience, to add excitement. The standard method of scoring is the "fall", which is accomplished by: These are each explained in greater detail below. Typically, pinfalls and submissions must occur within the ring area, however there are times where it may be stipulated otherwise. Most wrestling matches last for a set number of falls, with the first side to achieve the majority number of pinfalls, submissions, or countouts being the winner. Historically, matches were wrestled to 3 falls ("best 2 out of 3") or 5 falls ("best 3 out of 5"). The standard for modern matches is one fall. However, even though it is now standard, many announcers will explicitly say so, e.g. "The following contest is set for one fall with a 20-minute time limit." These matches are given a time limit; if not enough falls are scored by the end of the time limit, the match is declared a draw. Modern matches are generally given a 10- to 30-minute time limit for standard matches; title matches can go for up to one hour. British wrestling matches held under Admiral-Lord Mountevans rules are 2 out of 3 falls. An alternative is a match set for a prescribed length of time, with a running tally of falls. The entrant with the most falls at the end of the time limit is declared the winner. This is usually for 20, 30 or 60 minutes, and is commonly called an Iron Man match. This type of match can be modified so that fewer types of falls are allowed. In matches with multiple competitors, an elimination system may be used. Any wrestler who has a fall scored against them is forced out of the match, and the match continues until only one remains. However, it is much more common when more than two wrestlers are involved to simply go one fall, with the one scoring the fall, regardless of who they scored it against, being the winner. In championship matches, this means that, unlike one-on-one matches (where the champion can simply disqualify himself or get himself counted out to retain the title via the Champion's Advantage), the champion does "not" have to be pinned or involved in the decision to lose the championship. However, heel champions often find advantages, not in Champion's Advantage, but in the use of weapons and outside interference, as these poly-sided matches tend to involve no holds barred rules. Many modern specialty matches have been devised, with unique winning conditions. The most common of these is the ladder match. In the basic ladder match, the wrestlers or teams of wrestlers must climb a ladder to obtain a prize that is hoisted above the ring. The key to winning this match is that the wrestler or team of wrestlers must try to incapacitate each other long enough for one wrestler to climb the ladder and secure that prize for their team. As a result, the ladder can be used as a weapon. The prizes include but are not limited to any given championship belt (the traditional prize), a document granting the winner the right to a future title shot, or any document that matters to the wrestlers involved in the match (such as one granting the winner a cash prize). Another common specialty match is known as the battle royal. In a battle royal, all the wrestlers enter the ring to the point that there are 20–30 wrestlers in the ring at one time. When the match begins, the simple objective is to throw the opponent over the top rope and out of the ring with both feet on the floor to eliminate that opponent. The last wrestler standing is declared the winner. A variant on this type of match is the WWE's Royal Rumble where two wrestlers enter the ring to start the match and other wrestlers follow in 90 second intervals (previously 2 minutes) until 30–40 wrestlers have entered the ring. All other rules stay the same. For more match types, see Professional wrestling match types. Every match must be assigned a rule keeper known as a referee, who is the final arbitrator. In multi-man lucha libre matches, two referees are used, one inside the ring and one outside. Due to the legitimate role that referees play in wrestling of serving as liaison between the bookers backstage and the wrestlers in the ring (the role of being a final arbitrator is merely kayfabe), the referee is present, even in matches that do not at first glance appear to require a referee (such as a ladder match, as it is no holds barred, and the criteria for victory could theoretically be assessed from afar). Although their actions are also frequently scripted for dramatic effect, referees are subject to certain general rules and requirements to maintain the theatrical appearance of unbiased authority. The most basic rule is that an action must be seen by a referee to be declared for a fall or disqualification. This allows for heel characters to gain a scripted advantage by distracting or disabling the referee to perform some ostensibly illegal maneuver on their opponent. Most referees are unnamed and essentially anonymous, though some wrestling promotions, most notably in the present All Elite Wrestling, have made officials known by their names (and there are some cases where fans have called their name during matches). Special guest referees may be used from time to time; by virtue of their celebrity status, they are often scripted to dispense with the appearance of neutrality and use their influence to unfairly influence the outcome of the match for added dramatic impact. Face special referees will often fight back against hostile heel wrestlers, particularly if the special referee is either a wrestler himself or a famous martial artist (such as Tito Ortiz at the main event at TNA Hard Justice 2005). For heel special referees, common ways of assisting the heel wrestler to obtain victory include, but are not limited to, the following: In some team matches, only one entrant from each team may be designated as the "legal" or "active" wrestler at any given moment. Two wrestlers must make physical contact (typically palm-to-palm) to transfer this legal status. This is known as a "tag", with the participants "tagging out" and "tagging in". Typically the wrestler who is tagging out has a 5-second count to leave the ring, whereas the one tagging in can enter the ring at any time, resulting in heels legally double-teaming a face. The non-legal wrestlers must remain outside the ring or other legal area at all times (and avoid purposeful contact with the opposing wrestlers) or face reprimand from the referee. In most promotions, the wrestler to be tagged in must be touching the turnbuckle on his corner, or a cloth strap attached to the turnbuckle. Some multi-wrestler matches allow for a set number of legal wrestlers, and a legal wrestler may tag out to any other wrestler, regardless of team. In these matches, the tag need not be a mutual effort, and this results in active wrestlers being tagged out against their will, or non-legal wrestlers forced to enter the battle. Sometimes, poly-sided matches that pit every man for himself will incorporate tagging rules. Outside of kayfabe, this is done to give wrestlers a break from the action (as these matches tend to go on for long periods of time), and to make the action in the ring easier to choreograph. One of the most mainstream examples of this is the Four-Corner match, the most common type of match in the WWE before it was replaced with its equivalent Fatal Four-Way; four wrestlers, each for himself, fight in a match, but only two wrestlers can be in the match at any given time. The other two are positioned in the corner, and tags can be made between any two wrestlers. In a Texas Tornado Tag Team match, all the competitors are legal in the match, and tagging in and out is not necessary. All matches fought under hardcore rules (such as no disqualification, no holds barred, ladder match, etc.) are all contested under "de facto" Texas Tornado rules, since the lack of ability of a referee to issue a disqualification renders any tagging requirements moot. Regardless of rules of tagging, a wrestler cannot pin his or her own tag team partner, even if it is technically possible from the rules of the match (e.g. Texas Tornado rules, or a three-way tag team match). This is called the "Outlaw Rule" because the first team to attempt to use that (in an attempt to unfairly retain their tag team titles) was the New Age Outlaws. To score by pinfall, a wrestler must pin both his opponent's shoulders against the mat while the referee slaps the mat three times (referred to as a "three count"). This is the most common form of defeat. The pinned wrestler must also be on his back and, if s/he is lying on his stomach, it usually does not count. A count may be started at any time that a wrestler's shoulders are down (both shoulders touching the mat), back-first and any part of the opponent's body is lying over the wrestler. This often results in pins that can easily be kicked out of, if the defensive wrestler is even slightly conscious. For example, an attacking wrestler who is half-conscious may simply drape an arm over an opponent, or a cocky wrestler may place his foot gently on the opponent's body, prompting a three-count from the referee. Illegal pinning methods include using the ropes for leverage and hooking the opponent's clothing, which are therefore popular cheating methods for heels, unless certain stipulations make such an advantage legal. Such pins as these are rarely seen by the referee and are subsequently often used by heels and on occasion by cheating faces to win matches. Even if it is noticed, it is rare for such an attempt to result in a disqualification (see below) and instead it simply results in nullification of the pin attempt, so the heel wrestler rarely has anything to lose for trying it anyway. Occasionally, there are instances where a pinfall is made where both wrestlers' shoulders were on the mat for the three-count. This situation will most likely lead to a draw, and in some cases a continuation of the match or a future match to determine the winner. To score by submission, the wrestler must make his opponent give up, usually, but not necessarily, by putting him in a submission hold (e.g. figure four leg-lock, arm-lock, sleeper-hold). A wrestler may voluntarily submit by verbally informing the referee (usually used in moves such as the Mexican Surfboard, where all four limbs are incapacitated, making tapping impossible). Also, since Ken Shamrock popularized it in 1997, a wrestler can indicate a voluntary submission by "tapping out", that is, tapping a free hand against the mat or against an opponent. Occasionally, a wrestler will reach for a rope (see rope breaks below), only to put his hand back on the mat so he can crawl towards the rope some more; this is not a submission, and the referee decides what his intent is. Submission was initially a large factor in professional wrestling, but following the decline of the submission-oriented catch-as-catch-can style from mainstream professional wrestling, the submission largely faded. Despite this, some wrestlers, such as Chris Jericho, Ric Flair, Bret Hart, Kurt Angle, Ken Shamrock, Dean Malenko, Chris Benoit, and Tazz, became famous for winning matches via submission. A wrestler with a signature submission technique is portrayed as better at applying the hold, making it more painful or more difficult to get out of than others who use it, or can be falsely credited as inventing the hold (such as when Tazz popularized the kata ha jime judo choke in pro wrestling as the "Tazzmission"). Since all contact between the wrestlers must cease if any part of the body is touching, or underneath, the ropes, many wrestlers will attempt to break submission holds by deliberately grabbing the bottom ropes. This is called a "rope break", and it is one of the most common ways to break a submission hold. Most holds leave an arm or leg free, so that the person can tap out if he wants. Instead, he uses these free limbs to either grab one of the ring ropes (the bottom one is the most common, as it is nearest the wrestlers, though other ropes sometimes are used for standing holds such as Chris Masters' Master Lock) or drape his foot across, or underneath one. Once this has been accomplished, and the accomplishment is witnessed by the referee, the referee will demand that the offending wrestler break the hold, and start counting to five if the wrestler does not. If the referee reaches the count of five, and the wrestler still does not break the hold, he is disqualified. If a manager decides that his client wrestler should tap out, but cannot convince the wrestler himself to do so, he may "throw in the towel" (by literally taking a gym towel and hurling it into the ring where the referee can see it). This is the same as a submission, as in kayfabe the manager is considered the wrestlers agent and therefore authorized to make formal decisions (such as forfeiting a match) on the client's behalf. Passing out in a submission hold constitutes a loss by knockout. To determine if a wrestler has passed out in WWE, the referee usually picks up and drops his hand. If it drops to the mat or floor three consecutive times without the wrestler having the strength to hold it up, the wrestler is considered to have passed out. At one point this was largely ignored. However, the rule is now much more commonly observed for safety reasons. If the wrestler has passed out, the opponent then scores by submission. A wrestler can also win by knockout if he does not resort to submission holds, but stills pummels his opponent to the point that he is completely out cold. To check for a knockout in this manner a referee would wave his hand in front of the wrestlers' face and, if the wrestler does not react in any way, the referee would award the victory to the other wrestler. A countout (alternatively "count-out" or "count out") happens when a wrestler is out of the ring long enough for the referee to count to ten (twenty in some promotions) and thus disqualified. The count is broken and restarted when a wrestler in the ring exits the ring. Playing into this, some wrestlers would "milk" the count by sliding in the ring and immediately sliding back out. As he was technically inside the ring for a split second before exiting again, it is sufficient to restart the count. This is often referred to by commentators as "breaking the count". Heels often use this tactic in order to buy themselves more time to catch their breath, or to attempt to frustrate their babyface opponents. If all the active wrestlers in a match are down inside the ring at the same time, the referee would begin a count (usually ten seconds, twenty in Japan). If nobody rises to their feet by the end of the count, the match is ruled a draw. Any participant who stands up in time would end the count for everyone else, while in a Last Man Standing match this form of a countout is the only way that the match can end, so the referee would count when one or more wrestlers are down and one wrestler standing up before the 10-count does not stop the count for another wrestler who is still down. In some promotions (and most major modern ones), Championships cannot change hands via a countout, unless the on-screen authority declares it for at least one match, although in others, championships may change hands via countout. Heels are known to take advantage of this and will intentionally get counted out when facing difficult opponents, especially when defending championships. Disqualification (sometimes abbreviated as "DQ") occurs when a wrestler violates the match's rules, thus losing automatically. Although a countout can technically be considered a disqualification (as it is, for all intents and purposes, an automatic loss suffered as a result of violating a match rule), the two concepts are often distinct in wrestling. A no disqualification match can still end by countout (although this is rare). Typically, a match must be declared a "no holds barred" match, a "street fight" or some other term, in order for both disqualifications and countouts to be waived. Disqualification from a match is called for a number of reasons: In practice, not all rule violations will result in a disqualification as the referee may use his own judgement and is not obligated to stop the match. Usually, the only offenses that the referee will see and "immediately" disqualify the match on (as opposed to having multiple offenses) are low blows, weapon usage, interference, or assaulting the referee. In WWE, a referee must see the violation with his own eyes to rule that the match end in a disqualification (simply watching the video tape is not usually enough) and the referee's ruling is almost always final, although "dusty finishes" (named after, and made famous by, Dusty Rhodes) will often result in the referee's decision being overturned. It is not uncommon for the referees themselves to get knocked out during a match, which is commonly referred to by the term "ref bump". While the referee remains "unconscious", wrestlers are free to violate rules until he is revived or replaced. In some cases, a referee might disqualify a person under the presumption that it was that wrestler who knocked him out; most referee knockouts are arranged to allow a wrestler, usually a heel, to gain an advantage. For example, a wrestler may get whipped into a referee at a slower speed, knocking the ref down for short amount of time; during that interim period, one wrestler may pin his opponent for a three-count and would have won the match but for the referee being down (sometimes, another referee will sprint to the ring from backstage to attempt to make the count, but by then, the other wrestler has had enough time to kick out on his own accord). In most promotions, a championship title cannot normally change hands via disqualification; this rule is explicitly enforced in a title match under special circumstances. If all participants in a match continue to breach the referee's instructions, the match may end in a double disqualification, where both wrestlers or teams (in a tag team match) have been disqualified. The match is essentially nullified, and called a draw or in some cases a restart or the same match being held at a pay-per-view or next night's show. Sometimes, however, if this happens in a match to determine the challenger for a heel champion's title, the champion is forced to face both opponents simultaneously for the title. Usually, the double disqualification is caused by the heel wrestler's associates in a match between two face wrestlers to determine his opponent. Although extremely rare, a match can end in a forfeit if the opponent either does not show up for the match, or shows up but refuses to compete. Although a championship usually cannot change hands except by pinfall or submission, a forfeit victory is enough to crown a new champion. The most famous example of this happened on the December 8, 1997 episode of "Raw is War", when Stone Cold Steve Austin handed the WWE Intercontinental Championship to The Rock after refusing to defend the title. Forfeit victories are extremely rare in wrestling. When a pay-per-view match is booked and one wrestler is unable to make it for one reason or another, it is usually customary to insert a last minute replacement rather than award a wrestler a victory by forfeit. Forfeit victories are almost always reserved for when the story the promotion is telling specifically requires such an ending. In addition to the aforementioned moment between Steve Austin and The Rock, other instance of this happening was in March 2009, when Triple H decided not to show up for a handicap match against Cody Rhodes and Ted Dibiase Jr., instead opting to attack Randy Orton in his own home. Despite being, statistically, an extremely rare occurrence, Charles Wright is one wrestler who is famous for turning forfeit victories into his own gimmick. During the late 1990s, Wright called himself "The Godfather" and portrayed the gimmick of a pimp. He would often bring multiple women, who he referred to as "hos," to the ring with him, and would offer the sexual services of these women to his opponents in exchange for them forfeiting their matches against him. A professional wrestling match can end in a draw. A draw occurs if both opponents are simultaneously disqualified (as via countout or if the referee loses complete control of the match and both opponents attack each other with no regard to being in a match, like Brock Lesnar vs. Undertaker at 2002 Unforgiven), neither opponent is able to answer a ten-count, or both opponents simultaneously win the match. The latter can occur if, for example, one opponent's shoulders touch the mat while maintaining a submission hold against another opponent. If the opponent in the hold begins to tap out at the same time a referee counts to three for pinning the opponent delivering the hold, both opponents have legally achieved scoring conditions simultaneously. Traditionally, a championship may not change hands in the event of a draw (though it may become vacant), though some promotions such as Total Nonstop Action Wrestling have endorsed rules where the champion may lose a title by disqualification. A variant of the draw is the time-limit draw, where the match does not have a winner by a specified time period (a one-hour draw, which was once common, is known in wrestling circles as a "Broadway"). Also if two wrestlers have been given a disqualification by either the referee or the chairman, this is a no contest and if there is a title on the line the champion keeps the championship. A wrestling match may be declared a no contest if the winning conditions are unable to occur. This can be due to excessive interference, loss of referee's control over the match, one or more participants sustaining debilitating injury not caused by the opponent, or the inability of a scheduled match to even begin. A no contest is a state separate and distinct from a draw — a draw indicates winning conditions were met. Although the terms are sometimes used interchangeably in practice, this usage is technically incorrect. While each wrestling match is ostensibly a competition of athletics and strategy, the goal from a business standpoint is to excite and entertain the audience. Although the competition is staged, dramatic emphasis draws out the most intense reaction. Heightened interest results in higher attendance, increased ticket sales, higher ratings on television broadcasts (greater ad revenue), higher pay-per-view buyrates, and sales of branded merchandise and recorded video footage. All of these contribute to the profit of the promotion company. In Latin America and English-speaking countries, most wrestlers (and other on-stage performers) portray character roles, sometimes with personalities wildly different from their own. These personalities are a gimmick intended to heighten interest in a wrestler without regard to athletic ability. Some can be unrealistic and cartoon-like (such as Doink the Clown), while others carry more verisimilitude (such as Chris Jericho, The Rock, John Cena, Steve Austin, and CM Punk). In lucha libre, many characters wear masks, adopting a secret identity akin to a superhero, a near-sacred tradition. An individual wrestler may use his real name, or a minor variation of it, for much of his career, such as Bret Hart, John Cena and Randy Orton. Others can keep one ring name for their entire career (Shawn Michaels, CM Punk and Ricky Steamboat), or may change from time to time to better suit the demands of the audience or company. Sometimes a character is owned and trademarked by the company, forcing the wrestler to find a new one when he leaves (although a simple typeset change, such as changing Rhyno to Rhino, can get around this), and sometimes a character is owned by the wrestler. Sometimes, a wrestler may change his legal name to obtain ownership of his ring name (Andrew Martin and Warrior). Many wrestlers (such as The Rock and The Undertaker) are strongly identified with their character, even responding to the name in public or between friends. It's actually considered proper decorum for fellow wrestlers to refer to each other by their stage names/characters rather than their birth/legal names, unless otherwise introduced. A character can become so popular that it appears in other media (Hulk Hogan and El Santo) or even gives the performer enough visibility to enter politics (Antonio Inoki and Jesse Ventura). Typically, matches are staged between a protagonist (historically an audience favorite, known as a babyface, or "the good guy") and an antagonist (historically a villain with arrogance, a tendency to break rules, or other unlikable qualities, called a heel). In recent years, however, antiheroes have also become prominent in professional wrestling. There is also a less common role of a "tweener", who is neither fully face nor fully heel yet able to play either role effectively (case in point, Samoa Joe during his first run in TNA Wrestling from June 2005 to November 2006). At times, a character may "turn", altering their face/heel alignment. This may be an abrupt, surprising event, or it may slowly build over time. It is almost always accomplished with a markable change in behavior. Some turns become defining points in a career, as when Hulk Hogan turned heel after being a top face for over a decade. Others may have no noticeable effect on the character's status. If a character repeatedly switches between face and heel, this lessens the effect of such turns, and may result in apathy from the audience. Vince McMahon is a good example of having more heel and face turns than anyone in WWE history. As with personae in general, a character's face or heel alignment may change with time, or remain constant over its lifetime (the most famous example of the latter is Ricky Steamboat, a WWE Hall of Famer who remained a babyface throughout his entire career). Sometimes a character's heel turn will become so popular that eventually the audience response will alter the character's heel-face cycle to the point where the heel persona will, in practice, become a face persona, and what was previously the face persona, will turn into the heel persona, such as when Dwayne Johnson first began using "The Rock" persona as a heel character, as opposed to his original "Rocky Maivia" babyface persona. Another legendary example is Stone Cold Steve Austin, who was originally booked as a heel, with such mannerisms as drinking on the job, using profanity, breaking company property, and even breaking into people's private homes. However, much to WWF's surprise, the fans got such a charge out of Austin's antics that he effectively became one of the greatest antiheroes in the history of the business. He, along with the stable of D-Generation X, is generally credited with ushering in the Attitude Era of WWF programming. While real exhibition matches are now not uncommon, most matches tell a story analogous to an episode of a serial drama: The face will from time to time win (triumph) or from time to time lose (tragedy), and longer story arcs can result from a couple of matches. Since most promotions have a championship title, opposition for the championship is a frequent impetus for stories. Also, whatever from a character's own hair to his job with the advertising can be wagered in a match. Some matches are designed to further the story of only one participant. It could be intended to portray an unstoppable force, a lucky underdog, a sore loser, or any other characterization. Sometimes non-wrestling vignettes are shown to enhance a character's image without the need for matches. Other stories result from a natural rivalry. Outside of performance, these are referred to as feuds. A feud can exist between any number of participants and can last from a few days to decades. The feud between Ric Flair and Ricky Steamboat lasted from the late 1970s into the early 1990s and allegedly spanned over two thousand matches (although most of those matches were mere dark matches). The career-spanning history between characters Mike Awesome and Masato Tanaka is another example of a long-running feud, as is the case of Steve Austin vs. Vince McMahon, one of the most lucrative feuds in the World Wrestling Federation during 1998 and 1999. In theory, the longer a feud is built up, the more audience interest (aka heat) lasts. The main event of a wrestling show is generally the most heated. Commonly, a heel will hold the upper hand over a face until a final showdown, heightening dramatic tension as the face's fans desire to see him win. Throughout the history of professional wrestling, many other elements of media have been utilized in professional wrestling storytelling: pre- and post-match interviews, "backstage" skits, positions of authority and worked behind-the-scenes feuds, division rankings (typically the #1-contendership spot), contracts, lotteries, news stories on websites, and in recent years social media. Also, anything that can be used as an element of drama can exist in professional wrestling stories: romantic relationships (including love triangles and marriage), racism, classism, nepotism, favoritism, corporate corruption, family bonds, personal histories, grudges, theft, cheating, assault, betrayal, bribery, seduction, stalking, confidence tricks, extortion, blackmail, substance abuse, self-doubt, self-sacrifice; even kidnapping, sexual fetishism, necrophilia, misogyny, rape and death have been portrayed in wrestling. Some promotions have included supernatural elements such as magic, curses, the undead and Satanic imagery (most notably the Undertaker and his Ministry of Darkness, a stable that regularly performed evil rituals and human sacrifice in Satanic-like worship of a hidden power figure). Celebrities would also be involved in storylines. Commentators have become important in communicating the relevance of the characters' actions to the story at hand, filling in past details and pointing out subtle actions that may otherwise go unnoticed. A main part of the story-telling part of wrestling is a promo, short for promotional interview. Promos are performed, or "cut", in wrestling jargon, for a variety of reasons, including to heighten interest in a wrestler, or to hype an upcoming match. Since the crowd is often too loud or the venue too large for promos to be heard naturally, wrestlers will use amplification when speaking in the ring. Unlike most Hollywood acting, large and highly visible handheld microphones are typically used and wrestlers often speak directly to the audience. Professional wrestling mimics the structure of title match combat sports. Participants compete for a championship and must defend it after winning it. These titles are represented physically by a title belt that can be worn by the champion. In the case of team wrestling, there is a title belt for each member of the team. Almost all professional wrestling promotions have one major title, and some have more. Championships are designated by divisions of weight, height, gender, wrestling style and other qualifications. Typically, each promotion only recognizes the "legitimacy" of their own titles, although cross-promotion does happen. When one promotion absorbs or purchases another, the titles from the defunct promotion may continue to be defended in the new promotion or be decommissioned. Behind the scenes, the bookers in a company will place the title on the most accomplished performer, or those the bookers believe will generate fan interest in terms of event attendance and television viewership. Lower ranked titles may also be used on the performers who show potential, thus allowing them greater exposure to the audience. However other circumstances may also determine the use of a championship. A combination of a championship's lineage, the caliber of performers as champion, and the frequency and manner of title changes, dictates the audience's perception of the title's quality, significance and reputation. A wrestler's championship accomplishments can be central to their career, becoming a measure of their performance ability and drawing power. In general, a wrestler with multiple title reigns or an extended title reign is indicative of a wrestler's ability to maintain audience interest or a wrestler's ability to perform in the ring. As such, the most accomplished or decorated wrestlers tend to be revered as legends due to the amount of title reigns they hold. American wrestler Ric Flair has had multiple world heavyweight championship reigns spanning over three decades. Japanese wrestler Último Dragón once held and defended a record 10 titles simultaneously. Often a match will take place under additional rules, usually serving as a special attraction or a climactic point in a feud or storyline. Sometimes this will be the culmination of an entire feud, ending it for the immediate future (known as a blowoff match). Perhaps the most well-known non-standard match is the cage match, in which the ring is surrounded by a fence or similar metal structure, with the express intention of preventing escape or outside interference—and with the added bonus of the cage being a potentially brutal weapon or platform for launching attacks. The WWE has another provision where a standard cage match can end with one wrestler or wrestling team escaping the cage through the door or over the top. Another example is the WWE's Royal Rumble match, which involves thirty participants in a random and unknown order. The Rumble match is itself a spectacle in that it is a once-yearly event with multiple participants, including individuals who might not interact otherwise. It also serves as a catalyst for the company's ongoing feuds, as well as a springboard for new storylines. While the wrestling matches themselves are the primary focus of professional wrestling, a key dramatic element of the business can be entrances of the wrestlers to the arena and ring. It is typical for a wrestler to get their biggest crowd reaction (or "pop") for their ring entrance, rather than for anything they do in the wrestling match itself, especially if former main event stars are returning to a promotion after a long absence. All notable wrestlers now enter the ring accompanied by music, and regularly add other elements to their entrance. The music played during the ring entrance will usually mirror the wrestler's personality. Many wrestlers, particularly in America, have music and lyrics specially written for their ring entrance. While invented long before, the practice of including music with the entrance gained rapid popularity during the 1980s, largely as a result of the huge success of Hulk Hogan and the WWF, and their Rock 'n' Wrestling Connection. When a match is won, the victor's theme music is usually also played in celebration. Because wrestling is predetermined, a wrestler's entrance music will play as they enter the arena, even if they are, in kayfabe, not supposed to be there. For example, in 2012 through 2014, The Shield was a trio of wrestlers who were (in kayfabe) not at the time under contract with WWE (hence their gimmick of entering the ring through the crowd), but they still had entrance music which was played whenever they entered the arena, despite the fact that they were kayfabe invaders. With the introduction of the Titantron entrance screen in 1997, WWF/WWE wrestlers also had entrance videos made that would play along with their entrance music. Other dramatic elements of a ring entrance can include: Another method of entry involves descending from the ceiling with a Zip-line or rappel line and stunt harness. This has been done by Shawn Michaels at WrestleMania XII, by Sting many times in WCW and TNA, gained major controversy over its role in the death of wrestler Owen Hart at Over the Edge. Special ring entrances are also developed for big occasions, most notably the WrestleMania event. For example, WrestleMania III and VI both saw all wrestlers enter the arena on motorized miniature wrestling rings. Live bands are sometimes hired to perform live entrance music at special events. John Cena and Triple H are particularly notable in recent years for their highly theatrical entrances at WrestleMania. The women's division of professional wrestling has maintained a recognized world champion since 1937, when Mildred Burke won the original World Women's title. She then formed the World Women's Wrestling Association in the early 1950s and recognized herself as the first champion, although the championship would be vacated upon her retirement in 1956. The NWA however, ceased to acknowledge Burke as "their" Women's World champion in 1954, and instead acknowledged June Byers as champion after a controversial finish to a high-profile match between Burke and Byers that year. Upon Byers' retirement in 1964, The Fabulous Moolah, who won a junior heavyweight version of the NWA World Women's Championship (the predecessor to the WWE Women's Championship) in a tournament back in 1958, was recognized by most NWA promoters as champion by default. For most of its history, men and women would rarely compete against each other in professional wrestling, as it was deemed to be unfair and unchivalrous. Andy Kaufman used this to gain notoriety when he created an Intergender Championship and declared it open to any female challenger. This led to a long (worked) feud with Jerry Lawler. In the 1980s, mixed tag team matches began to take place, with a male and female on each team and a rule stating that each wrestler could only attack the opponent of the same gender. If a tag was made, the other team had to automatically switch their legal wrestler as well. Despite these restrictions, many mixed tag matches do feature some physical interaction between participants of different genders. For example, a heel may take a cheap shot at the female wrestler of the opposing team to draw a negative crowd reaction. In lucha libre, cheap-shots and male-female attacks are not uncommon. Intergender singles bouts were first fought on a national level in the 1990s. This began with Luna Vachon, who faced men in ECW and WWF. Later, Chyna became the first female to hold a belt that was not exclusive to women when she won the WWF Intercontinental Championship. While it is a rare feat in WWE, in TNA, ODB participates in singles intergender matches. Also, ODB's kayfabe husband and tag team partner Eric Young held the Knockouts tag team titles for a record 478 days before it was stripped by Brooke Hogan because Young was a male. Midget wrestling can be traced to professional wrestling's carnival and vaudeville origins. In recent years, the popularity and prevalence of midgets in wrestling has greatly decreased due to wrestling companies depriving midget divisions of storyline or feud. However, WWE has made a few attempts to enter this market with their "minis" in the 1990s and the "junior's league" as recent as 2006. It is still a popular form of entertainment in Mexican wrestling, mostly as a "sideshow". Some wrestlers may have their own specific "mini me", like Mascarita Sagrada, Alebrije has Quije, etc. There are also cases in which midgets can become valets for a wrestler, and even get physically involved in matches, like Alushe, who often accompanies Tinieblas, or KeMonito, who is portrayed as Consejo Mundial de Lucha Libre's mascot and is also a valet for Mistico. Dave Finlay was often aided in his matches by a midget known mainly as Hornswoggle while in WWE, who hid under the ring and gave a shillelagh to Finlay to use on his opponent. Finlay also occasionally threw him at his opponent(s). Hornswoggle has also been given a run with the WWE Cruiserweight Championship and feuded with D-X in 2009. The U.S., Japan and Mexico are three countries where there is a huge market and high popularity for professional wrestling. But the styles of professional wrestling are different, given their independent development for a long period. Professional wrestling in the U.S. tends to have a heavy focus on story building and the establishment of characters (and their personalities). There is a story for each match, and even a longer story for successive matches. The stories usually contain characters like faces and heels, and less often antiheroes and tweeners. It is a "triumph" if the face wins, while it is a "tragedy" if the heel wins. The characters usually have strong and sharp personalities. The opposition between faces and heels is very intense in the story, and the heels may even attack the faces during TV interviews. The relationship between different characters can also be very complex. Although professional wrestling in Mexico (Lucha libre) also has stories and characters, they are less emphasized. Wrestlers in Mexico are traditionally more agile and perform more aerial maneuvers than professional wrestlers in the U.S. who, more often, rely on power moves and strikes to subdue their opponents. The difference in styles is due to the independent evolution of the sport in Mexico beginning in the 1930s and the fact that wrestlers in the cruiserweight division () are often the most popular wrestlers in Mexican lucha libre. Wrestlers often execute high flying moves characteristic of lucha libre by utilizing the wrestling ring's ropes to catapult themselves towards their opponents, using intricate combinations in rapid-fire succession, and applying complex submission holds. Lucha libre is also known for its tag team wrestling matches, in which the teams are often made up of three members, instead of two as is common in the U.S. The style of Japanese professional wrestling (puroresu) is again different. With its origins in traditional American style of wrestling and still being under the same genre, it has become an entity in itself. Despite the similarity to its American counterpart in that the outcome of the matches remains predetermined, the phenomena are different in the form of the psychology and presentation of the sport. In most of the largest promotions, such as New Japan Pro-Wrestling, All Japan Pro Wrestling and Pro Wrestling Noah, it is treated as a full contact combat sport as it mixes hard hitting martial arts strikes with shoot style submission holds, while in the U.S. it is rather more regarded as an entertainment show. Wrestlers incorporate kicks and strikes from martial arts disciplines, and a strong emphasis is placed on submission wrestling, and unlike the use of involved storylines in the U.S., they are not as intricate in Japan, more emphasis is placed on the concept of "fighting spirit", meaning the wrestlers display of physical and mental stamina are valued a lot more than theatrics. Many of Japan's wrestlers including top stars such as Shinya Hashimoto, Riki Chōshū and Keiji Mutoh came from a legitimate martial arts background and many Japanese wrestlers in the 1990s began to pursue careers in mixed martial arts organizations such as Pancrase and Shooto which at the time retained the original look of puroresu but were actual competitions. Other companies, such as Michinoku Pro Wrestling and Dragon Gate, wrestle in a style similar to Mexican companies like AAA and CMLL. This is known as "Lucharesu". Professional wrestling has developed its own cultures, both internal and external. Those involved in producing professional wrestling have developed a kind of global fraternity, with familial bonds, shared language and passed-down traditions. New performers are expected to "pay their dues" for a few years by working in lower-profile promotions and working as ring crew before working their way upward. The permanent rosters of most promotions develop a backstage pecking order, with veterans mediating conflicts and mentoring younger wrestlers. For many decades (and still to a lesser extent today) performers were expected to keep the illusions of wrestling's legitimacy alive even while not performing, essentially acting in character any time they were in public. Some veterans speak of a "sickness" among wrestling performers, an inexplicable pull to remain active in the wrestling world despite the devastating effects the job can have on one's life and health. Fans of professional wrestling have their own subculture, comparable to those of science fiction, video games, or comic books. Those who are interested in the backstage occurrences, future storylines and reasonings behind company decisions read newsletters written by journalists with inside ties to the wrestling industry. These "rags" or "dirt sheets" have expanded into the Internet, where their information can be dispensed on an up-to-the-minute basis. Some have expanded into radio shows. Some fans enjoy a pastime of collecting tapes of wrestling shows from specific companies, of certain wrestlers, or of specific genres. The internet has given fans exposure to worldwide variations of wrestling they would be unable to see otherwise. Since the 1990s, many companies have been founded which deal primarily in wrestling footage. When the WWE purchased both WCW and ECW in 2001, they also obtained the entire past video libraries of both productions and have released many past matches online and on home video. Like some other sports, fantasy leagues have developed around professional wrestling. Some take this concept further by creating E-feds (electronic federations), where a user can create their own fictional wrestling character, and role-playing storylines with other users, leading to scheduled "shows" where match results are determined by the organizers, usually based on a combination of the characters' statistics and the players' roleplaying aptitude, sometimes with audience voting. From the first established world championship, the top professional wrestlers have garnered fame within mainstream society. Each successive generation has produced a number of wrestlers who extend their careers into the realms of music, acting, writing, business, politics or public speaking, and are known to those who are unfamiliar with wrestling in general. Conversely, celebrities from other sports or general pop culture also become involved with wrestling for brief periods of time. A prime example of this is The Rock 'n' Wrestling Connection of the 1980s, which combined wrestling with MTV. Professional wrestling is often portrayed within other works using parody, and its general elements have become familiar tropes and memes in American culture. Some terminology originating in professional wrestling has found its way into the common vernacular. Phrases such as "body slam", "sleeper hold" and "tag team" are used by those who do not follow professional wrestling. The term "smackdown", popularized by The Rock and "SmackDown!" in the 1990s, has been included in Merriam-Webster dictionaries since 2007. Many television shows and films have been produced which portray in-character professional wrestlers as protagonists, such as "Ready to Rumble", "¡Mucha Lucha!", "Nacho Libre", and the Santo film series. At least two stage plays set in the world of pro wrestling have been produced: "The Baron" is a comedy that retells the life of an actual performer known as Baron von Raschke. "From Parts Unknown..." is an award-nominated Canadian drama about the rise and fall of a fictional wrestler. The 2009 "South Park" episode "W.T.F." played on the soap operatic elements of professional wrestling. One of the lead characters on the Disney Channel series "Kim Possible" was a huge fan of pro wrestling and actually featured it on an episode (with two former WWE wrestlers voicing the two fictitious wrestlers featured in the episode). The 2008 film "The Wrestler", about a washed-up professional wrestler, garnered several Oscar nominations. The 1950 film noir "Night and the City", directed by Jules Dassin and starring Richard Widmark and Gene Tierney, told the story of a promoter in London trying to make it big, and featured a match involving real professional wrestler Stanislaus Zbyszko. Wrestling has also gained a major following on YouTube with WWE being the being the most subscribed to Wrestling channel and sixth most subscribed to channel in the world. AEW also hosts it's AEW Dark show on YouTube. In addition to this, channels such as Wrestletalk, Wrestling With Wregret, Whatculture Wrestling, Cultaholic and The Whole Fucking Show have benefited gaining massive cult followings. With its growing popularity, professional wrestling has attracted attention as a subject of serious academic study and journalistic criticism. Many courses, theses, essays and dissertations have analyzed wrestling's conventions, content, and its role in modern society. It is often included as part of studies on theatre, sociology, performance, and media. The Massachusetts Institute of Technology developed a course of study on the cultural significance of professional wrestling, and anthropologist Heather Levi has written an ethnography about the culture of lucha libre in Mexico This was not always the case. In the early 20th century, once it became apparent that the "sport" was worked, pro wrestling was looked down on as a cheap entertainment for the uneducated working class, an attitude that still exists to varying degrees today. The French theorist Roland Barthes was among the first to propose that wrestling was worthy of deeper analysis, in his essay "The World of Wrestling" from his book "Mythologies", first published in 1957. Barthes argued that it should be looked at not as a scamming of the ignorant, but as spectacle; a mode of theatric performance for a willing, if bloodthirsty, audience. Wrestling is described as performed art which demands an immediate reading of the juxtaposed meanings. The logical conclusion is given least importance over the theatrical performers of the wrestlers and the referee. According to Barthes, the function of a wrestler is not to win: it is to go exactly through the motions which are expected of him and to give the audience a theatrical spectacle. This work is considered a foundation of all later study. While pro wrestling is often described simplistically as a "soap opera for males", it has also been cited as filling the role of past forms of literature and theatre; a of classical heroics, commedia dell'arte, revenge tragedies, morality plays, and burlesque. The characters and storylines portrayed by a successful promotion are seen to reflect the current mood, attitudes, and concerns of that promotion's society and can in turn influence those same things. Wrestling's high levels of violence and masculinity make it a vicarious outlet for aggression during peacetime. Documentary filmmakers have studied the lives of wrestlers and the effects the profession has on them and their families. The 1999 theatrical documentary "Beyond the Mat" focused on Terry Funk, a wrestler nearing retirement; Mick Foley, a wrestler within his prime; Jake Roberts, a former star fallen from grace; and a school of wrestling student trying to break into the business. The 2005 release "" chronicled the development of women's wrestling throughout the 20th century. Pro wrestling has been featured several times on HBO's "Real Sports with Bryant Gumbel". MTV's documentary series "True Life" featured two episodes titled "I'm a Professional Wrestler" and "I Want to Be a Professional Wrestler." Other documentaries have been produced by The Learning Channel ("The Secret World of Professional Wrestling") and A&E (""). "Bloodstained Memoirs" explored the careers of several pro wrestlers, including Chris Jericho, Rob Van Dam and Roddy Piper. Although professional wrestling is choreographed, there is a high chance of injury, and even death. Strikes are often stiff, especially in Japan and in independent wrestling promotions such as Combat Zone Wrestling and Ring of Honor. The ring is often made out of 2 inch by 8 inch timber planks. There have been many brutal accidents, hits and injuries. Many of the injuries that occur in pro wrestling are shoulders, knee, back, neck, and rib injuries. Professional wrestler Davey Richards said in 2015, "We train to take damage, we know we are going to take damage and we accept that." In April 2014, less than 25 years after the 1990 WrestleMania VI, one-third of its 36 competitors had died, including André the Giant and main event winner The Ultimate Warrior, with none of the deceased having reached the age of 64.
https://en.wikipedia.org/wiki?curid=24864
Pauli matrices In mathematical physics and mathematics, the Pauli matrices are a set of three complex matrices which are Hermitian and unitary. Usually indicated by the Greek letter sigma (), they are occasionally denoted by tau () when used in connection with isospin symmetries. They are These matrices are named after the physicist Wolfgang Pauli. In quantum mechanics, they occur in the Pauli equation which takes into account the interaction of the spin of a particle with an external electromagnetic field. Each Pauli matrix is Hermitian, and together with the identity matrix (sometimes considered as the zeroth Pauli matrix ), the Pauli matrices form a basis for the real vector space of Hermitian matrices. This means that any Hermitian matrix can be written in a unique way as a linear combination of Pauli matrices, with all coefficients being real numbers. Hermitian operators represent observables in quantum mechanics, so the Pauli matrices span the space of observables of the -dimensional complex Hilbert space. In the context of Pauli's work, represents the observable corresponding to spin along the th coordinate axis in three-dimensional Euclidean space . The Pauli matrices (after multiplication by to make them anti-Hermitian) also generate transformations in the sense of Lie algebras: the matrices form a basis for the real Lie algebra formula_2, which exponentiates to the special unitary group SU(2). The algebra generated by the three matrices is isomorphic to the Clifford algebra of , and the (unital associative) algebra generated by is isomorphic to that of quaternions. All three of the Pauli matrices can be compacted into a single expression: where is the imaginary unit, and is the Kronecker delta, which equals +1 if and 0 otherwise. This expression is useful for "selecting" any one of the matrices numerically by substituting values of , in turn useful when any of the matrices (but no particular one) is to be used in algebraic manipulations. The matrices are involutory: where is the identity matrix. The determinants and traces of the Pauli matrices are: From which, we can deduce that the eigenvalues of each are . With the inclusion of the identity matrix, (sometimes denoted ), the Pauli matrices form an orthogonal basis (in the sense of Hilbert–Schmidt) of the real Hilbert space of complex Hermitian matrices, formula_6, and the complex Hilbert space of all matrices, formula_7. The complex conjugate of a Pauli matrix can be compactly expressed as formula_8. Each of the (Hermitian) Pauli matrices has two eigenvalues, and . Using a convention in which prior to normalization, the 1 is placed into the top and bottom positions of the + and – wavefunctions respectively, the corresponding normalized eigenvectors are: An advantage of using this convention is that the + and – wavefunctions may be related to one another, using the Pauli matrices themselves, by formula_10, formula_11 and formula_11. The Pauli vector is defined by and provides a mapping mechanism from a vector basis to a Pauli matrix basis as follows, using the summation convention. Further, its eigenvalues being formula_16, and moreover (see completeness, below) Its (unnormalized) eigenvectors are The Pauli matrices obey the following commutation relations: and anticommutation relations: where the structure constant is the Levi-Civita symbol, Einstein summation notation is used, is the Kronecker delta, and is the identity matrix. For example, Pauli vectors elegantly map these commutation and anticommutation relations to corresponding vector products. Adding the commutator to the anticommutator gives so that, Contracting each side of the equation with components of two -vectors and (which commute with the Pauli matrices, i.e., for each matrix and vector component (and likewise with ), and relabeling indices , to prevent notational conflicts, yields Finally, translating the index notation for the dot product and cross product results in If formula_24 is identified with the pseudoscalar formula_25 then the right hand side becomes formula_26 which is also the definition for the product of two vectors in geometric algebra. Following traces can be derived using the commutation and anticommutation relations. If the matrix formula_28 is thrown into the mix, these relationships become where greek indices formula_30 and formula_31 assume values from formula_32 and the notation formula_33 is used to denote the sum over the cyclic permutation of the included indices. For one has, for even powers, formula_35 which can be shown first for the formula_37 case using the anticommutation relations. For convenience, the case formula_38 is taken to be formula_39 by convention. For odd powers, formula_40 Matrix exponentiating, and using the Taylor series for sine and cosine, In the last line, the first sum is the cosine, while the second sum is the sine; so, finally, which is analogous to Euler's formula, extended to quaternions. Note that while the determinant of the exponential itself is just , which makes it the generic group element of SU(2). A more abstract version of formula for a general matrix can be found in the article on matrix exponentials. A general version of for an analytic (at "a" and −"a") function is provided by application of Sylvester's formula, A straightforward application of formula provides a parameterization of the composition law of the group . One may directly solve for in which specifies the generic group multiplication, where, manifestly, the spherical law of cosines. Given , then, Consequently, the composite rotation parameters in this group element (a closed form of the respective BCH expansion in this case) simply amount to It is also straightforward to likewise work out the adjoint action on the Pauli vector, namely rotation effectively by double the angle , An alternative notation that is commonly used for the Pauli matrices is to write the vector index in the superscript, and the matrix indices as subscripts, so that the element in row and column of the -th Pauli matrix is . In this notation, the completeness relation for the Pauli matrices can be written As noted above, it is common to denote the 2 × 2 unit matrix by "σ"0, so "σ"0"αβ" = "δ""αβ". The completeness relation can alternatively be expressed as The fact that any 2 × 2 complex Hermitian matrices can be expressed in terms of the identity matrix and the Pauli matrices also leads to the Bloch sphere representation of 2 × 2 mixed states' density matrix, (2 × 2 positive semidefinite matrices with unit trace. This can be seen by first expressing an arbitrary Hermitian matrix as a real linear combination of as above, and then imposing the positive-semidefinite and trace conditions. For a pure state, in polar coordinates, formula_56, the idempotent density matrix acts on the state eigenvector formula_58 with eigenvalue 1, hence like a projection operator for it. Let be the transposition (also known as a permutation) between two spins and living in the tensor product space , This operator can also be written more explicitly as Dirac's spin exchange operator, Its eigenvalues are therefore 1 or −1. It may thus be utilized as an interaction term in a Hamiltonian, splitting the energy eigenvalues of its symmetric versus antisymmetric eigenstates. The group SU(2) is the Lie group of unitary matrices with unit determinant; its Lie algebra is the set of all anti-Hermitian matrices with trace 0. Direct calculation, as above, shows that the Lie algebra formula_62 is the 3-dimensional real algebra spanned by the set }. In compact notation, As a result, each can be seen as an infinitesimal generator of SU(2). The elements of SU(2) are exponentials of linear combinations of these three generators, and multiply as indicated above in discussing the Pauli vector. Although this suffices to generate SU(2), it is not a proper representation of, as the Pauli eigenvalues are scaled unconventionally. The conventional normalization is  , so that As SU(2) is a compact group, its Cartan decomposition is trivial. The Lie algebra is isomorphic to the Lie algebra , which corresponds to the Lie group SO(3), the group of rotations in three-dimensional space. In other words, one can say that the are a realization (and, in fact, the lowest-dimensional realization) of "infinitesimal" rotations in three-dimensional space. However, even though and are isomorphic as Lie algebras, and are not isomorphic as Lie groups. is actually a double cover of , meaning that there is a two-to-one group homomorphism from to , see relationship between SO(3) and SU(2). The real linear span of is isomorphic to the real algebra of quaternions . The isomorphism from to this set is given by the following map (notice the reversed signs for the Pauli matrices): Alternatively, the isomorphism can be achieved by a map using the Pauli matrices in reversed order, As the set of versors "U" ⊂ ℍ forms a group isomorphic to , "U" gives yet another way of describing . The two-to-one homomorphism from to may be given in terms of the Pauli matrices in this formulation. Quaternions form a division algebra—every non-zero element has an inverse—whereas Pauli matrices do not. In classical mechanics, Pauli matrices are useful in the context of the Cayley-Klein parameters. The matrix "P" corresponding to the position formula_67 of a point in space is defined in terms of the above Pauli vector matrix, Consequently, the transformation matrix formula_69 for rotations about the "x"-axis through an angle "θ" may be written in terms of Pauli matrices and the unit matrix as Similar expressions follow for general Pauli vector rotations as detailed above. In quantum mechanics, each Pauli matrix is related to an angular momentum operator that corresponds to an observable describing the spin of a spin ½ particle, in each of the three spatial directions. As an immediate consequence of the Cartan decomposition mentioned above, are the generators of a projective representation (spin representation) of the rotation group SO(3) acting on non-relativistic particles with spin ½. The states of the particles are represented as two-component spinors. In the same way, the Pauli matrices are related to the isospin operator. An interesting property of spin ½ particles is that they must be rotated by an angle of 4 in order to return to their original configuration. This is due to the two-to-one correspondence between SU(2) and SO(3) mentioned above, and the fact that, although one visualizes spin up/down as the north/south pole on the 2-sphere 2, they are actually represented by orthogonal vectors in the two dimensional complex Hilbert space. For a spin ½ particle, the spin operator is given by , the fundamental representation of "SU(2)". By taking Kronecker products of this representation with itself repeatedly, one may construct all higher irreducible representations. That is, the resulting spin operators for higher spin systems in three spatial dimensions, for arbitrarily large "j", can be calculated using this spin operator and ladder operators. They can be found in Rotation group SO(3)#A note on Lie algebra. The analog formula to the above generalization of Euler's formula for Pauli matrices, the group element in terms of spin matrices, is tractable, but less simple. Also useful in the quantum mechanics of multiparticle systems, the general Pauli group is defined to consist of all -fold tensor products of Pauli matrices. In relativistic quantum mechanics, the spinors in four dimensions are 4 × 1 (or 1 × 4) matrices. Hence the Pauli matrices or the Sigma matrices operating on these spinors have to be 4 × 4 matrices. They are defined in terms of 2 × 2 Pauli matrices as It follows from this definition that formula_72 matrices have the same algebraic properties as formula_73 matrices. However, relativistic angular momentum is not a three-vector, but a second order four-tensor. Hence formula_72 needs to be replaced by formula_75, the generator of Lorentz transformations on spinors. By the antisymmetry of angular momentum, the formula_75 are also antisymmetric. Hence there are only six independent matrices. The first three are the formula_77 The remaining three, formula_78, where the Dirac formula_79 matrices are defined as The relativistic spin matrices formula_75 are written in compact form in terms of commutator of gamma matrices as In quantum information, single-qubit quantum gates are "2" × "2" unitary matrices. The Pauli matrices are some of the most important single-qubit operations. In that context, the Cartan decomposition given above is called the "Z–Y decomposition of a single-qubit gate". Choosing a different Cartan pair gives a similar "X–Y decomposition of a single-qubit gate".
https://en.wikipedia.org/wiki?curid=24868
Pie menu In user interface design, a pie menu (also known as a radial menu) is a circular context menu where selection depends on direction. It is a graphical control element. A pie menu is made of several "pie slices" around an inactive center and works best with stylus input, and well with a mouse. Pie slices are drawn with a hole in the middle for an easy way to exit the menu. Pie menus work well with keyboard acceleration, particularly four and eight item menus, on the cursor keys and the number pad. A goal of pie menus is to provide a smooth, reliable gestural style of interaction for novices and experts. A slice can lead to another pie menu; selecting this may center the pointer in the new menu. A marking menu is a variant of this technique that makes the menu less sensitive to variance in gesture size. As a kind of context menu, pie menus are often context-sensitive, showing different options depending on what the pointer was pointing at when the menu was requested. The first documented radial menu is attributed to a system called PIXIE in 1969. Some universities explored alternative visual layouts. In 1986, Mike Gallaher and Don Hopkins together independently arrived at the concept of a context menu based on the angle to the origin where the exact angle and radius could be passed as parameters to a command, or the radius could be used to trigger a submenu. The first performance comparison to linear menus was performed in 1988 showing an increase in performance of 15% less time and a reduction of selection errors. The role-playing video game "Secret of Mana" featured an innovative icon-based radial menu system in 1993. Its ring menu system was adopted by later video games. For novice users, pie menus are easy because they are a self-revealing gestural interface: They show what you can do and direct you how to do it. By clicking and popping up a pie menu, looking at the labels, moving the pointer in the desired direction, then clicking to make a selection, users learn the menu and practice the gesture to "mark ahead" ("mouse ahead" in the case of a mouse, "wave ahead" in the case of a dataglove). With a little practice, it becomes quite easy to mark ahead even through nested pie menus. For the expert, the pie menus are more efficient. Because they might have built up the muscle memory for certain menu actions, and able to select the option they want without looking the pop up selections. In some cases, only when used more slowly like a traditional menu, does a pie menu pop up on the screen, to reveal the available selections. Moreover, novices can gradually become experts when they practice the same pie menu selection for many times and start to remember the menu and the motion. As Jaron Lanier of VPL Research has remarked, "The mind may forget, but the body remembers." Pie menus take advantage of the body's ability to remember muscle motion and direction, even when the mind has forgotten the corresponding symbolic labels. Pie menus are faster and more reliable to select from than linear menus, because selection depends on direction instead of distance. The circular menu slices are large in size and near the pointer for fast interaction (see Fitts's law). Experienced users use muscle memory without looking at the menu while selecting from it. Nested pie menus can efficiently offer many options, and some pie menus can pop up linear menus, and combine linear and radial items in the same menu. Pie menus just like any popup menu are shown only when requested, resulting in less visual distraction and cognitive load than toolbars and menu bars that are always shown. Pie menus show available options, in contrast to invisible mouse gestures. Pie menus, which delay appearance until the pointer is not moving, reduce intrusiveness to the same level as mouse gestures for experienced users. Pie menus take up more screen space than linear menus, and the number of slices in an individual menu must be kept low for effectiveness by using submenus. When using pie menus, submenus may overlap with the parent menu, but the parent menu may become translucent or hidden. Pie menus are most suited for actions that have been laid out by humans, and have logical grouping choices. Linear menus are most suited for dynamic, large menus that have many possible options, without any logical grouping, since pie menus can only show a limited number of menu items. Around 3-12 items can be reasonably accommodated in a radial layout, but additional items past that tend to counteract the benefits of using pie menus in the first place. This can be overcome with related techniques that allow chaining commands in one single gesture through submenus. However, using interaction techniques that are not pointer-based have proven problematic with both pie and linear menus for cluttered digital tabletop, where physical objects might occlude menu items. Pie menus are unavailable as standard graphical control element in common commercial toolkits. Video games often require custom widget development, so pie menu cost is lower in that particular scenario.
https://en.wikipedia.org/wiki?curid=24869
Pollution Pollution is the introduction of contaminants into the natural environment that cause adverse change. Pollution can take the form of chemical substances or energy, such as noise, heat or light. Pollutants, the components of pollution, can be either foreign substances/energies or naturally occurring contaminants. Pollution is often classed as point source or nonpoint source pollution. In 2015, pollution killed 9 million people in the world. Major forms of pollution include: Air pollution, light pollution, littering, noise pollution, plastic pollution, soil contamination, radioactive contamination, thermal pollution, visual pollution, water pollution. Air pollution has always accompanied civilizations. Pollution started from prehistoric times, when man created the first fires. According to a 1983 article in the journal "Science," "soot" found on ceilings of prehistoric caves provides ample evidence of the high levels of pollution that was associated with inadequate ventilation of open fires." Metal forging appears to be a key turning point in the creation of significant air pollution levels outside the home. Core samples of glaciers in Greenland indicate increases in pollution associated with Greek, Roman, and Chinese metal production. The burning of coal and wood, and the presence of many horses in concentrated areas made the cities the primary sources of pollution. The Industrial Revolution brought an infusion of untreated chemicals and wastes into local streams that served as the water supply. King Edward I of England banned the burning of sea-coal by proclamation in London in 1272, after its smoke became a problem; the fuel was so common in England that this earliest of names for it was acquired because it could be carted away from some shores by the wheelbarrow. It was the Industrial Revolution that gave birth to environmental pollution as we know it today. London also recorded one of the earlier extreme cases of water quality problems with the Great Stink on the Thames of 1858, which led to construction of the London sewerage system soon afterward. Pollution issues escalated as population growth far exceeded viability of neighborhoods to handle their waste problem. Reformers began to demand sewer systems and clean water. In 1870, the sanitary conditions in Berlin were among the worst in Europe. August Bebel recalled conditions before a modern sewer system was built in the late 1870s: The primitive conditions were intolerable for a world national capital, and the Imperial German government brought in its scientists, engineers, and urban planners to not only solve the deficiencies, but to forge Berlin as the world's model city. A British expert in 1906 concluded that Berlin represented "the most complete application of science, order and method of public life," adding "it is a marvel of civic administration, the most modern and most perfectly organized city that there is." The emergence of great factories and consumption of immense quantities of coal gave rise to unprecedented air pollution and the large volume of industrial chemical discharges added to the growing load of untreated human waste. Chicago and Cincinnati were the first two American cities to enact laws ensuring cleaner air in 1881. Pollution became a major issue in the United States in the early twentieth century, as progressive reformers took issue with air pollution caused by coal burning, water pollution caused by bad sanitation, and street pollution caused by the 3 million horses who worked in American cities in 1900, generating large quantities of urine and manure. As historian Martin Melosi notes, the generation that first saw automobiles replacing the horses saw cars as "miracles of cleanliness". By the 1940s, however, automobile-caused smog was a major issue in Los Angeles. Other cities followed around the country until early in the 20th century, when the short lived Office of Air Pollution was created under the Department of the Interior. Extreme smog events were experienced by the cities of Los Angeles and Donora, Pennsylvania in the late 1940s, serving as another public reminder. Air pollution would continue to be a problem in England, especially later during the industrial revolution, and extending into the recent past with the Great Smog of 1952. Awareness of atmospheric pollution spread widely after World War II, with fears triggered by reports of radioactive fallout from atomic warfare and testing. Then a non-nuclear event – the Great Smog of 1952 in London – killed at least 4000 people. This prompted some of the first major modern environmental legislation: the Clean Air Act of 1956. Pollution began to draw major public attention in the United States between the mid-1950s and early 1970s, when Congress passed the Noise Control Act, the Clean Air Act, the Clean Water Act, and the National Environmental Policy Act. Severe incidents of pollution helped increase consciousness. PCB dumping in the Hudson River resulted in a ban by the EPA on consumption of its fish in 1974. National news stories in the late 1970s – especially the long-term dioxin contamination at Love Canal starting in 1947 and uncontrolled dumping in Valley of the Drums – led to the Superfund legislation of 1980. The pollution of industrial land gave rise to the name brownfield, a term now common in city planning. The development of nuclear science introduced radioactive contamination, which can remain lethally radioactive for hundreds of thousands of years. Lake Karachay – named by the Worldwatch Institute as the "most polluted spot" on earth – served as a disposal site for the Soviet Union throughout the 1950s and 1960s. Chelyabinsk, Russia, is considered the "Most polluted place on the planet". Nuclear weapons continued to be tested in the Cold War, especially in the earlier stages of their development. The toll on the worst-affected populations and the growth since then in understanding about the critical threat to human health posed by radioactivity has also been a prohibitive complication associated with nuclear power. Though extreme care is practiced in that industry, the potential for disaster suggested by incidents such as those at Three Mile Island, Chernobyl, and Fukushima pose a lingering specter of public mistrust. Worldwide publicity has been intense on those disasters. Widespread support for test ban treaties has ended almost all nuclear testing in the atmosphere. International catastrophes such as the wreck of the Amoco Cadiz oil tanker off the coast of Brittany in 1978 and the Bhopal disaster in 1984 have demonstrated the universality of such events and the scale on which efforts to address them needed to engage. The borderless nature of atmosphere and oceans inevitably resulted in the implication of pollution on a planetary level with the issue of global warming. Most recently the term persistent organic pollutant (POP) has come to describe a group of chemicals such as PBDEs and PFCs among others. Though their effects remain somewhat less well understood owing to a lack of experimental data, they have been detected in various ecological habitats far removed from industrial activity such as the Arctic, demonstrating diffusion and bioaccumulation after only a relatively brief period of widespread use. A much more recently discovered problem is the Great Pacific Garbage Patch, a huge concentration of plastics, chemical sludge and other debris which has been collected into a large area of the Pacific Ocean by the North Pacific Gyre. This is a less well known pollution problem than the others described above, but nonetheless has multiple and serious consequences such as increasing wildlife mortality, the spread of invasive species and human ingestion of toxic chemicals. Organizations such as 5 Gyres have researched the pollution and, along with artists like Marina DeBris, are working toward publicizing the issue. Pollution introduced by light at night is becoming a global problem, more severe in urban centres, but nonetheless contaminating also large territories, far away from towns. Growing evidence of local and global pollution and an increasingly informed public over time have given rise to environmentalism and the environmental movement, which generally seek to limit human impact on the environment. The major forms of pollution are listed below along with the particular contaminant relevant to each of them: A pollutant is a waste material that pollutes air, water, or soil. Three factors determine the severity of a pollutant: its chemical nature, the concentration, the area affected and the persistence. Pollution has a cost. Manufacturing activities that cause air pollution impose health and clean-up costs on the whole of society, whereas the neighbors of an individual who chooses to fire-proof his home may benefit from a reduced risk of a fire spreading to their own homes. A manufacturing activity that causes air pollution is an example of a negative externality in production. A negative externality in production occurs “when a firm’s production reduces the well-being of others who are not compensated by the firm." For example, if a laundry firm exists near a polluting steel manufacturing firm, there will be increased costs for the laundry firm because of the dirt and smoke produced by the steel manufacturing firm. If external costs exist, such as those created by pollution, the manufacturer will choose to produce more of the product than would be produced if the manufacturer were required to pay all associated environmental costs. Because responsibility or consequence for self-directed action lies partly outside the self, an element of externalization is involved. If there are external benefits, such as in public safety, less of the good may be produced than would be the case if the producer were to receive payment for the external benefits to others. However, goods and services that involve negative externalities in production, such as those that produce pollution, tend to be over-produced and underpriced since the externality is not being priced into the market. Pollution can also create costs for the firms producing the pollution. Sometimes firms choose, or are forced by regulation, to reduce the amount of pollution that they are producing. The associated costs of doing this are called abatement costs, or marginal abatement costs if measured by each additional unit. In 2005 pollution abatement capital expenditures and operating costs in the US amounted to nearly $27 billion. Society derives some indirect utility from pollution, otherwise there would be no incentive to pollute. This utility comes from the consumption of goods and services that create pollution. Therefore, it is important that policymakers attempt to balance these indirect benefits with the costs of pollution in order to achieve an efficient outcome. It is possible to use environmental economics to determine which level of pollution is deemed the social optimum. For economists, pollution is an “external cost and occurs only when one or more individuals suffer a loss of welfare,” however, there exists a socially optimal level of pollution at which welfare is maximized. This is because consumers derive utility from the good or service manufactured, which will outweigh the social cost of pollution until a certain point. At this point the damage of one extra unit of pollution to society, the marginal cost of pollution, is exactly equal to the marginal benefit of consuming one more unit of the good or service. In markets with pollution, or other negative externalities in production, the free market equilibrium will not account for the costs of pollution on society. If the social costs of pollution are higher than the private costs incurred by the firm, then the true supply curve will be higher. The point at which the social marginal cost and market demand intersect gives the socially optimal level of pollution. At this point, the quantity will be lower and the price will be higher in comparison to the free market equilibrium. Therefore, the free market outcome could be considered a market failure because it “does not maximize efficiency”. This model can be used as a basis to evaluate different methods of internalizing the externality. Some examples include tariffs, a carbon tax and cap and trade systems. Air pollution comes from both natural and human-made (anthropogenic) sources. However, globally human-made pollutants from combustion, construction, mining, agriculture and warfare are increasingly significant in the air pollution equation. Motor vehicle emissions are one of the leading causes of air pollution. China, United States, Russia, India Mexico, and Japan are the world leaders in air pollution emissions. Principal stationary pollution sources include chemical plants, coal-fired power plants, oil refineries, petrochemical plants, nuclear waste disposal activity, incinerators, large livestock farms (dairy cows, pigs, poultry, etc.), PVC factories, metals production factories, plastics factories, and other heavy industry. Agricultural air pollution comes from contemporary practices which include clear felling and burning of natural vegetation as well as spraying of pesticides and herbicides About 400 million metric tons of hazardous wastes are generated each year. The United States alone produces about 250 million metric tons. Americans constitute less than 5% of the world's population, but produce roughly 25% of the world's , and generate approximately 30% of world's waste. In 2007, China overtook the United States as the world's biggest producer of , while still far behind based on per capita pollution (ranked 78th among the world's nations). In February 2007, a report by the Intergovernmental Panel on Climate Change (IPCC), representing the work of 2,500 scientists, economists, and policymakers from more than 120 countries, confirmed that humans have been the primary cause of global warming since 1950. Humans have ways to cut greenhouse gas emissions and avoid the consequences of global warming, a major climate report concluded. But to change the climate, the transition from fossil fuels like coal and oil needs to occur within decades, according to the final report this year from the UN's Intergovernmental Panel on Climate Change (IPCC). Some of the more common soil contaminants are chlorinated hydrocarbons (CFH), heavy metals (such as chromium, cadmium – found in rechargeable batteries, and lead – found in lead paint, aviation fuel and still in some countries, gasoline), MTBE, zinc, arsenic and benzene. In 2001 a series of press reports culminating in a book called "Fateful Harvest" unveiled a widespread practice of recycling industrial byproducts into fertilizer, resulting in the contamination of the soil with various metals. Ordinary municipal landfills are the source of many chemical substances entering the soil environment (and often groundwater), emanating from the wide variety of refuse accepted, especially substances illegally discarded there, or from pre-1970 landfills that may have been subject to little control in the U.S. or EU. There have also been some unusual releases of polychlorinated dibenzodioxins, commonly called "dioxins" for simplicity, such as TCDD. Pollution can also be the consequence of a natural disaster. For example, hurricanes often involve water contamination from sewage, and petrochemical spills from ruptured boats or automobiles. Larger scale and environmental damage is not uncommon when coastal oil rigs or refineries are involved. Some sources of pollution, such as nuclear power plants or oil tankers, can produce widespread and potentially hazardous releases when accidents occur. In the case of noise pollution the dominant source class is the motor vehicle, producing about ninety percent of all unwanted noise worldwide. Adverse air quality can kill many organisms, including humans. Ozone pollution can cause respiratory disease, cardiovascular disease, throat inflammation, chest pain, and congestion. Water pollution causes approximately 14,000 deaths per day, mostly due to contamination of drinking water by untreated sewage in developing countries. An estimated 500 million Indians have no access to a proper toilet, Over ten million people in India fell ill with waterborne illnesses in 2013, and 1,535 people died, most of them children. Nearly 500 million Chinese lack access to safe drinking water. A 2010 analysis estimated that 1.2 million people died prematurely each year in China because of air pollution. The high smog levels China has been facing for a long time can do damage to civilians' bodies and cause different diseases. The WHO estimated in 2007 that air pollution causes half a million deaths per year in India. Studies have estimated that the number of people killed annually in the United States could be over 50,000. Oil spills can cause skin irritations and rashes. Noise pollution induces hearing loss, high blood pressure, stress, and sleep disturbance. Mercury has been linked to developmental deficits in children and neurologic symptoms. Older people are majorly exposed to diseases induced by air pollution. Those with heart or lung disorders are at additional risk. Children and infants are also at serious risk. Lead and other heavy metals have been shown to cause neurological problems. Chemical and radioactive substances can cause cancer and as well as birth defects. An October 2017 study by the Lancet Commission on Pollution and Health found that global pollution, specifically toxic air, water, soils and workplaces, kills nine million people annually, which is triple the number of deaths caused by AIDS, tuberculosis and malaria combined, and 15 times higher than deaths caused by wars and other forms of human violence. The study concluded that "pollution is one of the great existential challenges of the Anthropocene era. Pollution endangers the stability of the Earth’s support systems and threatens the continuing survival of human societies." Pollution has been found to be present widely in the environment. There are a number of effects of this: The Toxicology and Environmental Health Information Program (TEHIP) at the United States National Library of Medicine (NLM) maintains a comprehensive toxicology and environmental health web site that includes access to resources produced by TEHIP and by other government agencies and organizations. This web site includes links to databases, bibliographies, tutorials, and other scientific and consumer-oriented resources. TEHIP also is responsible for the Toxicology Data Network (TOXNET) an integrated system of toxicology and environmental health databases that are available free of charge on the web. TOXMAP is a Geographic Information System (GIS) that is part of TOXNET. TOXMAP uses maps of the United States to help users visually explore data from the United States Environmental Protection Agency's (EPA) Toxics Release Inventory and Superfund Basic Research Programs. A 2019 paper linked pollution to adverse school outcomes for children. A number of studies show that pollution has an adverse effect on the productivity of both indoor and outdoor workers. To protect the environment from the adverse effects of pollution, many nations worldwide have enacted legislation to regulate various types of pollution as well as to mitigate the adverse effects of pollution. Pollution control is a term used in environmental management. It means the control of emissions and effluents into air, water or soil. Without pollution control, the waste products from overconsumption, heating, agriculture, mining, manufacturing, transportation and other human activities, whether they accumulate or disperse, will degrade the environment. In the hierarchy of controls, pollution prevention and waste minimization are more desirable than pollution control. In the field of land development, low impact development is a similar technique for the prevention of urban runoff. The earliest precursor of pollution generated by life forms would have been a natural function of their existence. The attendant consequences on viability and population levels fell within the sphere of natural selection. These would have included the demise of a population locally or ultimately, species extinction. Processes that were untenable would have resulted in a new balance brought about by changes and adaptations. At the extremes, for any form of life, consideration of pollution is superseded by that of survival. For humankind, the factor of technology is a distinguishing and critical consideration, both as an enabler and an additional source of byproducts. Short of survival, human concerns include the range from quality of life to health hazards. Since science holds experimental demonstration to be definitive, modern treatment of toxicity or environmental harm involves defining a level at which an effect is observable. Common examples of fields where practical measurement is crucial include automobile emissions control, industrial exposure (e.g. Occupational Safety and Health Administration (OSHA) PELs), toxicology (e.g. ), and medicine (e.g. medication and radiation doses). "The solution to pollution is dilution", is a dictum which summarizes a traditional approach to pollution management whereby sufficiently diluted pollution is not harmful. It is well-suited to some other modern, locally scoped applications such as laboratory safety procedure and hazardous material release emergency management. But it assumes that the diluent is in virtually unlimited supply for the application or that resulting dilutions are acceptable in all cases. Such simple treatment for environmental pollution on a wider scale might have had greater merit in earlier centuries when physical survival was often the highest imperative, human population and densities were lower, technologies were simpler and their byproducts more benign. But these are often no longer the case. Furthermore, advances have enabled measurement of concentrations not possible before. The use of statistical methods in evaluating outcomes has given currency to the principle of probable harm in cases where assessment is warranted but resorting to deterministic models is impractical or infeasible. In addition, consideration of the environment beyond direct impact on human beings has gained prominence. Yet in the absence of a superseding principle, this older approach predominates practices throughout the world. It is the basis by which to gauge concentrations of effluent for legal release, exceeding which penalties are assessed or restrictions applied. One such superseding principle is contained in modern hazardous waste laws in developed countries, as the process of diluting hazardous waste to make it non-hazardous is usually a regulated treatment process. Migration from pollution dilution to elimination in many cases can be confronted by challenging economical and technological barriers. Carbon dioxide, while vital for photosynthesis, is sometimes referred to as pollution, because raised levels of the gas in the atmosphere are affecting the Earth's climate. Disruption of the environment can also highlight the connection between areas of pollution that would normally be classified separately, such as those of water and air. Recent studies have investigated the potential for long-term rising levels of atmospheric carbon dioxide to cause slight but critical increases in the acidity of ocean waters, and the possible effects of this on marine ecosystems. Air pollution fluctuations have been known to strongly depend on the weather dynamics. A recent study developed a multi-layered network analysis and detected strong interlinks between the geopotential height of the upper air ( 5 km) and surface air pollution in both China and the USA. This study indicates that Rossby waves significantly affect air pollution fluctuations through the development of cyclone and anticyclone systems, and further affect the local stability of the air and the winds. The Rossby waves impact on air pollution has been observed in the daily fluctuations in surface air pollution. Thus, the impact of Rossby waves on human life is significant and rapid warming of the Arctic could slow down Rossby waves, thus increasing human health risks. The Pure Earth, an international non-for-profit organization dedicated to eliminating life-threatening pollution in the developing world, issues an annual list of some of the world's most polluting industries. A 2018 report by the Institute for Agriculture and Trade Policy and GRAIN says that the meat and dairy industries are poised to surpass the oil industry as the world's worst polluters. Pure Earth issues an annual list of some of the world's worst polluted places.
https://en.wikipedia.org/wiki?curid=24872
Robotech Robotech is a science fiction franchise that began with an 85-episode anime television series produced by Harmony Gold USA in association with Tatsunoko Production and first released in the United States in 1985. The show was adapted from three original and distinct, though visually similar, Japanese anime television series ("Super Dimension Fortress Macross", "Super Dimension Cavalry Southern Cross" and "Genesis Climber MOSPEADA") to make a series suitable for syndication. In the series, "Robotechnology" refers to the scientific advances discovered in an alien starship that crashed on a South Pacific island. With this technology, Earth developed robotic technologies, such as transformable mecha, to fight three successive extraterrestrial invasions. Prior to the release of the TV series, the name "Robotech" was used by model kit manufacturer Revell on their "Robotech Defenders" line in the mid-1980s. The line consisted of mecha model kits imported from Japan and featured in anime titles such as "Super Dimension Fortress Macross" (1982), "Super Dimension Century Orguss" (1983) and "Fang of the Sun Dougram" (1981). The kits were originally intended to be a marketing tie-in to a similarly named comic book series by DC Comics, which was cancelled after only two issues. At the same time, Harmony Gold licensed the "Macross" TV series for direct-to-video distribution in 1984, but their merchandising plans were compromised by Revell's prior distribution of the "Macross" kits. In the end, both parties signed a co-licensing agreement and the "Robotech" name was adopted for the TV syndication of "Macross" combined with "Super Dimension Cavalry Southern Cross" (1984) and "Genesis Climber MOSPEADA" (1983). "The "Robotech" chronology, according to Harmony Gold, is illustrated below: Note: Asterisked works are now considered 'secondary continuity'—that is, that their events exist in the continuity of "Robotech", but 'don't count' when conflicts arise with the primary continuity that comprises the three-part "Robotech" TV series and 2006's "". In 2002, with the publication of the WildStorm (DC) comics, Harmony Gold officially decided to retcon the "Robotech" Universe. The following "Robotech" material is now relegated to the status of secondary continuity: While these materials are not precisely 'retired' or 'removed' from the continuity, their events are subject to critical review, and are strictly subordinate to the 'official' events of the 85-episode animated series. "Robotech" (1985) is an original story adapted with edited content and revised dialogue from the animation of three different mecha anime series: Harmony Gold's cited reasoning for combining these unrelated series was its decision to market "Macross" for American weekday syndication television, which required a minimum of 65 episodes at the time (thirteen weeks at five episodes per week). "Macross" and the two other series each had fewer episodes than required, since they originally aired in Japan as weekly series. On some television stations, the syndicated run was preceded by the broadcast premiere of "", a feature-length pilot. This combination resulted in a storyline that spans three generations, as mankind must fight three destructive 'Robotech Wars' in succession with various invading forces, each of which is motivated in one way or another by a desire for a powerful energy source called 'protoculture'. While each of the three animated series used for its footage informs its content, the Robotech storyline is distinct and separate from each of them. "Robotech: The Movie", also called "Robotech: The Untold Story", is a feature film and was the first new "Robotech" adventure created after the premiere of the original series. It uses footage from the "Megazone 23 – Part 1" OVA (original video animation; made-for-video animated feature) combined with scenes from "Southern Cross" and additional original animation produced for the film. The original plan for the film was to have it set during the Macross Saga, parallel to the SDF-1's return to Earth from Pluto. The film would also have served as a prequel to the Sentinels, as both projects were initially meant to share many characters. Harmony Gold producer Carl Macek worked with the OVA's original creators to make the story and the new ending work. The film had to be changed again after the distributor of the film, Cannon Films, saw an incomplete rough cut of the film and were upset by it. They ordered Macek to remove multiple scenes from the film and to add more violence (most of the scenes removed were scenes setting up characters and showing female characters interacting). Macek reluctantly did what they ordered, and created a new script and rough edit for the film in less than 24 hours. When the distributors saw Macek act out the new film, they were much more pleased with the new cut. The opening night in Texas received a positive response, but Cannon Films pulled out after noting that most attendants were adults; the bulk of the scheduled advertising for the series was targeted to children. The film had limited success in Argentina and Belgium. In 2011, A&E Home Video released, as a part of their "Robotech: The Complete Series" collection, a 29-minute version of "Robotech: The Movie" containing only footage used from "Southern Cross". There was no attempt to remaster the footage. This aborted American-produced series would have followed the continuing adventures of Rick and Lisa Hunter and the Robotech Expedition during the events of "The Masters" and "The New Generation". The feature-length pilot is composed of the first three (and only) episodes that were produced. "The Sentinels" featured characters from all three "Robotech" sagas and introduced the SDF-3 along with an overview of their new mission. The series was planned to have a total of 65 episodes. In "Robotech Art 3: The Sentinels", Carl Macek blamed the cancellation of the series on the crash of the Yen/Dollar exchange rate, which caused toy partner Matchbox to withdraw from the project. Harmony Gold lacked the funds to produce the series on its own, and production ceased after only three episodes. "Robotech II: The Sentinels" was released on VHS by Palladium Books. In 2011, a "remastered" version was released on the A&E DVD set, "Robotech: The Complete Original Series" DVD. This version has opening titles resembling those found on the "Robotech Remastered" DVDs, as well as a new ending with text explaining the fate of the SDF-3. Also, all of the flashback footage used from "The Macross Saga" has been removed, including the re-used footage from the episode "Wedding Bells". In 2002, Tommy Yune announced development of a new sequel film, which was untitled until 2004 as "Robotech: Shadow Force". The storyline overlaps with and continues from the unresolved ending of the original series. The title of the story arc was soon changed to "". The first trailers with finished animation were shown at Anime Expo and Comic-Con International in 2005. It was not until February 2006, when Kevin McKeever, operations coordinator at Harmony Gold, was able to confirm that the pilot movie had been completed. After a series of delays, FUNimation Entertainment was finally announced as the home video, broadcast, and theatrical distributor at the 2006 Comic-Con International in San Diego with the possibility of producing further sequels. Harmony Gold premiered the movie at various film festivals in 2006, and it was first seen by a public audience at MechaCon on August 9, 2006, where it was showcased as a charity screening to help raise funds for the ongoing Hurricane Katrina and Hurricane Rita recovery effort. A limited theatrical run followed in January 2007, and the film was released on DVD on February 6, 2007. A two-disc collector's edition was released in November 2007. First revealed in late 2011 in the final minutes of "Carl Macek's Robotech Universe", a documentary on the making of "Robotech" dedicated to the then-recent passing of Macek, "Love Live Alive" is an adaptation of the 1985 "Genesis Climber Mospeada" OVA, "Love Live Alive", incorporating some brand-new animation. The film was released on DVD on July 23, 2013, by Lionsgate Home Entertainment in North America. This promotional VHS tape created by Matchbox was included with their "Robotech Wars" playset. This video includes two episodes cobbled together from clips of "The Macross Saga". Titled "To the End of the Universe" and "Battle Royale", these episodes contain no new footage, and are not meant to follow any continuity established in the TV series. Carl Macek revealed ideas for another proposed series, "Robotech: The Odyssey", which would have picked up where "The New Generation" and end of "Robotech: The Sentinels" left off, and eventually created a circular storyline that would end where the original "Robotech" began in a giant 260-episode cycle to fill up all the weekdays in a year. According to Macek, "The Odyssey" would have involved the SDF-3 travelling back into the past to the days before the birth of Zor (as well as Scott Bernard's search for the SDF-3). The SDF-3's crew would become citizens of the Robotech Masters' homeworld and change time by becoming a part of its history. Ultimately, it would be revealed that Lynn Minmei was the mother of Zor, making Minmei the focal point of Robotech. The final episode of the Odyssey would be of Zor dying and his Super Dimension Fortress (the SDF-1) being launched into space, and eventually crash landing on Earth in 1999. The next episode after that would be "Boobytrap", episode 1 of the original series which in turn will create an endless loop within the "Robotech" universe. After the failure of "Sentinels", "Odyssey" never went into development, although some of its ideas were worked into the final Jack McKinney novel "The End of the Circle", which wrapped up all of the outstanding plot threads left by the original series and the previous "Robotech" novels. Fan publication "Macross Life" interviewed Harmony Gold executive Richard Firth in 1986, where he revealed that Macek had "plans through "ROBOTECH V", which would give us an episode for each day of the year for a year and a half." He also said that these two installments would have brought the series to 285 episodes. Regarding the plot, Firth mentioned a "retired Commodore Hunter, whomever that may be, could very well be speaking at the graduation of the later day cadets or whatever, and they ask him to tell them the story all over again: it comes back [to the first episode of the series]." Macek himself described a fourth and fifth series envisioned for Robotech in Chris Meadow's Space Station Liberty podcast in 2007. Macek mentioned the original series as the first, The Sentinels as Robotech II, The Odyssey as Robotech III, and then 2 further series detailing the evolution of Zor, bringing the combined series' total to approximately 300 episodes. Macek attempted another sequel with the development of "Robotech 3000". This all-CGI series would have been set a millennium in the future of the "Robotech" universe and feature none of the old series' characters. In the three-minute trailer, an expedition is sent to check on a non-responsive mining outpost and is attacked by "infected" Veritech mecha. The idea was abandoned midway into production after negative reception within the company, negative fan reactions at the FanimeCon anime convention in 2000, and financial difficulties within Netter Digital who was animating the show. The trailer is hosted on the official "Robotech" website, and was included in the 2007 release of the "Robotech: The Shadow Chronicles" 2-disc collector's DVD, along with behind-the-scenes motion capture footage. In October 2004, veteran animation writer and producer Greg Weisman revealed that he wrote developed an animated spin-off series titled "Robotech: Mars Force". When asked about the project, Weisman said that he was under a non-disclosure agreement with Harmony Gold and was only allowed to mention that he developed the series. In 2006, Harmony Gold Creative Director Tommy Yune elaborated on the project in the Space Station Liberty Podcast, saying that Mars Force was a series geared at younger audiences, following the children Robotech Expeditionary Force. A similar plot would later be used for canceled 2014 spin-off, "Robotech Academy". A sixty-second public service announcement for the 60th anniversary of the United Nations, featuring Scott Bernard and Ariel, was animated during the production of "The Shadow Chronicles". Although it did not use the original voice actors and the dialogue was somewhat out-of-character, it nonetheless marked the first fully completed "Robotech" footage in many years. On July 27, 2007, at their Comic-Con International panel, Harmony Gold and Yune unveiled the second entry of the "Shadow Chronicles" production, titled "Robotech: Shadow Rising" and was to be a co-production with FUNimation Entertainment. Pre-production reportedly began on February 2007 and a projected release date of sometime in 2009 was originally expected. Production ceased after Harmony Gold terminated their deal with FUNimation Entertainment due to creative differences. At Comic-Con 2012, Tommy Yune announced that "Love Live Alive" would pave the way for "Shadow Rising". As of 2015, the Shadow Rising trademark remains abandoned since 2007. On July 5, 2014, Harmony Gold started a Kickstarter project for "Robotech Academy", which Macek had developed before he died. The goal of this project was to raise US$500,000 to produce a new 24-minute pilot episode. The crowdfunding project was to have closed on August 9, 2014; however, on August 2, the project was canceled with a pledge level of US$194,574, or 39% of its target. Harmony Gold, however, announced that further plans to fund the project were being explored. At the 2014 Long Beach Comic Con, it was announced that the producers at Harmony Gold were in talks with at least one new media network on the prospect of producing the show. As of December 7, 2015, the project remains abandoned. In the 1990s, Seishun Shitemasu, an anime fandubbing group, produced the parodies "Robotech III: Not Necessarily the Sentinels" and "Robotech IV: Khyron's Counterattack", using footage from, respectively, "Gunbuster" and "", continuing the tradition of the original Robotech's adaptation of unrelated anime series into a single continuity. On July 2, 2010, Ecuadorian animator Patricio "Pat" Mosquera uploaded to YouTube a teaser for "Robotech Skull Knights". On August 17, 2010, second teaser revealed Rick Hunter standing in front of an image of the VF-4 shown in the final episodes of the original series. "Robotech Skull Knights" has not been released yet. In July 2013, Patricio Mosquera was included as an animation director in the staff list in the IMDb page of "Love Live Alive". On December 31, 2012, Cesar Turturro uploaded to YouTube an Argentinian fan trailer for "Robotech Valkyrie Project". On December, 2013 the first episode was uploaded to YouTube, and in January 2014, the second episode was also uploaded. The series was cancelled after Harmony Gold issued a "cease and desist" letter to the producers. The team was, however, hired to do the CGI effects for "Robotech: Academy". On September 7, 2007, "The Hollywood Reporter" stated that Warner Bros. had acquired the film rights to "Robotech" and would be producing a live-action film with an as-yet-unknown release date. Tobey Maguire is producing the film through his Maguire Entertainment banner and is pursuing the lead role, in what the studio plans to be a tentpole science fiction franchise. Maguire stated, "We are very excited to bring "Robotech" to the big screen. There is a rich mythology that will be a great foundation for a sophisticated, smart and entertaining film." In an interview, Harmony Gold representative Kevin McKeever said that Warner Bros. had approached Harmony Gold about the project, that Harmony Gold would have "a say" in its creative direction, and that it was not expected to affect the production schedule for "". He was unable to confirm any details of budget, casting, expected release date, or storyline, explaining that it was too early in the life of the project for these things to have been decided. In June 2008, it was reported that Lawrence Kasdan had been hired to write the film, with Charles Roven and Akiva Goldsman joining Tobey Maguire as producers. During the Robotech Panel at Anime Expo 2008, the involvement of Maguire and Kasdan was confirmed, with Kasdan writing the script for the live-action film. Tommy Yune also revealed that the film is planned as a re-imagining of the original "Robotech" universe (with new updated mecha and character designs) and will take place several years in the future, departing from the original cartoon's 2009 setting. As of November 2008, Alfred Gough and Miles Millar (who both worked in "Smallville", "Spider-Man 2", "", and "") are the writers for the film. Roven is currently no longer working on the proposed film adaptation of the "Robotech" animated series, but he wished the remaining producers Goldsman and Maguire "fantastic luck" on the project. The Mania.com website reported on June 23, 2009, that British television writer and novelist Tom Rob Smith "has taken over writing duties" for the proposed film adaptation. Smith wrote for the British soaps "Family Affairs" and "Bad Girls" before writing the critically acclaimed crime suspense novel "Child 44". Smith will be the fourth writer or writing team to be reportedly attached to the upcoming film's pre-production. In early 2013, "The Hollywood Reporter" announced that Warner Bros. was in talks with commercial director Nic Mathieu to direct the film. On July 24, 2013, it was reported that Leonardo DiCaprio had turned down a role in "" and has shown interest to star as a main character in the upcoming big screen version of "Robotech". DiCaprio is a longtime friend of Tobey Maguire; they co-starred in "The Great Gatsby". Maguire will probably participate in the film as another one of the lead actors—while Nic Mathieu will direct. On February 4, 2015, Deadline.com reported Gianni Nunori and Mark Canton selected Michael B. Gordon to write the film's script and are looking at Andy Muschietti to direct it. On March 25, 2015, Variety announced that the "Robotech" franchise had been acquired by Sony Pictures, who views "Robotech" as a potential film franchise. On April 29, 2015, Deadline reported that James Wan is in talks to direct the film. On June 3, 2015, "The Hollywood Reporter" reported that Wan is confirmed to direct the film. On July 3, 2015, Kevin McKeever of Harmony Gold announced at Anime Expo that Sony has the rights to release this film worldwide, with the exception of Japan. On March 27, 2016, Wan told IGN that the film will follow the roots of the franchise. As of April 4, 2016, Harmony Gold's Kevin McKeever revealed on the official Robotech Facebook page that their deal with Sony is still not finalized. On July 17, 2017, it was reported by The Hollywood Reporter that Argentine filmmaker Andy Muschietti will direct the project, after Wan dropped out to work on "Aquaman". On September 12, 2017, Jason Fuchs was reportedly hired by the studio to write the script for the film. In September 2019, when asked about which project he would choose to direct next, director Andy Muschietti told the Argentinian news agency Télam that "Robotech" is a “complicated” property, citing its lack of popularity in the United States compared to other properties, and its requirement of a $100 million budget.  He did, however, say that the script is completed. At the time of its broadcast, Harmony Gold also launched "Robotech" through a popular line of comics to be followed by novels, role-playing games, toys, and other consumer products. With the cancellation of "Robotech II: The Sentinels", many of these licensed products were discontinued, and led to a drought of "Robotech" product through much of the 1990s, except for publishers who continued "The Sentinels" storyline in print. In 1986, Starblaze Graphics published "Robotech Art 1", a reference book containing artwork, Japanese production designs, and episode guides from the original television series. This was followed by "Robotech Art 2", which was largely a collection of art by various American artists and fans. In 1988, Carl Macek collected much of the unused designs from "Robotech II: The Sentinels" into "Robotech Art 3: The Sentinels", which also included his story outline for the rest of the unfinished series, with an explanation behind its cancellation. In 2007, Stone Bridge Press published "The Art of Robotech: The Shadow Chronicles". "Robotech" comics were first published in 1984 with DC Comics' short-lived "Robotech Defenders" and Comico's adaptation of the first episode of the Japanese version of "Macross". However, the first adaptation of the "Robotech" television series did not arrive until 1985 with Comico's "Robotech: The Macross Saga" Number 2, which continued from the first "Macross" issue. The various comic publishers include: The first "Robotech" collectible card game was released in 2006 by Hero Factory, which had previously produced "Robotech" trading cards. Various "Robotech" soundtracks have been released on records, cassettes, and compact discs since 1988. Since 1987, "Robotech" was adapted into novel form by "Jack McKinney", a pseudonym for the team of James Luceno and the late Brian Daley, a pair of writers who had been working with Macek since they had collaborated on the animated series "Galaxy Rangers". Using fictitious epigraphs in the style of "Dune", McKinney's novels fleshed out the chronology (including adapting the incomplete "Sentinels" source material) in far greater detail than the original animation. Many "Robotech" fans consider the McKinney series to be an unofficial canon of its own, despite notable divergences in the writing from Harmony Gold's current official animation-based canon. Despite no longer being considered core-continuity by Harmony Gold, the novels have been recently re-issued by Del Rey Books as Omnibus compilations. In 1986, Palladium Books published a role-playing game based on the "Robotech" series, including several books covering the "Sentinels" portion of the storyline. The original "Robotech" RPG line went out of print as of June 30, 2001, but Harmony Gold and Palladium Books signed an agreement in 2007 to produce a new line of Robotech RPG books, beginning with a book covering and promoting the feature-length film "The Shadow Chronicles". The "" sourcebook first book was released on March 21, 2008, followed by sourcebooks covering the Macross, Masters, and New Generation chapters of Robotech (redrafted to reflect the Harmony Gold canon). Other sourcebooks and supplements are reflected in the Palladium Books production pipeline. On April 18, 2013, Palladium started a campaign on the crowdfunding site Kickstarter for a tabletop miniatures game based on the Robotech RPG called "Robotech: RPG Tactics". The miniatures are being produced by Ninja Division (combining sculpting talents from Soda Pop Miniatures and Cipher Studios), and will feature multi part plastic miniatures that can be posed during assembly. The campaign reached its goal in 3 hours, and was initially scheduled to release in December 2013, but delays have persisted into 2018. In May 2019, under licencing from Harmony Gold and Strange Machine Games, Battlefield Press International produced a game book for the new Savage Worlds Adventure Edition. Action figures in the size of the three "Robotech" generations were initially released in 1985 by Matchbox toy company, but then reissued in 1992 by Harmony Gold (Lunk and Corg were only released by Matchbox and Lynn Minmei was only released by Harmony Gold). Each included a weapon and helmet where appropriate. Matchbox also released figures of Zentraedi characters from the first generation. These figures were supposed to represent the size difference between the Humans and the giant Zentraedi forces, but to be correct these figures would have to have been made about tall. None of the larger figures came with weapons but the Armored Zentraedi came with a removable helmet. Also many toys depicting the vehicles and mecha from the series were released by Matchbox in 1985, Harmony Gold in 1992 and Playmates Toys in 1994 (under the Exosquad line). There were major differences in packaging, toy stickers and colors between the different releases. The vehicles were designed to be used only with the 3¾-inch figures. The SDF-1 Playset was only released under the Matchbox line in the 1980s and could be used with both the 3¾- and six-inch figures. Harmony Gold and Matchbox were unable to sell the 1/55 VF-1 Valkyrie toy originally sold in Japan by Takatoku Toys due to Hasbro licensing it as Jetfire in the Transformers toy line. Because of this, they settled with manufacturing a non-transformable Veritech Fighter that could fit any of the 3¾-inch action figures, as well as importing the transformable super deformed Veritech Fighters (originally manufactured in Japan by Bandai as "Macross" VF-1 Valkyrie "Joke machines"). Since the late 1990s, there has been a resurgence of "Robotech"-related toys. In 2001, Toynami released the "Robotech Masterpiece Collection" line, featuring replicas of the Veritech Fighters of "The Macross Saga". Since then, Toynami has become the exclusive toy manufacturer of the "Robotech" franchise—having covered mecha from "The Macross Saga", "The New Generation" and "The Shadow Chronicles". "Robotech" spawned five video game licenses, of which the most recent three were released: "Robotech" is often a polarizing subject amongst anime fans. Some critics look down upon the show for its extensive edits to the source material (Westernizing character names, editing for content and chiefly, forging a connection between previously unrelated series), while supporters of the adaptation have pointed out that the weaving of three unrelated series into a contiguous whole necessarily required reworking, and that it helped to maintain a slow but continuous rise in the consumption of anime in the US. Series writer/actor Gregory Snegoff said in an interview on the now-defunct "Shadow Chronicles News" fansite that, "afterward, we received compliments from the Japanese who thought our dialogue and stories were better than the original," likely a reference to the creators of the latter two series, both of whom worked with the team on "The Sentinels". The producers of "Megazone 23 – Part 1" were very happy with the original plans for "" (where the incomplete film would have been added to the "Robotech" mythos to play part in "The Sentinels" storyline), and worked closely with Carl Macek to plan the new ending and animation. When the film reached a limited release, the new ending was released on a LaserDisc of "Megazone 23", with the title "Present For You." However, "Animag" magazine (issue 11) and "Animerica" magazine (issue 9, volume 4) reports that the staff of "Macross" at Studio Nue and Artland, such as the original story creator and mecha designer Shōji Kawamori and chief director Noboru Ishiguro, expressed their concern over the "Robotech" adaptation, and surprise at its differences. In 2009, "IGN" ranked "Robotech" as the 34th-greatest animated show of all time in their Top 100 list. Following the original broadcast, the series enjoyed popularity on home video in VHS and DVD formats from the following distributors:
https://en.wikipedia.org/wiki?curid=26333
Radioteletype Radioteletype (RTTY) is a telecommunications system consisting originally of two or more electromechanical teleprinters in different locations connected by radio rather than a wired link. These machines were superseded by personal computers (PCs) running software to emulate teleprinters. Radioteletype evolved from earlier landline teleprinter operations that began in the mid-1800s. The US Navy Department successfully tested printing telegraphy between an airplane and ground radio station in 1922. Later that year, the Radio Corporation of America successfully tested printing telegraphy via their Chatham, Massachusetts, radio station to the R.M.S. Majestic. Commercial RTTY systems were in active service between San Francisco and Honolulu as early as April 1932 and between San Francisco and New York City by 1934. The US military used radioteletype in the 1930s and expanded this usage during World War II. From the 1980s, teleprinters were replaced by computers running teleprinter emulation software. The term radioteletype is used to describe both the original radioteletype system, sometimes described as "Baudot", as well as the entire family of systems connecting two or more teleprinters or PCs using software to emulate teleprinters, over radio, regardless of alphabet, link system or modulation. In some applications, notably military and government, radioteletype is known by the acronym RATT (Radio Automatic Teletype). Landline teleprinter operations began in 1849 when a circuit was put in service between Philadelphia and New York City. Émile Baudot designed a system using a five unit code in 1874 that is still in use today. Teleprinter system design was gradually improved until, at the beginning of World War II, it represented the principal distribution method used by the news services. Radioteletype evolved from these earlier landline teleprinter operations. The US Department of the Navy successfully tested printing telegraphy between an airplane and ground radio station in August 1922. Later that year, the Radio Corporation of America successfully tested printing telegraphy via their Chatham, MA radio station to the R.M.S. Majestic. An early implementation of the Radioteletype was the Watsongraph, named after Detroit inventor Glenn Watson in March 1931. Commercial RTTY systems were in active service between San Francisco and Honolulu as early as April 1932 and between San Francisco and New York City by 1934. The US Military used radioteletype in the 1930s and expanded this usage during World War II. The Navy called radioteletype "RATT" (Radio Automatic Teletype) and the Army Signal Corps called radioteletype "SCRT", an abbreviation of Single-Channel Radio Teletype. The military used frequency shift keying technology and this technology proved very reliable even over long distances. From the 1980s, teleprinters were replaced by computers running teleprinter emulation software. A radioteletype station consists of three distinct parts: the Teletype or teleprinter, the modem and the radio. The Teletype or teleprinter is an electromechanical or electronic device. The word "Teletype" was a trademark of the Teletype Corporation, so the terms "TTY", "RTTY", "RATT" and "teleprinter" are usually used to describe a generic device without reference to a particular manufacturer. Electromechanical teleprinters were heavy, complex and noisy, and have been replaced with electronic units. The teleprinter includes a keyboard, which is the main means of entering text, and a printer or visual display unit (VDU). An alternative input device is a perforated tape reader and, more recently, computer storage media (such as floppy disks). Alternative output devices are tape perforators and computer storage media. The line output of a teleprinter can be at either digital logic levels (+5 V signifies a logical "1" or "mark" and 0 V signifies a logical "0" or "space") or line levels (−80 V signifies a "1" and +80 V a "0"). When no traffic is passed, the line idles at the "mark" state. When a key of the teleprinter keyboard is pressed, a 5-bit character is generated. The teleprinter converts it to serial format and transmits a sequence of a "start bit" (a logical 0 or space), then one after the other the 5 data bits, finishing with a "stop bit" (a logical 1 or mark, lasting 1, 1.5 or 2 bits). When a sequence of start bit, 5 data bits and stop bit arrives at the input of the teleprinter, it is converted to a 5-bit word and passed to the printer or VDU. With electromechanical teleprinters, these functions required complicated electromechanical devices, but they are easily implemented with standard digital electronics using shift registers. Special integrated circuits have been developed for this function, for example the Intersil 6402 and 6403. These are stand-alone UART devices, similar to computer serial port peripherals. The 5 data bits allow for only 32 different codes, which cannot accommodate the 26 letters, 10 figures, space, a few punctuation marks and the required control codes, such as carriage return, new line, bell, etc. To overcome this limitation, the teleprinter has two "states", the "unshifted" or "letters" state and the "shifted" or "numbers" or "figures" state. The change from one state to the other takes place when the special control codes "LETTERS" and "FIGURES" are sent from the keyboard or received from the line. In the "letters" state the teleprinter prints the letters and space while in the shifted state it prints the numerals and punctuation marks. Teleprinters for languages using other alphabets also use an additional "third shift" state, in which they print letters in the alternative alphabet. The modem is sometimes called the terminal unit and is an electronic device which is connected between the teleprinter and the radio transceiver. The transmitting part of the modem converts the digital signal transmitted by the teleprinter or tape reader to one or the other of a pair of audio frequency tones, traditionally 2295/2125 Hz (US) or 2125/1955 Hz (Europe). One of the tones corresponds to the "mark" condition and the other to the "space" condition. These audio tones, then, modulate an SSB transmitter to produce the final audio-frequency shift keying (AFSK) radio frequency signal. Some transmitters are capable of direct frequency-shift keying (FSK) as they can directly accept the digital signal and change their transmitting frequency according to the "mark" or "space" input state. In this case the transmitting part of the modem is bypassed. On reception, the FSK signal is converted to the original tones by mixing the FSK signal with a local oscillator called the BFO or "beat frequency oscillator". These tones are fed to the demodulator part of the modem, which processes them through a series of filters and detectors to recreate the original digital signal. The FSK signals are audible on a communications radio receiver equipped with a BFO, and have a distinctive "beedle-eeeedle-eedle-eee" sound, usually starting and ending on one of the two tones ("idle on mark"). The transmission speed is a characteristic of the teleprinter while the shift (the difference between the tones representing mark and space) is a characteristic of the modem. These two parameters are therefore independent, provided they have satisfied the minimum shift size for a given transmission speed. Electronic teleprinters can readily operate in a variety of speeds, but mechanical teleprinters require the change of gears in order to operate at different speeds. Today, both functions can be performed with modern computers equipped with digital signal processors or sound cards. The sound card performs the functions of the modem and the CPU performs the processing of the digital bits. This approach is very common in amateur radio, using specialized computer programs like fldigi, MMTTY or MixW. Before the computer mass storage era, most RTTY stations stored text on paper tape using paper tape punchers and readers. The operator would type the message on the TTY keyboard and punch the code onto the tape. The tape could then be transmitted at a steady, high rate, without typing errors. A tape could be reused, and in some cases - especially for use with ASCII on NC Machines - might be made of plastic or even very thin metal material in order to be reused many times. The most common test signal is a series of "RYRYRY" characters, as these form an alternating tone pattern exercising all bits and are easily recognized. Pangrams are also transmitted on RTTY circuits as test messages, the most common one being "The quick brown fox jumps over the lazy dog", and in French circuits, "Voyez le brick géant que j'examine près du wharf" The original (or "Baudot") radioteletype system is based almost invariably on the Baudot code or ITA-2 5 bit alphabet. The link is based on character asynchronous transmission with 1 start bit and 1, 1.5 or 2 stop bits. Transmitter modulation is normally FSK (F1B). Occasionally, an AFSK signal modulating an RF carrier (A2B, F2B) is used on VHF or UHF frequencies. Standard transmission speeds are 45.45, 50, 75, 100, 150 and 300 baud. Common carrier shifts are 85 Hz (used on LF and VLF frequencies), 170 Hz, 425 Hz, 450 Hz and 850 Hz, although some stations use non-standard shifts. There are variations of the standard Baudot alphabet to cover languages written in Cyrillic, Arabic, Greek etc., using special techniques. Some combinations of speed and shift are standardized for specific services using the original radioteletype system: After World War II, amateur radio operators in the US started to receive obsolete but usable Teletype Model 26 equipment from commercial operators with the understanding that this equipment would not be used for or returned to commercial service. "The Amateur Radioteletype and VHF Society" was founded in 1946 in Woodside, NY. This organization soon changed its name to "The VHF Teletype Society" and started US Amateur Radio operations on 2 meters using audio frequency shift keying (AFSK). The first two-way amateur radioteletype QSO of record took place in May 1946 between Dave Winters, W2AUF, Brooklyn, NY and W2BFD, John Evans Williams, Woodside Long Island, NY. On the west coast, amateur RTTY also started on 2 meters. Operation on 80 meters, 40 meters and the other High Frequency (HF) amateur radio bands was initially accomplished using make and break keying since frequency shift keying (FSK) was not yet authorized. In early 1949, the first American transcontinental two-way RTTY QSO was accomplished on 11 meters using AFSK between Tom McMullen (W1QVF) operating at W1AW and Johnny Agalsoff, W6PSW. The stations effected partial contact on January 30, 1949, and repeated more successfully on January 31. On February 1, 1949, the stations exchanged solid print congratulatory message traffic and rag-chewed. Earlier, on January 23, 1949, William T. Knott, W2QGH, Larchmont, NY, had been able to make rough copy of W6PSW's test transmissions. While QSOs could be accomplished, it was quickly realized that FSK was technically superior to make and break keying. Due to the efforts of Merrill Swan, W6AEE, of "The RTTY Society of Southern California" publisher of "RTTY" and Wayne Green, W2NSD, of "CQ Magazine", Amateur Radio operators successfully petitioned the U.S. Federal Communications Commission (FCC) to amend Part 12 of the Regulations, which was effective on February 20, 1953. The amended Regulations permitted FSK in the non-voice parts of the 80, 40 and 20 meter bands and also specified the use of single channel 60 words-per-minute five unit code corresponding to ITA2. A shift of 850 hertz plus or minus 50 hertz was specified. Amateur Radio operators also had to identify their station callsign at the beginning and the end of each transmission and at ten-minute intervals using International Morse code. Use of this wide shift proved to be a problem for Amateur Radio operations. Commercial operators had already discovered that narrow shift worked best on the HF bands. After investigation and a petition to the FCC, Part 12 was amended, in March 1956, to allow Amateur Radio Operators to use any shift that was less than 900 hertz. The FCC Notice of Proposed Rule Making (NPRM) that resulted in the authorization of Frequency Shift Keying (FSK) in the amateur high frequency (HF) bands responded to petitions by the American Radio Relay League (ARRL), the National Amateur Radio Council and Mr. Robert Weinstein. The NPRM specifically states this, and this information may be found in its entirety in the December 1951 Issue of QST. While the New RTTY Handbook gives ARRL no credit, it was published by CQ Magazine and its author was a CQ columnist (CQ generally opposed ARRL at that time). The first RTTY Contest was held by the RTTY Society of Southern California from October 31 to November 1, 1953. Named the RTTY Sweepstakes Contest, twenty nine participants exchanged messages that contained a serial number, originating station call, check or RST report of two or three numbers, ARRL section of originator, local time (0000-2400 preferred) and date. Example: NR 23 W0BP CK MINN 1325 FEB 15. By the late 1950s, the contest exchange was expanded to include band used. Example: NR 23 W0BP CK MINN 1325 FEB 15 FORTY METERS. The contest was scored as follows: one point for each message sent and received entirely by RTTY and one point for each message received and acknowledged by RTTY. The final score was computed by multiplying the total number of message points by the number of ARRL sections worked. Two stations could exchange messages again on a different band for added points, but the section multiplier did not increase when the same section was reworked on a different band. Each DXCC entity was counted as an additional ARRL section for RTTY multiplier credit. "RTTY", later named "RTTY Journal", also published the first listing of stations, mostly located in the continental US, that were interested in RTTY in 1956. Amateur Radio operators used this callbook information to contact other operators both inside and outside the United States. For example, the first recorded USA to New Zealand two-way RTTY QSO took place in 1956 between W0BP and ZL1WB. By the late 1950s, new organizations focused on amateur radioteletype started to appear. The "British Amateur Radio Teletype Group", BARTG, now known as the "British Amateur Radio Teledata Group" was formed in June 1959. The Florida RTTY Society was formed in September 1959. Amateur Radio operators outside of Canada and the United States began to acquire surplus teleprinter and receive permission to get on the air. The first recorded RTTY QSO in the UK occurred in September 1959 between G2UK and G3CQE. A few weeks later, G3CQE had the first G/VE RTTY QSO with VE7KX. This was quickly followed up by G3CQE QSOs with VK3KF and ZL3HJ. Information on how to acquire surplus teleprinter equipment continued to spread and before long it was possible to work all continents on RTTY. Amateur Radio operators used various equipment designs to get on the air using RTTY in the 1950s and 1960s. Amateurs used their existing receivers for RTTY operation but needed to add a terminal unit, sometimes called a demodulator, to convert the received audio signals to DC signals for the teleprinter. Most of the terminal unit equipment used for receiving RTTY signals was homebuilt, using designs published in amateur radio publications. These original designs can be divided into two classes of terminal units: audio-type and intermediate frequency converters. The audio-type converters proved to be more popular with amateur radio operators. The Twin City, W2JAV and W2PAT designs were examples of typical terminal units that were used into the middle 1960s. The late 1960s and early 1970s saw the emergence of terminal units designed by W6FFC, such as the TT/L, ST-3, ST-5, and ST-6. These designs were first published in "RTTY Journal" starting in September 1967 and ending in 1970. An adaptation of the W6FFC TT/L terminal unit was developed by Keith Petersen, W8SDZ, and it was first published in the RTTY Journal in September of 1967. The drafting of the schematic in the article was done by Ralph Leland, W8DLT. Amateur Radio operators needed to modify their transmitters to allow for HF RTTY operation. This was accomplished by adding a frequency shift keyer that used a diode to switch a capacitor in and out of the circuit, shifting the transmitter’s frequency in synchronism with the teleprinter signal changing from mark to space to mark. A very stable transmitter was required for RTTY. The typical frequency multiplication type transmitter that was popular in the 1950s and 1960s would be relatively stable on 80 meters but become progressively less stable on 40 meters, 20 meters and 15 meters. By the middle 1960s, transmitter designs were updated, mixing a crystal-controlled high frequency oscillator with a variable low frequency oscillator, resulting in better frequency stability across all Amateur Radio HF bands. During the early days of Amateur RTTY, the Worked All Continents – RTTY Award was conceived by the RTTY Society of Southern California and issued by RTTY Journal. The first Amateur Radio station to achieve this WAC – RTTY Award was VE7KX. The first stations recognized as having achieved single band WAC RTTY were W1MX (3.5 MHz); DL0TD (7.0 MHz); K3SWZ (14.0 MHz); W0MT (21.0 MHz) and FG7XT (28.0 MHz). The ARRL began issuing WAC RTTY certificates in 1969. By the early 1970s, Amateur Radio RTTY had spread around the world and it was finally possible to work more than 100 countries via RTTY. FG7XT was the first Amateur Radio station to claim to achieve this honor. However, Jean did not submit his QSL cards for independent review. ON4BX, in 1971, was the first Amateur Radio station to submit his cards to the DX Editor of RTTY Journal and to achieve this honor. The ARRL began issuing DXCC RTTY Awards on November 1, 1976. Prior to that date, an award for working 100 countries on RTTY was only available via RTTY Journal. In the 1950s through the 1970s, "RTTY art" was a popular on-air activity. This consisted of (sometimes very elaborate and artistic) pictures sent over rtty through the use of lengthy punched tape transmissions and then printed by the receiving station on paper. On January 7, 1972, the FCC amended Part 97 to allow faster RTTY speeds. Four standard RTTY speeds were authorized, namely, 60 (45 baud), 67 (50 baud), 75 (56.25 baud) and 100 (75 baud) words per minute. Many Amateur Radio operators had equipment that was capable of being upgraded to 75 and 100 words per minute by changing teleprinter gears. While there was an initial interest in 100 words per minute operation, many Amateur Radio operators moved back to 60 words per minute. Some of the reasons for the failure of 100 words per minute HF RTTY included poor operation of improperly maintained mechanical teleprinters, narrow bandwidth terminal units, continued use of 170 Hz shift at 100 words per minute and excessive error rates due to multipath distortion and the nature of ionospheric propagation. The FCC approved the use of ASCII by Amateur Radio stations on March 17, 1980 with speeds up to 300 baud from 3.5 to 21.25 MHz and 1200 baud between 28 and 225 MHz. Speeds up to 19.2 kilobaud was authorized on Amateur frequencies above 420 MHz. The requirement for Amateur Radio operators in the United States to identify their station callsign at the beginning and the end of each digital transmission and at ten-minute intervals using International Morse code was finally lifted by the FCC on June 15, 1983. RTTY has a typical baud rate for Amateur operation of 45.45 baud (approximately 60 words per minute). It remains popular as a "keyboard to keyboard" mode in Amateur Radio. RTTY has declined in commercial popularity as faster, more reliable alternative data modes have become available, using satellite or other connections. For its transmission speed, RTTY has low spectral efficiency. The typical RTTY signal with 170 Hz shift at 45.45 baud requires around 250 Hz receiver bandwidth, more than double that required by PSK31. In theory, at this baud rate, the shift size can be decreased to 22.725 Hz, reducing the overall band footprint substantially. Because RTTY, using either AFSK or FSK modulation, produces a waveform with constant power, a transmitter does not need to use a linear amplifier, which is required for many digital transmission modes. A more efficient Class C amplifier may be used. RTTY, using either AFSK or FSK modulation, is moderately resistant to vagaries of HF propagation and interference, however modern digital modes, such as MFSK, use Forward Error Correction to provide much better data reliability. Principally, the primary users are those who need robust shortwave communications. Examples are: One regular service transmitting RTTY meteorological information is the German Meteorological Service (Deutscher Wetterdienst or DWD). The DWD regularly transmit two programs on various frequencies on LF and HF in standard RTTY (ITA-2 alphabet). The list of callsigns, frequencies, baudrates and shifts are as follows: The DWD signals can be easily received in Europe, North Africa and parts of North America. RTTY (in English) may be spoken as "radioteletype", by its letters: R-T-T-Y, or simply as "ritty".
https://en.wikipedia.org/wiki?curid=26341
Register transfer language In computer science, register transfer language (RTL) is a kind of intermediate representation (IR) that is very close to assembly language, such as that which is used in a compiler. It is used to describe data flow at the register-transfer level of an architecture. Academic papers and textbooks often use a form of RTL as an architecture-neutral assembly language. RTL is used as the name of a specific intermediate representation in several compilers, including the GNU Compiler Collection (GCC), Zephyr, and the European compiler projects CerCo and CompCert. In GCC, RTL is generated from the GIMPLE representation, transformed by various passes in the GCC 'middle-end', and then converted to assembly language. GCC's RTL is usually written in a form which looks like a Lisp S-expression: This "side-effect expression" says "sum the contents of register 138 with the contents of register 139 and store the result in register 140". The SI specifies the access mode for each register. In the example it is "SImode", i.e. "access the register as 32-bit integer". The sequence of RTL generated has some dependency on the characteristics of the processor for which GCC is generating code. However, the meaning of the RTL is more-or-less independent of the target: it would usually be possible to read and understand a piece of RTL without knowing what processor it was generated for. Similarly, the meaning of the RTL doesn't usually depend on the original high-level language of the program. A register transfer language is a system for expressing in symbolic form the microoperation sequences among the registers of a digital module. It is a convenient tool for describing the internal organization of digital computers in concise and precise manner. It can also be used to facilitate the design process of digital systems. The idea behind RTL was first described in: Davidson and Fraser; The Design and Application of a Retargetable Peephole Optimizer; ToPLaS v2(2) 191-202 (April 1980)
https://en.wikipedia.org/wiki?curid=26344
Remote procedure call In distributed computing, a remote procedure call (RPC) is when a computer program causes a procedure (subroutine) to execute in a different address space (commonly on another computer on a shared network), which is coded as if it were a normal (local) procedure call, without the programmer explicitly coding the details for the remote interaction. That is, the programmer writes essentially the same code whether the subroutine is local to the executing program, or remote. This is a form of client–server interaction (caller is client, executor is server), typically implemented via a request–response message-passing system. In the object-oriented programming paradigm, RPCs are represented by remote method invocation (RMI). The RPC model implies a level of location transparency, namely that calling procedures are largely the same whether they are local or remote, but usually they are not identical, so local calls can be distinguished from remote calls. Remote calls are usually orders of magnitude slower and less reliable than local calls, so distinguishing them is important. RPCs are a form of inter-process communication (IPC), in that different processes have different address spaces: if on the same host machine, they have distinct virtual address spaces, even though the physical address space is the same; while if they are on different hosts, the physical address space is different. Many different (often incompatible) technologies have been used to implement the concept. Request–response protocols date to early distributed computing in the late 1960s, theoretical proposals of remote procedure calls as the model of network operations date to the 1970s, and practical implementations date to the early 1980s. Bruce Jay Nelson is generally credited with coining the term "remote procedure call" in 1981. Remote procedure calls used in modern operating systems trace their roots back to the RC 4000 multiprogramming system, which used a request-response communication protocol for process synchronization. The idea of treating network operations as remote procedure calls goes back at least to the 1970s in early ARPANET documents. In 1978, Per Brinch Hansen proposed Distributed Processes, a language for distributed computing based on "external requests" consisting of procedure calls between processes. One of the earliest practical implementations was in 1982 by Brian Randell and colleagues for their Newcastle Connection between UNIX machines. This was soon followed by "Lupine" by Andrew Birrell and Bruce Nelson in the Cedar environment at Xerox PARC. Lupine automatically generated stubs, providing type-safe bindings, and used an efficient protocol for communication. One of the first business uses of RPC was by Xerox under the name "Courier" in 1981. The first popular implementation of RPC on Unix was Sun's RPC (now called ONC RPC), used as the basis for Network File System (NFS). In the 1990s, with the popularity of object-oriented programming, an alternative model of remote method invocation (RMI) was widely implemented, such as in Common Object Request Broker Architecture (CORBA, 1991) and Java remote method invocation. RMIs in turn fell in popularity with the rise of the internet, particularly in the 2000s. RPC is a request–response protocol. An RPC is initiated by the "client", which sends a request message to a known remote "server" to execute a specified procedure with supplied parameters. The remote server sends a response to the client, and the application continues its process. While the server is processing the call, the client is blocked (it waits until the server has finished processing before resuming execution), unless the client sends an asynchronous request to the server, such as an XMLHttpRequest. There are many variations and subtleties in various implementations, resulting in a variety of different (incompatible) RPC protocols. An important difference between remote procedure calls and local calls is that remote calls can fail because of unpredictable network problems. Also, callers generally must deal with such failures without knowing whether the remote procedure was actually invoked. Idempotent procedures (those that have no additional effects if called more than once) are easily handled, but enough difficulties remain that code to call remote procedures is often confined to carefully written low-level subsystems. To let different clients access servers, a number of standardized RPC systems have been created. Most of these use an interface description language (IDL) to let various platforms call the RPC. The IDL files can then be used to generate code to interface between the client and servers. Notable RPC implementations and analogues include:
https://en.wikipedia.org/wiki?curid=26346
Russian submarine Kursk (K-141) K-141 "Kursk" (, transl. , meaning "Nuclear-powered submarine "Kursk"") was an Oscar II-class nuclear-powered cruise missile submarine of the Russian Navy. On 12 August 2000, K-141 "Kursk" was lost when it sank in the Barents Sea, killing all 118 personnel on board. K-141 "Kursk" was a Project 949A class "Antey" () submarine of the Oscar class, known as the Oscar II by its NATO reporting name, and was the penultimate submarine of the Oscar II class designed and approved in the Soviet Union. Construction began in 1990 at the Soviet Navy military shipyards in Severodvinsk, near Arkhangelsk, in the northern Russian SFSR. During the construction of K-141, the Soviet Union collapsed; work continued, and she became one of the first naval vessels completed after the collapse. K-141 was inherited by Russia and launched in 1994, before being commissioned by the Russian Navy on December 30, as part of the Russian Northern Fleet. K-141 was named "Kursk" after the city, following the naming system for Soviet submarines. "Kursk" was assigned to the home port of Vidyayevo, Murmansk Oblast. The Antey design represented the highest achievement of Soviet nuclear submarine technology. They are the second-largest cruise missile submarines ever built, after some ballistic missile submarines were converted to carry cruise missiles in 2007. It was built to defeat an entire United States aircraft carrier group. A single Type 65 torpedo carried a warhead powerful enough to sink an aircraft carrier. Both missiles and torpedoes could be equipped with nuclear warheads. She was longer than the preceding Oscar I-class of submarines. The senior officers had individual staterooms and the entire crew had access to a gymnasium. The outer hull, made of high-nickel, high-chromium stainless steel thick, had exceptionally good resistance to corrosion and a weak magnetic signature which helped prevent detection by U.S. magnetic anomaly detector (MAD) systems. There was a gap to the -thick steel pressure hull. She was designed to remain submerged for up to 120 days. The sail superstructure was reinforced to allow it to break through the Arctic ice. The submarine was armed with 24 SS-N-19/P-700 Granit cruise missiles, and eight torpedo tubes in the bow: four and four . The Granit missiles with a range of , were capable of supersonic flight at altitudes over . They were designed to swarm enemy vessels and intelligently choose individual targets which terminated with a dive onto the target. The torpedo tubes could be used to launch either torpedoes or anti-ship missiles with a range of . Her weapons included 18 SS-N-16 "Stallion" anti-submarine missiles. "Kursk" was part of Russia's Northern Fleet, which had suffered funding cutbacks throughout the 1990s. Many of its submarines were anchored and rusting in Zapadnaya Litsa Naval Base, from Murmansk. Little work to maintain all but the most essential front-line equipment, including search and rescue equipment, had occurred. Northern Fleet sailors had gone unpaid in the mid-1990s. During her five years of service, "Kursk" completed only one mission, a six-month deployment to the Mediterranean Sea during the summer of 1999 to monitor the United States Sixth Fleet responding to the Kosovo crisis. This was due to a lack of funds for fuel. As a result, many of her crew had spent little time at sea and were inexperienced. "Kursk" joined the "Summer-X" exercise, the first large-scale naval exercise planned by the Russian Navy in more than a decade, on 10 August 2000. It included 30 ships including the fleet's flagship "Pyotr Velikiy" ("Peter the Great"), four attack submarines, and a flotilla of smaller ships. The crew had recently won a citation for its excellent performance and been recognized as the best submarine crew in the Northern Fleet. While it was on an exercise, "Kursk" loaded a full complement of combat weapons. It was one of the few ships authorized to carry a combat load at all times. On the first day of the exercise, "Kursk" successfully launched a Granit missile armed with a dummy warhead. Two days later, on the morning of 12 August, "Kursk" prepared to fire dummy torpedoes at the ("Peter The Great"). These practice torpedoes had no explosive warheads and were manufactured and tested at a much lower quality standard. On 12 August 2000, at 11:28 local time (07:28 UTC), there was an explosion while preparing to fire. The Russian Navy's final report on the disaster concluded the explosion was due to the failure of one of "Kursk"'s hydrogen peroxide-fueled Type 65 torpedoes. A subsequent investigation concluded that high-test peroxide (HTP), a form of highly concentrated hydrogen peroxide used as propellant for the torpedo, seeped through a faulty weld in the torpedo casing. When HTP comes into contact with a catalyst, it rapidly expands by a factor of 5000, generating vast quantities of steam and oxygen. The pressure produced by the expanding HTP ruptured the kerosene fuel tank in the torpedo and set off an explosion equal to of TNT. The submarine sank in relatively shallow water, bottoming at about off Severomorsk, at . A second explosion 135 seconds after the initial event was equivalent to 3-7 tons of TNT. The explosions blew a large hole in the hull and caused the first three compartments of the submarine to collapse, killing or incapacitating all but 23 of the 118 personnel on board. The British and Norwegian navies offered assistance, but Russia initially refused all help. All 118 sailors and officers aboard "Kursk" died. The Russian Admiralty initially told the public that the majority of the crew died within minutes of the explosion, but on 21 August, Norwegian and Russian divers found 24 bodies in the ninth compartment, the turbine room at the stern of the boat. Captain-lieutenant Dmitri Kolesnikov wrote a note listing the names of 23 sailors who were alive in the compartment after the ship sank. "Kursk" carried a potassium superoxide cartridge of a chemical oxygen generator; these are used to absorb carbon dioxide and chemically release oxygen during an emergency. However, the cartridge became contaminated with sea water and the resulting chemical reaction caused a flash fire which consumed the available oxygen. The investigation showed that some men temporarily survived the fire by plunging under water, as fire marks on the bulkheads indicated the water was at waist level at the time. Ultimately, the remaining crew burned to death or suffocated. Russian President Vladimir Putin, though immediately informed of the tragedy, was told by the navy that they had the situation under control and that rescue was imminent. He waited for five days before he ended his holiday at a presidential resort in Sochi on the Black Sea. Putin was only four months into his tenure as President, and the public and media were extremely critical of his decision to remain at a seaside resort. His highly favourable ratings dropped dramatically. The President's response appeared callous and the government's actions looked incompetent. A year later he said, "I probably should have returned to Moscow, but nothing would have changed. I had the same level of communication both in Sochi and in Moscow, but from a PR point of view I could have demonstrated some special eagerness to return." A consortium formed by the Dutch companies Mammoet and Smit International was awarded a contract by Russia to raise the vessel, excluding the bow. They modified the barge "Giant 4" which raised "Kursk" and recovered the remains of the sailors. During salvage operations in 2001, the team first cut the bow off the hull using a tungsten carbide-studded cable. As this tool had the potential to cause sparks which could ignite remaining pockets of reactive gases, such as hydrogen, the operation was executed with care. Most of the bow was abandoned and the rest of the vessel was towed to Severomorsk and placed in a floating dry dock for analysis. The remains of "Kursk"s reactor compartment were towed to Sayda Bay on Russia's northern Kola Peninsula, where more than 50 reactor compartments were afloat at pier points, after a shipyard had removed all the fuel from the boat in early 2003. Some torpedo and torpedo tube fragments from the bow were recovered and the rest was destroyed by explosives in 2002. Notwithstanding the navy's oft-stated position that a collision with a foreign vessel had triggered the event, a report issued by the government attributed the disaster to a torpedo explosion caused when high-test peroxide (HTP), a form of highly concentrated hydrogen peroxide, leaked from a faulty weld in the torpedo's casing. The report found that the initial explosion destroyed the torpedo room compartment and killed everyone in the first compartment. The blast entered the second and perhaps the third and fourth compartments through an air conditioning vent. All of the 36 men in the command post located in the second compartment were immediately incapacitated by the blast wave and possibly killed. The first explosion caused a fire that raised the temperature of the compartment to more than . The heat caused the warheads of between five and seven additional torpedoes to detonate, creating an explosion equivalent to 2–3 tons of TNT that measured 4.2 on the Richter magnitude scale on seismographs across Europe and was detected as far away as Alaska. Vice-Admiral Valery Ryazantsev differed with the government's official conclusion. He cited inadequate training, poor maintenance, and incomplete inspections that caused the crew to mishandle the weapon. During the examination of the wrecked sub, investigators recovered a partially burned copy of the safety instructions for loading HTP torpedoes, but the instructions were for a significantly different type of torpedo and failed to include essential steps for testing an air valve. The 7th Division, 1st Submarine Flotilla never inspected "Kursk"s crew's qualifications and readiness to fire HTP torpedoes. "Kursk"s crew had no prior experience with HTP-powered torpedoes and had not been trained in handling or firing HTP-powered torpedoes. Due to their inexperience and lack of training, compounded by incomplete inspections and oversight, and because the "Kursk"s crew followed faulty instructions when loading the practice torpedo, Ryazantsev believes they set off a chain of events that led to the explosion.
https://en.wikipedia.org/wiki?curid=26347
Radiation therapy Radiation therapy or radiotherapy, often abbreviated RT, RTx, or XRT, is a therapy using ionizing radiation, generally as part of cancer treatment to control or kill malignant cells and normally delivered by a linear accelerator. Radiation therapy may be curative in a number of types of cancer if they are localized to one area of the body. It may also be used as part of adjuvant therapy, to prevent tumor recurrence after surgery to remove a primary malignant tumor (for example, early stages of breast cancer). Radiation therapy is synergistic with chemotherapy, and has been used before, during, and after chemotherapy in susceptible cancers. The subspecialty of oncology concerned with radiotherapy is called radiation oncology. Radiation therapy is commonly applied to the cancerous tumor because of its ability to control cell growth. Ionizing radiation works by damaging the DNA of cancerous tissue leading to cellular death. To spare normal tissues (such as skin or organs which radiation must pass through to treat the tumor), shaped radiation beams are aimed from several angles of exposure to intersect at the tumor, providing a much larger absorbed dose there than in the surrounding healthy tissue. Besides the tumour itself, the radiation fields may also include the draining lymph nodes if they are clinically or radiologically involved with the tumor, or if there is thought to be a risk of subclinical malignant spread. It is necessary to include a margin of normal tissue around the tumor to allow for uncertainties in daily set-up and internal tumor motion. These uncertainties can be caused by internal movement (for example, respiration and bladder filling) and movement of external skin marks relative to the tumor position. Radiation oncology is the medical specialty concerned with prescribing radiation, and is distinct from radiology, the use of radiation in medical imaging and diagnosis. Radiation may be prescribed by a radiation oncologist with intent to cure ("curative") or for adjuvant therapy. It may also be used as palliative treatment (where cure is not possible and the aim is for local disease control or symptomatic relief) or as therapeutic treatment (where the therapy has survival benefit and can be curative). It is also common to combine radiation therapy with surgery, chemotherapy, hormone therapy, immunotherapy or some mixture of the four. Most common cancer types can be treated with radiation therapy in some way. The precise treatment intent (curative, adjuvant, neoadjuvant therapeutic, or palliative) will depend on the tumor type, location, and stage, as well as the general health of the patient. Total body irradiation (TBI) is a radiation therapy technique used to prepare the body to receive a bone marrow transplant. Brachytherapy, in which a radioactive source is placed inside or next to the area requiring treatment, is another form of radiation therapy that minimizes exposure to healthy tissue during procedures to treat cancers of the breast, prostate and other organs. Radiation therapy has several applications in non-malignant conditions, such as the treatment of trigeminal neuralgia, acoustic neuromas, severe thyroid eye disease, pterygium, pigmented villonodular synovitis, and prevention of keloid scar growth, vascular restenosis, and heterotopic ossification. The use of radiation therapy in non-malignant conditions is limited partly by worries about the risk of radiation-induced cancers. Different cancers respond to radiation therapy in different ways. The response of a cancer to radiation is described by its radiosensitivity. Highly radiosensitive cancer cells are rapidly killed by modest doses of radiation. These include leukemias, most lymphomas and germ cell tumors. The majority of epithelial cancers are only moderately radiosensitive, and require a significantly higher dose of radiation (60-70 Gy) to achieve a radical cure. Some types of cancer are notably radioresistant, that is, much higher doses are required to produce a radical cure than may be safe in clinical practice. Renal cell cancer and melanoma are generally considered to be radioresistant but radiation therapy is still a palliative option for many patients with metastatic melanoma. Combining radiation therapy with immunotherapy is an active area of investigation and has shown some promise for melanoma and other cancers. It is important to distinguish the radiosensitivity of a particular tumor, which to some extent is a laboratory measure, from the radiation "curability" of a cancer in actual clinical practice. For example, leukemias are not generally curable with radiation therapy, because they are disseminated through the body. Lymphoma may be radically curable if it is localised to one area of the body. Similarly, many of the common, moderately radioresponsive tumors are routinely treated with curative doses of radiation therapy if they are at an early stage. For example, non-melanoma skin cancer, head and neck cancer, breast cancer, non-small cell lung cancer, cervical cancer, anal cancer, and prostate cancer. Metastatic cancers are generally incurable with radiation therapy because it is not possible to treat the whole body. Before treatment, a CT scan is often performed to identify the tumor and surrounding normal structures. The patient receives small skin marks to guide the placement of treatment fields. Patient positioning is crucial at this stage as the patient will have to be placed in an identical position during each treatment. Many patient positioning devices have been developed for this purpose, including masks and cushions which can be molded to the patient. The response of a tumor to radiation therapy is also related to its size. Due to complex radiobiology, very large tumors respond less well to radiation than smaller tumors or microscopic disease. Various strategies are used to overcome this effect. The most common technique is surgical resection prior to radiation therapy. This is most commonly seen in the treatment of breast cancer with wide local excision or mastectomy followed by adjuvant radiation therapy. Another method is to shrink the tumor with neoadjuvant chemotherapy prior to radical radiation therapy. A third technique is to enhance the radiosensitivity of the cancer by giving certain drugs during a course of radiation therapy. Examples of radiosensitizing drugs include Cisplatin, Nimorazole, and Cetuximab. The impact of radiotherapy varies between different types of cancer and different groups. For example, for breast cancer after breast-conserving surgery, radiotherapy has been found to halve the rate at which the disease recurs. Radiation therapy is in itself painless. Many low-dose palliative treatments (for example, radiation therapy to bony metastases) cause minimal or no side effects, although short-term pain flare-up can be experienced in the days following treatment due to oedema compressing nerves in the treated area. Higher doses can cause varying side effects during treatment (acute side effects), in the months or years following treatment (long-term side effects), or after re-treatment (cumulative side effects). The nature, severity, and longevity of side effects depends on the organs that receive the radiation, the treatment itself (type of radiation, dose, fractionation, concurrent chemotherapy), and the patient. Most side effects are predictable and expected. Side effects from radiation are usually limited to the area of the patient's body that is under treatment. Side effects are dose- dependent; for example higher doses of head and neck radiation can be associated with cardiovascular complications, thyroid dysfunction, and pituitary axis dysfunction. Modern radiation therapy aims to reduce side effects to a minimum and to help the patient understand and deal with side effects that are unavoidable. The main side effects reported are fatigue and skin irritation, like a mild to moderate sun burn. The fatigue often sets in during the middle of a course of treatment and can last for weeks after treatment ends. The irritated skin will heal, but may not be as elastic as it was before. Late side effects occur months to years after treatment and are generally limited to the area that has been treated. They are often due to damage of blood vessels and connective tissue cells. Many late effects are reduced by fractionating treatment into smaller parts. Cumulative effects from this process should not be confused with long-term effects—when short-term effects have disappeared and long-term effects are subclinical, reirradiation can still be problematic. These doses are calculated by the radiation oncologist and many factors are taken into account before the subsequent radiation takes place. During the first two weeks after fertilization, radiation therapy is lethal but not teratogenic. High doses of radiation during pregnancy induce anomalies, impaired growth and intellectual disability, and there may be an increased risk of childhood leukemia and other tumours in the offspring. In males previously having undergone radiotherapy, there appears to be no increase in genetic defects or congenital malformations in their children conceived after therapy. However, the use of assisted reproductive technologies and micromanipulation techniques might increase this risk. Hypopituitarism commonly develops after radiation therapy for sellar and parasellar neoplasms, extrasellar brain tumours, head and neck tumours, and following whole body irradiation for systemic malignancies. Radiation-induced hypopituitarism mainly affects growth hormone and gonadal hormones. In contrast, adrenocorticotrophic hormone (ACTH) and thyroid stimulating hormone (TSH) deficiencies are the least common among people with radiation-induced hypopituitarism. Changes in prolactin-secretion is usually mild, and vasopressin deficiency appears to be very rare as a consequence of radiation. There are rigorous procedures in place to minimise the risk of accidental overexposure of radiation therapy to patients. However, mistakes do occasionally occur; for example, the radiation therapy machine Therac-25 was responsible for at least six accidents between 1985 and 1987, where patients were given up to one hundred times the intended dose; two people were killed directly by the radiation overdoses. From 2005 to 2010, a hospital in Missouri overexposed 76 patients (most with brain cancer) during a five-year period because new radiation equipment had been set up incorrectly. Although medical errors are exceptionally rare, radiation oncologists, medical physicists and other members of the radiation therapy treatment team are working to eliminate them. ASTRO has launched a safety initiative called Target Safely that, among other things, aims to record errors nationwide so that doctors can learn from each and every mistake and prevent them from happening. ASTRO also publishes a list of questions for patients to ask their doctors about radiation safety to ensure every treatment is as safe as possible. Radiation therapy is used to treat early stage Dupuytren's disease and Ledderhose disease. When Dupuytren's disease is at the nodules and cords stage or fingers are at a minimal deformation stage of less than 10 degrees, then radiation therapy is used to prevent further progress of the disease. Radiation therapy is also used post surgery in some cases to prevent the disease continuing to progress. Low doses of radiation are used typically three gray of radiation for five days, with a break of three months followed by another phase of three gray of radiation for five days. Radiation therapy works by damaging the DNA of cancerous cells. This DNA damage is caused by one of two types of energy, photon or charged particle. This damage is either direct or indirect ionization of the atoms which make up the DNA chain. Indirect ionization happens as a result of the ionization of water, forming free radicals, notably hydroxyl radicals, which then damage the DNA. In photon therapy, most of the radiation effect is through free radicals. Cells have mechanisms for repairing single-strand DNA damage and double-stranded DNA damage. However, double-stranded DNA breaks are much more difficult to repair, and can lead to dramatic chromosomal abnormalities and genetic deletions. Targeting double-stranded breaks increases the probability that cells will undergo cell death. Cancer cells are generally less differentiated and more stem cell-like; they reproduce more than most healthy differentiated cells, and have a diminished ability to repair sub-lethal damage. Single-strand DNA damage is then passed on through cell division; damage to the cancer cells' DNA accumulates, causing them to die or reproduce more slowly. One of the major limitations of photon radiation therapy is that the cells of solid tumors become deficient in oxygen. Solid tumors can outgrow their blood supply, causing a low-oxygen state known as hypoxia. Oxygen is a potent radiosensitizer, increasing the effectiveness of a given dose of radiation by forming DNA-damaging free radicals. Tumor cells in a hypoxic environment may be as much as 2 to 3 times more resistant to radiation damage than those in a normal oxygen environment. Much research has been devoted to overcoming hypoxia including the use of high pressure oxygen tanks, hyperthermia therapy (heat therapy which dilates blood vessels to the tumor site), blood substitutes that carry increased oxygen, hypoxic cell radiosensitizer drugs such as misonidazole and metronidazole, and hypoxic cytotoxins (tissue poisons), such as tirapazamine. Newer research approaches are currently being studied, including preclinical and clinical investigations into the use of an oxygen diffusion-enhancing compound such as trans sodium crocetinate (TSC) as a radiosensitizer. Charged particles such as protons and boron, carbon, and neon ions can cause direct damage to cancer cell DNA through high-LET (linear energy transfer) and have an antitumor effect independent of tumor oxygen supply because these particles act mostly via direct energy transfer usually causing double-stranded DNA breaks. Due to their relatively large mass, protons and other charged particles have little lateral side scatter in the tissue—the beam does not broaden much, stays focused on the tumor shape, and delivers small dose side-effects to surrounding tissue. They also more precisely target the tumor using the Bragg peak effect. See proton therapy for a good example of the different effects of intensity-modulated radiation therapy (IMRT) vs. charged particle therapy. This procedure reduces damage to healthy tissue between the charged particle radiation source and the tumor and sets a finite range for tissue damage after the tumor has been reached. In contrast, IMRT's use of uncharged particles causes its energy to damage healthy cells when it exits the body. This exiting damage is not therapeutic, can increase treatment side effects, and increases the probability of secondary cancer induction. This difference is very important in cases where the close proximity of other organs makes any stray ionization very damaging (example: head and neck cancers). This x-ray exposure is especially bad for children, due to their growing bodies, and they have a 30% chance of a second malignancy after 5 years post initial RT. The amount of radiation used in photon radiation therapy is measured in grays (Gy), and varies depending on the type and stage of cancer being treated. For curative cases, the typical dose for a solid epithelial tumor ranges from 60 to 80 Gy, while lymphomas are treated with 20 to 40 Gy. Preventive (adjuvant) doses are typically around 45–60 Gy in 1.8–2 Gy fractions (for breast, head, and neck cancers.) Many other factors are considered by radiation oncologists when selecting a dose, including whether the patient is receiving chemotherapy, patient comorbidities, whether radiation therapy is being administered before or after surgery, and the degree of success of surgery. Delivery parameters of a prescribed dose are determined during treatment planning (part of dosimetry). Treatment planning is generally performed on dedicated computers using specialized treatment planning software. Depending on the radiation delivery method, several angles or sources may be used to sum to the total necessary dose. The planner will try to design a plan that delivers a uniform prescription dose to the tumor and minimizes dose to surrounding healthy tissues. In radiation therapy, three-dimensional dose distributions may be evaluated using the dosimetry technique known as gel dosimetry. The total dose is fractionated (spread out over time) for several important reasons. Fractionation allows normal cells time to recover, while tumor cells are generally less efficient in repair between fractions. Fractionation also allows tumor cells that were in a relatively radio-resistant phase of the cell cycle during one treatment to cycle into a sensitive phase of the cycle before the next fraction is given. Similarly, tumor cells that were chronically or acutely hypoxic (and therefore more radioresistant) may reoxygenate between fractions, improving the tumor cell kill. Fractionation regimens are individualised between different radiation therapy centers and even between individual doctors. In North America, Australia, and Europe, the typical fractionation schedule for adults is 1.8 to 2 Gy per day, five days a week. In some cancer types, prolongation of the fraction schedule over too long can allow for the tumor to begin repopulating, and for these tumor types, including head-and-neck and cervical squamous cell cancers, radiation treatment is preferably completed within a certain amount of time. For children, a typical fraction size may be 1.5 to 1.8 Gy per day, as smaller fraction sizes are associated with reduced incidence and severity of late-onset side effects in normal tissues. In some cases, two fractions per day are used near the end of a course of treatment. This schedule, known as a concomitant boost regimen or hyperfractionation, is used on tumors that regenerate more quickly when they are smaller. In particular, tumors in the head-and-neck demonstrate this behavior. Patients receiving palliative radiation to treat uncomplicated painful bone metastasis should not receive more than a single fraction of radiation.
https://en.wikipedia.org/wiki?curid=26350
Ronald Coase Ronald Harry Coase (; 29 December 1910 – 2 September 2013) was a British economist and author. He was the Clifton R. Musser Professor of Economics at the University of Chicago Law School, where he arrived in 1964 and remained for the rest of his life. He received the Nobel Memorial Prize in Economic Sciences in 1991. Coase, who believed economists should study real markets and not theoretical ones, established the case for the corporation as a means to pay the costs of operating a marketplace. Coase is best known for two articles in particular: "The Nature of the Firm" (1937), which introduces the concept of transaction costs to explain the nature and limits of firms; and "The Problem of Social Cost" (1960), which suggests that well-defined property rights could overcome the problems of externalities (see Coase theorem). Additionally, Coase's transaction costs approach is currently influential in modern organizational economics, where it was reintroduced by Oliver E. Williamson. Ronald Harry Coase was born in Willesden, a suburb of London, on 29 December 1910. His father, Henry Joseph Coase (1884–1973) was a telegraphist for the post office, as was his mother, Rosalie Elizabeth Coase (née Giles; 1882–1972), before marriage. As a child, Coase had a weakness in his legs, for which he was required to wear leg-irons. Due to this problem, he attended the school for physical defectives. At the age of 12, he was able to enter Kilburn Grammar School on scholarship. At Kilburn, he studied for the intermediate examination of the University of London as an external student in 1927–29. Coase married Marion Ruth Hartung of Chicago, Illinois in Willesden, England, 7 August 1937. Although they were unable to have children, they were married 75 years until her death in 2012, making him one of the longest-married Nobel Prize laureates. Coase attended the London School of Economics, where he took courses with Arnold Plant and received a bachelor of commerce degree in 1932. During his undergraduate studies, Coase received the Sir Ernest Cassel Travelling Scholarship, awarded by the University of London. He used this to visit the University of Chicago in 1931–1932 and studied with Frank Knight and Jacob Viner. Coase's colleagues would later admit that they did not remember this first visit. Between 1932–34, Coase was an assistant lecturer at the Dundee School of Economics and Commerce, which later became part of the University of Dundee. Subsequently, Coase was an assistant lecturer in commerce at the University of Liverpool between 1934–1935 before returning to London School of Economics as a member of staff until 1951. He then started to work at the University at Buffalo and retained his British citizenship after moving to the United States in the 1950s. In 1958, he moved to the University of Virginia. Coase settled at the University of Chicago in 1964 and became the co-editor of the "Journal of Law and Economics" with Aaron Director. He was also for a time a trustee of the Philadelphia Society. He received the Nobel Prize in Economics in 1991. Nearing his 100th birthday, Coase was working on a book concerning the rise of the economies of China and Vietnam. In an interview, Coase explained the mission of the Coase China Society and his vision of economics and the part to be played by Chinese economists. This became "How China Became Capitalist" (2012) co-authored with Ning Wang. Coase was honoured and received an honorary doctorate from the University at Buffalo Department of Economics in May 2012. Coase died in Chicago on 2 September 2013 at the age of 102. His wife had died on 17 October 2012. He was praised across the political spectrum, with Slate Magazine calling him "one of the most distinguished economists in the world" and "Forbes" magazine calling him "the greatest of the many great University of Chicago economists". The Washington Post called his work over eight decades "impossible to summarize" while recommending five of his papers to read. In "The Nature of the Firm" (1937), a brief but highly influential essay, Coase attempts to explain why the economy features a number of business firms instead of consisting exclusively of a multitude of independent, self-employed people who contract with one another. Given that "production could be carried on without any organization [that is, firms] at all", Coase asks, why and under what conditions should we expect firms to emerge? Since modern firms can only emerge when an entrepreneur of some sort begins to hire people, Coase's analysis proceeds by considering the conditions under which it makes sense for an entrepreneur to seek hired help instead of contracting out for some particular task. The traditional economic theory of the time (in the tradition of Adam Smith) suggested that, because the market is "efficient" (that is, those who are best at providing each good or service most cheaply are already doing so), it should always be cheaper to contract out than to hire. Coase noted, however, a number of transaction costs involved in using the market; the cost of obtaining a good or service via the market actually exceeds the price of the good. Other costs, including search and information costs, bargaining costs, keeping trade secrets, and policing and enforcement costs, can all potentially add to the cost of procuring something from another party. This suggests that firms will arise which can internalise the production of goods and services required to deliver a product, thus avoiding these costs. This argument sets the stage for the later contributions by Oliver Williamson: markets and hierarchies are alternative co-ordination mechanisms for economic transactions. There is a natural limit to what a firm can produce internally, however. Coase notices "decreasing returns to the entrepreneur function", including increasing overhead costs and increasing propensity for an overwhelmed manager to make mistakes in resource allocation. These factors become countervailing costs to the use of the firm. Coase argues that the size of a firm (as measured by how many contractual relations are "internal" to the firm and how many "external") is a result of finding an optimal balance between the competing tendencies of the costs outlined above. In general, making the firm larger will initially be advantageous, but the decreasing returns indicated above will eventually kick in, preventing the firm from growing indefinitely. Other things being equal, therefore, a firm will tend to be larger: The first two costs will increase with the spatial distribution of the transactions organised and the dissimilarity of the transactions. This explains why firms tend to either be in different geographic locations or to perform different functions. Additionally, technology changes that mitigate the cost of organising transactions across space may allow firms to become larger – the advent of the telephone and of cheap air travel, for example, would be expected to increase the size of firms. A further exploration of the dichotomy between markets and hierarchies as co-ordination mechanisms for economic transactions, derived a third alternative way called Commons based peer production, in which individuals successfully collaborate on large-scale projects following a diverse cluster of motivational drives and social signals. Upon publishing his article The Federal Communications Commission in 1959, Coase received negative feedback from the faculty at the University of Chicago over his conclusions and apparent conflicts with A. C. Pigou. According to Coase, "What I said was thought to run counter to Pigou's analysis by a number of economists at the University of Chicago and was therefore, according to them, wrong. At a meeting in Chicago I was able to convince these economists that I was right and Pigou's analysis faulty." Coase had presented his paper in 1960 during a seminar in Chicago, to twenty senior economist including George Stigler and Milton Friedman. He gradually won over the usually skeptic audience, in what has later been considered a "paradigm-shifting moment" in the genesis of Chicago Law and Economics. Coase would join the Chicago faculty four years later. Published in the "Journal of Law and Economics" in 1960, while Coase was a member of the Economics department at the University of Virginia, "The Problem of Social Cost" provided the key insight that it is unclear where the blame for externalities lies. The example he gave was of a rancher whose cattle stray onto the cropland of his neighbour. If the rancher is made to restrict his cattle, he is harmed just as the farmer is if the cattle remain unrestrained. Coase argued that without transaction costs the initial assignment of property rights makes no difference to whether or not the farmer and rancher can achieve the economically efficient outcome. If the cost of restraining cattle by, say, building a fence, is less than the cost of crop damage, the fence will be built. The initial assignment of property rights determines who builds the fence. If the farmer is responsible for the crop damage, the farmer will pay for the fence (as long the fence costs less than the crop damage). The allocation of property rights is primarily an equity issue, with consequences for the distribution of income and wealth, rather than an efficiency issue. With sufficient transaction costs, initial property rights matter for both equity and efficiency. From the point of view of economic efficiency, property rights should be assigned such that the owner of the rights wants to take the economically efficient action. To elaborate, if it is efficient not to restrict the cattle, the rancher should be given the rights (so that cattle can move about freely), whereas if it is efficient to restrict the cattle, the farmer should be given the rights over the movement of the cattle (so the cattle are restricted). This seminal argument forms the basis of the famous Coase theorem as labelled by Stigler. Though trained as an economist, Coase spent much of his career working in a law school. He is a central figure in the development of the subfield of law and economics. He viewed law and economics as having two parts, the first "using the economists' approach and concepts to analyze the working of the legal system, often called the economic analysis of the law"; and the second "a study of the influence of the legal system on the working of the economic system." Coase said that the second part "is the part of law and economics in which I am most interested." In his Simons Lecture celebrating the centennial of the University of Chicago, titled "Law and Economics at Chicago", Coase noted that he only accidentally wandered into the field: Despite wandering accidentally into law and economics, the opportunity to edit the Journal of Law and Economics was instrumental in bringing him to the University of Chicago: Coase believed that the University of Chicago was the intellectual center of law and economics. He concluded his Simons lecture by stating: I am very much aware that, in concentrating in this lecture on law and economics at Chicago, I have neglected other significant contributions to the subject made elsewhere such as those by Guido Calabresi at Yale, by Donald Turner at Harvard, and by others. But it can hardly be denied that in the emergence of the subject of law and economics, Chicago has played a very significant part and one of which the University can be proud. Another important contribution of Coase is the Coase conjecture, which states that an informal argument that durable-goods monopolists do not have market power because they are unable to commit to not lowering their prices in future periods. When asked what he considered his politics to be, Coase stated, I really don't know. I don't reject any policy without considering what its results are. If someone says there's going to be regulation, I don't say that regulation will be bad. Let's see. What we discover is that most regulation does produce, or has produced in recent times, a worse result. But I wouldn't like to say that all regulation would have this effect because one can think of circumstances in which it doesn't. Coase admitted that early in life, he aligned himself with socialism. Guido Calabresi wrote that Coase's focus on transaction costs in "The Nature of the Firm" was the result of his socialist beliefs. Reflecting on this, Coase wrote: "It is very difficult to know where one's ideas come from but for all I know he may well be right." Coase continued: Coase was research advisor to the Ronald Coase Institute, an organisation that promotes research on institutions and organizations – the laws, rules, customs, and norms – that govern real economic systems, with particular support for young scholars from developing and transitional countries. The University of Chicago Law School carries on the legacy of Ronald Coase through the mission of the Coase-Sandor Institute for Law and Economics. Each year, the University of Chicago Law School hosts the Coase Lecture, which was delivered in 2003 by Ronald Coase himself.
https://en.wikipedia.org/wiki?curid=26354
Richard R. Ernst Richard Robert Ernst (born 14 August 1933) is a Swiss physical chemist and Nobel Laureate. Ernst was awarded the Nobel Prize in Chemistry in 1991 for his contributions towards the development of Fourier transform Nuclear Magnetic Resonance (NMR) spectroscopy while at Varian Associates, Palo Alto and the subsequent development of multi-dimensional NMR techniques. These underpin applications to both to chemistry with NMR spectroscopy and to medicine with Magnetic Resonance Imaging (MRI). Richard R. Ernst was born in 1933 in Winterthur, Switzerland. He received both his diploma in chemistry in 1957 and his Ph.D. in physical chemistry in 1962 from ETH Zurich. His dissertation was on nuclear magnetic resonance in the field of physical chemistry. Ernst entered Varian Associates as a scientist in 1963 and invented Fourier transform NMR, noise decoupling, and a number of other methods. He returned to ETH Zurich in 1968 and became a lecturer. His career developed to Assistant Professor in 1970, Associate Professor in 1972. Since 1976, Richard R. Ernst was Full Professor of Physical Chemistry. He led a research group dedicated to magnetic resonance spectroscopy, was the director of the Physical Chemistry Laboratory at the ETH Zurich. He developed two-dimensional NMR and several novel pulse techniques. He retired in 1998. He participated in the development of medical magnetic resonance tomography, as well as the NMR structure determination of biopolymers in solution collaborating with Professor Kurt Wüthrich. He also participated in the study of intra-molecular dynamics. Ernst is a foreign fellow of the Estonian Academy of Sciences (elected 2002), the US National Academy of Sciences, the Royal Academy of Sciences, London, the German National Academy of Sciences Leopoldina, the Russian Academy of Sciences, the Korean Academy of Science and Technology and Bangladesh Academy of Sciences. He was elected a Foreign Member of the Royal Society (ForMemRS) in 1993. He was awarded the John Gamble Kirkwood Medal in 1989. The Nobel Prize in Chemistry 1991 was awarded to Richard R. Ernst "for his contributions to the development of the methodology of high resolution nuclear magnetic resonance (NMR) spectroscopy" A strong proponent of Ernst's nomination was the long-time Danish colleague and member of the Nobel Committee Professor Børge Bak. He holds Honorary Doctorates from the Technical University of Munich, ETH Lausanne, University of Zurich, University Antwerpen, Babes-Bolyai University, and University Montpellier. Ernst is member of the World Knowledge Dialogue Scientific Board. Ernst was awarded the Marcel Benoist Prize in 1986, the Wolf Prize for Chemistry in 1991 and Louisa Gross Horwitz Prize of Columbia University in 1991. He was also awarded the Tadeus Reichstein Medal in 2000 and the Order of the Star of Romania in 2004. The 2009 Bel Air Film Festival featured the world premiere of a documentary film on Ernst "Science Plus Dharma Equals Social Responsibility". Produced by Carlo Burton, the film takes place in Ernst's hometown in Switzerland.
https://en.wikipedia.org/wiki?curid=26361
RIPEMD RIPEMD (RIPE Message Digest) is a family of cryptographic hash functions developed in 1992 (the original RIPEMD) and 1996 (other variants). There are five functions in the family: RIPEMD, RIPEMD-128, RIPEMD-160, RIPEMD-256, and RIPEMD-320, of which RIPEMD-160 is the most common. The original RIPEMD, as well as RIPEMD-128, is not considered secure because 128-bit result is too small and also (for the original RIPEMD) because of design weaknesses. The 256- and 320-bit versions of RIPEMD provide the same level of security as RIPEMD-128 and RIPEMD-160, respectively; they are designed for applications where the security level is sufficient but longer hash result is necessary. While RIPEMD functions are less popular than SHA-1 and SHA-2, they are used, among others, in Bitcoin and other cryptocurrencies based on Bitcoin. The original RIPEMD function was designed in the framework of the EU project RIPE (RACE Integrity Primitives Evaluation) in 1992. Its design was based on the MD4 hash function. In 1996, in response to security weaknesses found in the original RIPEMD, Hans Dobbertin, Antoon Bosselaers and Bart Preneel at the COSIC research group at the Katholieke Universiteit Leuven in Leuven, Belgium published four strengthened variants: RIPEMD-128, RIPEMD-160, RIPEMD-256, and RIPEMD-320. In August 2004, a collision was reported for the original RIPEMD. This does not apply to RIPEMD-160. The 160-bit RIPEMD-160 hashes (also termed RIPE "message digests") are typically represented as 40-digit hexadecimal numbers. The following demonstrates a 43-byte ASCII input and the corresponding RIPEMD-160 hash: RIPEMD-160 behaves with the desired avalanche effect of cryptographic hash functions (small changes, e.g. changing d to c, result in a completely different hash): The hash of a zero-length string is: Below is a list of cryptography libraries that support RIPEMD (specifically RIPEMD-160):
https://en.wikipedia.org/wiki?curid=26363
Roman law Roman law is the legal system of ancient Rome, including the legal developments spanning over a thousand years of jurisprudence, from the Twelve Tables (c. 449 BC), to the "Corpus Juris Civilis" (AD 529) ordered by Eastern Roman Emperor Justinian I. Roman law forms the basic framework for civil law, the most widely used legal system today, and the terms are sometimes used synonymously. The historical importance of Roman law is reflected by the continued use of Latin legal terminology in many legal systems influenced by it, including common law. After the dissolution of the Western Roman Empire, the Roman law remained in effect in the Eastern Roman Empire. From the 7th century onward, the legal language in the East was Greek. "Roman law" also denoted the legal system applied in most of Western Europe until the end of the 18th century. In Germany, Roman law practice remained in place longer under the Holy Roman Empire (963–1806). Roman law thus served as a basis for legal practice throughout Western continental Europe, as well as in most former colonies of these European nations, including Latin America, and also in Ethiopia. English and Anglo-American common law were influenced also by Roman law, notably in their Latinate legal glossary (for example, "stare decisis", "culpa in contrahendo", "pacta sunt servanda"). Eastern Europe was also influenced by the jurisprudence of the "Corpus Juris Civilis", especially in countries such as medieval Romania (Wallachia, Moldavia, and some other medieval provinces/historical regions) which created a new system, a mixture of Roman and local law. Also, Eastern European law was influenced by the "Farmer's Law" of the medieval Byzantine legal system. Before the Twelve Tables (754–449 BC), private law comprised the Roman civil law ("ius civile Quiritium") that applied only to Roman citizens, and was bonded to religion; undeveloped, with attributes of strict formalism, symbolism, and conservatism, e.g. the ritual practice of mancipatio (a form of sale). The jurist Sextus Pomponius said, "At the beginning of our city, the people began their first activities without any fixed law, and without any fixed rights: all things were ruled despotically, by kings". It is believed that Roman Law is rooted in the Etruscan religion, emphasizing ritual. The first legal text is the Law of the Twelve Tables, dating from the mid-5th century BC. The plebeian tribune, C. Terentilius Arsa, proposed that the law should be written in order to prevent magistrates from applying the law arbitrarily. After eight years of political struggle, the plebeian social class convinced the patricians to send a delegation to Athens to copy the Laws of Solon; they also dispatched delegations to other Greek cities for like reason. In 451 BC, according to the traditional story (as Livy tells it), ten Roman citizens were chosen to record the laws ("decemviri legibus scribundis"). While they were performing this task, they were given supreme political power ("imperium"), whereas the power of the magistrates was restricted. In 450 BC, the "decemviri" produced the laws on ten tablets ("tabulae"), but these laws were regarded as unsatisfactory by the plebeians. A second decemvirate is said to have added two further tablets in 449 BC. The new Law of the Twelve Tables was approved by the people's assembly. Modern scholars tend to challenge the accuracy of Roman historians. They generally do not believe that a second decemvirate ever took place. The decemvirate of 451 is believed to have included the most controversial points of customary law, and to have assumed the leading functions in Rome. Furthermore, questions concerning Greek influence on early Roman Law are still much discussed. Many scholars consider it unlikely that the patricians sent an official delegation to Greece, as the Roman historians believed. Instead, those scholars suggest, the Romans acquired Greek legislations from the Greek cities of Magna Graecia, the main portal between the Roman and Greek worlds. The original text of the Twelve Tables has not been preserved. The tablets were probably destroyed when Rome was conquered and burned by the Gauls in 387 BC. The fragments which did survive show that it was not a law code in the modern sense. It did not provide a complete and coherent system of all applicable rules or give legal solutions for all possible cases. Rather, the tables contained specific provisions designed to change the then-existing customary law. Although the provisions pertain to all areas of law, the largest part is dedicated to private law and civil procedure. Many laws include "Lex Canuleia" (445 BC; which allowed the marriage—"ius connubii"—between patricians and plebeians), "Leges Licinae Sextiae" (367 BC; which made restrictions on possession of public lands—"ager publicus"—and also made sure that one of the consuls was plebeian), "Lex Ogulnia" (300 BC; plebeians received access to priest posts), and "Lex Hortensia" (287 BC; verdicts of plebeian assemblies—"plebiscita"—now bind all people). Another important statute from the Republican era is the "Lex Aquilia" of 286 BC, which may be regarded as the root of modern tort law. However, Rome's most important contribution to European legal culture was not the enactment of well-drafted statutes, but the emergence of a class of professional jurists ("prudentes", sing. "prudens", or "jurisprudentes") and of a legal science. This was achieved in a gradual process of applying the scientific methods of Greek philosophy to the subject of law, a subject which the Greeks themselves never treated as a science. Traditionally, the origins of Roman legal science are connected to Gnaeus Flavius. Flavius is said to have published around the year 300 BC the formularies containing the words which had to be spoken in court to begin a legal action. Before the time of Flavius, these formularies are said to have been secret and known only to the priests. Their publication made it possible for non-priests to explore the meaning of these legal texts. Whether or not this story is credible, jurists were active and legal treatises were written in larger numbers before the 2nd century BC. Among the famous jurists of the republican period are Quintus Mucius Scaevola who wrote a voluminous treatise on all aspects of the law, which was very influential in later times, and Servius Sulpicius Rufus, a friend of Marcus Tullius Cicero. Thus, Rome had developed a very sophisticated legal system and a refined legal culture when the Roman republic was replaced by the monarchical system of the principate in 27 BC. In the period between about 201 to 27 BC, we can see the development of more flexible laws to match the needs of the time. In addition to the old and formal "ius civile" a new juridical class is created: the "ius honorarium", which can be defined as "The law introduced by the magistrates who had the right to promulgate edicts in order to support, supplement or correct the existing law." With this new law the old formalism is being abandoned and new more flexible principles of "ius gentium" are used. The adaptation of law to new needs was given over to juridical practice, to magistrates, and especially to the praetors. A praetor was not a legislator and did not technically create new law when he issued his edicts ("magistratuum edicta"). In fact, the results of his rulings enjoyed legal protection ("actionem dare") and were in effect often the source of new legal rules. A Praetor's successor was not bound by the edicts of his predecessor; however, he did take rules from edicts of his predecessor that had proved to be useful. In this way a constant content was created that proceeded from edict to edict ("edictum traslatitium"). Thus, over the course of time, parallel to the civil law and supplementing and correcting it, a new body of praetoric law emerged. In fact, praetoric law was so defined by the famous Roman jurist Papinian (142–212 AD): ""Ius praetorium est quod praetores introduxerunt adiuvandi vel supplendi vel corrigendi iuris civilis gratia propter utilitatem publicam"" ("praetoric law is that law introduced by praetors to supplement or correct civil law for public benefit"). Ultimately, civil law and praetoric law were fused in the "Corpus Juris Civilis". The first 250 years of the current era are the period during which Roman law and Roman legal science reached its greatest degree of sophistication. The law of this period is often referred to as the "classical period of Roman law". The literary and practical achievements of the jurists of this period gave Roman law its unique shape. The jurists worked in different functions: They gave legal opinions at the request of private parties. They advised the magistrates who were entrusted with the administration of justice, most importantly the praetors. They helped the praetors draft their edicts, in which they publicly announced at the beginning of their tenure, how they would handle their duties, and the formularies, according to which specific proceedings were conducted. Some jurists also held high judicial and administrative offices themselves. The jurists also produced all kinds of legal punishments. Around AD 130 the jurist Salvius Iulianus drafted a standard form of the praetor's edict, which was used by all praetors from that time onwards. This edict contained detailed descriptions of all cases, in which the praetor would allow a legal action and in which he would grant a defense. The standard edict thus functioned like a comprehensive law code, even though it did not formally have the force of law. It indicated the requirements for a successful legal claim. The edict therefore became the basis for extensive legal commentaries by later classical jurists like Paulus and Ulpian. The new concepts and legal institutions developed by pre-classical and classical jurists are too numerous to mention here. Only a few examples are given here: The Roman Republic had three different branches: The Assemblies could decide whether war or peace. The Senate had complete control over the Treasury, and the Consuls had the highest juridical power. By the middle of the 3rd century, the conditions for the flourishing of a refined legal culture had become less favourable. The general political and economic situation deteriorated as the emperors assumed more direct control of all aspects of political life. The political system of the principate, which had retained some features of the republican constitution, began to transform itself into the absolute monarchy of the dominate. The existence of a legal science and of jurists who regarded law as a science, not as an instrument to achieve the political goals set by the absolute monarch, did not fit well into the new order of things. The literary production all but ended. Few jurists after the mid-3rd century are known by name. While legal science and legal education persisted to some extent in the eastern part of the Empire, most of the subtleties of classical law came to be disregarded and finally forgotten in the west. Classical law was replaced by so-called vulgar law. The Roman Republic's constitution or "mos maiorum" ("custom of the ancestors") was an unwritten set of guidelines and principles passed down mainly through precedent. Concepts that originated in the Roman constitution live on in constitutions to this day. Examples include checks and balances, the separation of powers, vetoes, filibusters, quorum requirements, term limits, impeachments, the powers of the purse, and regularly scheduled elections. Even some lesser used modern constitutional concepts, such as the block voting found in the electoral college of the United States, originate from ideas found in the Roman constitution. The constitution of the Roman Republic was not formal or even official. Its constitution was largely unwritten, and was constantly evolving throughout the life of the Republic. Throughout the 1st century BC, the power and legitimacy of the Roman constitution was progressively eroding. Even Roman constitutionalists, such as the senator Cicero, lost a willingness to remain faithful to it towards the end of the republic. When the Roman Republic ultimately fell in the years following the Battle of Actium and Mark Antony's suicide, what was left of the Roman constitution died along with the Republic. The first Roman Emperor, Augustus, attempted to manufacture the appearance of a constitution that still governed the Empire, by utilising that constitution's institutions to lend legitimacy to the Principate, e.g. reusing prior grants of greater imperium to substantiate Augustus' greater imperium over the Imperial provinces and the prorogation of different magistracies to justify Augustus' receipt of tribunician power. The belief in a surviving constitution lasted well into the life of the Roman Empire. "Stipulatio" was the basic form of contract in Roman law. It was made in the format of question and answer. The precise nature of the contract was disputed, as can be seen below. "Rei vindicatio" is a legal action by which the plaintiff demands that the defendant return a thing that belongs to the plaintiff. It may only be used when plaintiff owns the thing, and the defendant is somehow impeding the plaintiff's possession of the thing. The plaintiff could also institute an "actio furti" (a personal action) to punish the defendant. If the thing could not be recovered, the plaintiff could claim damages from the defendant with the aid of the "condictio furtiva" (a personal action). With the aid of the "actio legis Aquiliae" (a personal action), the plaintiff could claim damages from the defendant. "Rei vindicatio" was derived from the ius civile, therefore was only available to Roman citizens. To describe a person's position in the legal system, Romans mostly used the expression "togeus". The individual could have been a Roman citizen ("status civitatis") unlike foreigners, or he could have been free ("status libertatis") unlike slaves, or he could have had a certain position in a Roman family ("status familiae") either as the head of the family ("pater familias"), or some lower "member"—"alieni iuris"—which lives by someone else's law. Two status types were senator and emperor. The history of Roman Law can be divided into three systems of procedure: that of "legis actiones", the "formulary system", and "cognitio extra ordinem". The periods in which these systems were in use overlapped one another and did not have definitive breaks, but it can be stated that the legis actio system prevailed from the time of the XII Tables (c. 450 BC) until about the end of the 2nd century BC, that the formulary procedure was primarily used from the last century of the Republic until the end of the classical period (c. AD 200), and that of cognitio extra ordinem was in use in post-classical times. Again, these dates are meant as a tool to help understand the types of procedure in use, not as a rigid boundary where one system stopped and another began. During the republic and until the bureaucratization of Roman judicial procedure, the judge was usually a private person ("iudex privatus"). He had to be a Roman male citizen. The parties could agree on a judge, or they could appoint one from a list, called "album iudicum". They went down the list until they found a judge agreeable to both parties, or if none could be found they had to take the last one on the list. No one had a legal obligation to judge a case. The judge had great latitude in the way he conducted the litigation. He considered all the evidence and ruled in the way that seemed just. Because the judge was not a jurist or a legal technician, he often consulted a jurist about the technical aspects of the case, but he was not bound by the jurist's reply. At the end of the litigation, if things were not clear to him, he could refuse to give a judgment, by swearing that it wasn't clear. Also, there was a maximum time to issue a judgment, which depended on some technical issues (type of action, etc.). Later on, with the bureaucratization, this procedure disappeared, and was substituted by the so-called "extra ordinem" procedure, also known as cognitory. The whole case was reviewed before a magistrate, in a single phase. The magistrate had obligation to judge and to issue a decision, and the decision could be appealed to a higher magistrate. When the centre of the Empire was moved to the Greek East in the 4th century, many legal concepts of Greek origin appeared in the official Roman legislation. The influence is visible even in the law of persons or of the family, which is traditionally the part of the law that changes least. For example, Constantine started putting restrictions on the ancient Roman concept of "patria potestas", the power held by the male head of a family over his descendants, by acknowledging that persons "in potestate", the descendants, could have proprietary rights. He was apparently making concessions to the much stricter concept of paternal authority under Greek-Hellenistic law. The "Codex Theodosianus" (438 AD) was a codification of Constantian laws. Later emperors went even further, until Justinian finally decreed that a child "in potestate" became owner of everything it acquired, except when it acquired something from its father. The codes of Justinian, particularly the "Corpus Juris Civilis" (529–534) continued to be the basis of legal practice in the Empire throughout its so-called "Byzantine" history. Leo III the Isaurian issued a new code, the "Ecloga", in the early 8th century. In the 9th century, the emperors Basil I and Leo VI the Wise commissioned a combined translation of the Code and the Digest, parts of Justinian's codes, into Greek, which became known as the "Basilica". Roman law as preserved in the codes of Justinian and in the Basilica remained the basis of legal practice in Greece and in the courts of the Eastern Orthodox Church even after the fall of the Byzantine Empire and the conquest by the Turks, and, along with the Syro-Roman law book, also formed the basis for much of the "Fetha Negest", which remained in force in Ethiopia until 1931. In the west, Justinian's political authority never went any farther than certain portions of the Italian and Hispanic peninsulas. In Law codes were issued by the Germanic kings, however, the influence of early Eastern Roman codes on some of these is quite discernible. In many early Germanic states, Roman citizens continued to be governed by Roman laws for quite some time, even while members of the various Germanic tribes were governed by their own respective codes. The "Codex Justinianus" and the Institutes of Justinian were known in Western Europe, and along with the earlier code of Theodosius II, served as models for a few of the Germanic law codes; however, the "Digest" portion was largely ignored for several centuries until around 1070, when a manuscript of the "Digest" was rediscovered in Italy. This was done mainly through the works of glossars who wrote their comments between lines ("glossa interlinearis"), or in the form of marginal notes ("glossa marginalis"). From that time, scholars began to study the ancient Roman legal texts, and to teach others what they learned from their studies. The center of these studies was Bologna. The law school there gradually developed into Europe's first university. The students who were taught Roman law in Bologna (and later in many other places) found that many rules of Roman law were better suited to regulate complex economic transactions than were the customary rules, which were applicable throughout Europe. For this reason, Roman law, or at least some provisions borrowed from it, began to be re-introduced into legal practice, centuries after the end of the Roman empire. This process was actively supported by many kings and princes who employed university-trained jurists as counselors and court officials and sought to benefit from rules like the famous "Princeps legibus solutus est" ("The sovereign is not bound by the laws", a phrase initially coined by Ulpian, a Roman jurist). There are several reasons that Roman law was favored in the Middle Ages. Roman law regulated the legal protection of property and the equality of legal subjects and their wills, and it prescribed the possibility that the legal subjects could dispose their property through testament. By the middle of the 16th century, the rediscovered Roman law dominated the legal practice of many European countries. A legal system, in which Roman law was mixed with elements of canon law and of Germanic custom, especially feudal law, had emerged. This legal system, which was common to all of continental Europe (and Scotland) was known as "Ius Commune". This "Ius Commune" and the legal systems based on it are usually referred to as civil law in English-speaking countries. Only England and the Nordic countries did not take part in the wholesale reception of Roman law. One reason for this is that the English legal system was more developed than its continental counterparts by the time Roman law was rediscovered. Therefore, the practical advantages of Roman law were less obvious to English practitioners than to continental lawyers. As a result, the English system of common law developed in parallel to Roman-based civil law, with its practitioners being trained at the Inns of Court in London rather than receiving degrees in Canon or Civil Law at the Universities of Oxford or Cambridge. Elements of Romano-canon law were present in England in the ecclesiastical courts and, less directly, through the development of the equity system. In addition, some concepts from Roman law made their way into the common law. Especially in the early 19th century, English lawyers and judges were willing to borrow rules and ideas from continental jurists and directly from Roman law. The practical application of Roman law, and the era of the European "Ius Commune", came to an end when national codifications were made. In 1804, the French civil code came into force. In the course of the 19th century, many European states either adopted the French model or drafted their own codes. In Germany, the political situation made the creation of a national code of laws impossible. From the 17th century, Roman law in Germany had been heavily influenced by domestic (customary) law, and it was called "usus modernus Pandectarum". In some parts of Germany, Roman law continued to be applied until the German civil code (Bürgerliches Gesetzbuch, BGB) went into effect in 1900. Colonial expansion spread the civil law system. Today, Roman law is no longer applied in legal practice, even though the legal systems of some countries like South Africa and San Marino are still based on the old "jus commune". However, even where the legal practice is based on a code, many rules deriving from Roman law apply: no code completely broke with the Roman tradition. Rather, the provisions of the Roman law were fitted into a more coherent system and expressed in the national language. For this reason, knowledge of the Roman law is indispensable to understand the legal systems of today. Thus, Roman law is often still a mandatory subject for law students in civil law jurisdictions. As steps towards a unification of the private law in the member states of the European Union are being taken, the old "jus commune", which was the common basis of legal practice everywhere in Europe, but allowed for many local variants, is seen by many as a model.
https://en.wikipedia.org/wiki?curid=26364
Rockwell International Rockwell International was a major American manufacturing conglomerate in the latter half of the 20th century, involved in aircraft, the space industry, both defense-oriented and commercial electronics, light & heavy vehicle components in the automotive industry, printing presses, avionics, power tools, valves and meters, and industrial automation. Rockwell International's predecessor was founded in 1919 by Willard Rockwell. At its peak in the 1990s, Rockwell International was No. 27 on the "Fortune" 500 list, with assets of over $8 billion, sales of $27 billion and 115,000 employees. Boston-born Willard Rockwell (1888–1978) made his fortune with the invention and successful launch of a new bearing system for truck axles in 1919. He merged his Oshkosh, Wisconsin-based operation with the Timken-Detroit Axle Company in 1928, rising to become chairman of its board in 1940. In 1945, Rockwell Manufacturing Company acquired Delta Machinery and renamed it the Delta Power Tool Division of Rockwell Manufacturing Company and continued to manufacture in Milwaukee. In 1966, Rockwell invented the world's first power miter saw. In 1981, Rockwell's power tool group was acquired by Pentair and re-branded Delta Machinery. Pentair's Tools group was acquired by Black & Decker in 2005. Since 1994, Rockwell power tools are now manufactured by Positec Tool Corporation In 1956, Rockwell Manufacturing Co. bought Walker-Turner from Kearney and Trecker. In 1957, Walker-Turner operations were closed down in Plainfield, New Jersey and moved to Bellefontaine, Ohio and Tupelo, Mississippi. Timken-Detroit merged in 1953 with the Standard Steel Spring Company, forming the Rockwell Spring and Axle Company. After various mergers with automotive suppliers, it comprised about 10 to 20 factories in the Upper Midwestern U.S. and southern Ontario, and in 1958 renamed itself Rockwell-Standard Corporation. Pittsburgh-based Rockwell Standard then acquired and merged with Los Angeles-based North American Aviation to form North American Rockwell in September 1967. It then purchased Miehle-Goss-Dexter, the largest supplier of printing presses, and in 1973, acquired Collins Radio, a major avionics supplier. In 1968, Sterling Faucet Company was bought by Rockwell Manufacturing Co. and it became a subsidiary of the company for the following years. In 1973, North American Rockwell merged with Rockwell Manufacturing, run by Willard Rockwell, Jr., to form Rockwell International. In the same year, the company acquired Admiral Radio and TV for $500 million. In 1979, the appliance division was sold to Magic Chef. Rockwell International also drew on the strengths of several of George Westinghouse's concerns, and Westinghouse is considered a co-founder of the company. With the death of company founder and first CEO Willard F. Rockwell in 1978, and the stepping down of his son Willard Rockwell, Jr. in 1979 as the second CEO, Bob Anderson became CEO and led the company through the 1980s when it became the largest U.S. defense contractor and largest NASA contractor. Rockwell acquired the privately held Allen-Bradley Company for $1.6 billion in February 1985 — $1 billion of which was cash — and became a producer of industrial automation hardware and software. During the 1980s, Anderson, his CFO Bob dePalma, and the Rockwell management team built the company to #27 on the "Fortune" 500 list. It boasted sales of $12 billion, roughly $32 billion in 2019, and assets of over $8 billion, roughly $21 billion in 2019. Its workforce of over 115,000 was organized into nine major divisions — Space, Aircraft, Defense Electronics, Commercial Electronics, Light Duty Automotive Components, Heavy Duty Automotive Components, Printing Presses, Valves and Meters, and Industrial Automation. Rockwell International was a major employer in Southern California, northern Ohio, northern Georgia, eastern Oklahoma, Michigan, west Texas, Iowa, Illinois, Wisconsin, and western Pennsylvania. Anderson stepped down as CEO in February 1988, leaving the company to president Donald R. Beall. The completion of the Space Shuttle program and the completion of the B-1 bomber program had led to a decline in revenues, and Beall sought to diversify the company away from government contracts. The end of the Cold War and the perceived "peace dividend", however, prompted accelerated divestitures and sweeping management reforms. From 1988 to 2001 the company moved its headquarters four times: from Pittsburgh, Pennsylvania where it had been for decades to El Segundo, California to Seal Beach, California to Costa Mesa, California to Milwaukee, Wisconsin. At the end of the 1980s, the company sold its valve and meter division, formerly Rockwell Manufacturing, to British Tyre & Rubber. Although Rockwell was the #1 Defense and NASA Contractor, the "peace dividend" perceived after the fall of the Soviet bloc, led the company to sell its defense and aerospace business, including what was once North American Aviation, the Defense Electronics Division and Rocketdyne, to Boeing Integrated Defense Systems in December 1996. In the 1990s, the company spun off its semiconductor products as Conexant Technologies (CNXT), which is publicly traded and based in Newport Beach, California. Rockwell International also spun off its two automotive divisions (light vehicles div. & heavy vehicles div.) as one publicly traded company, Meritor Automotive, based in Troy, Michigan, which then merged with Arvin Industries to form Arvin Meritor. That company is now known as Meritor, Inc. In 1996, Rockwell International sold Graphic Systems (formerly Miehle-Goss-Dexter), an Illinois-based newspaper and commercial printing press manufacturer, to its internal management team Stonington Partners as part of a new corporation for $600 million. In 2001, what remained of Rockwell International was split into two publicly-traded companies, Rockwell Automation and Rockwell Collins, ending the run of what had once been a massive and diverse conglomerate. The split was structured so that Rockwell Automation was the legal successor of the old Rockwell International, while Rockwell Collins was the spin-off. At the end, the result had been four spin-offs and three sales combined from Rockwell's nine divisions. The various Rockwell companies list a large number of firsts in their histories, including the World War II-era P-51 Mustang fighter and the B-25 Mitchell bomber, and the Korean War-era F-86 Sabre fighter jet, as well as the Apollo spacecraft, the B-1 Lancer bomber, the Space Shuttle orbiter, and most of the Navstar Global Positioning System satellites. Rocketdyne, which had been spun off by North American in 1955, was re-merged into Rockwell, and by that time produced most of the rocket engines used in the United States. Rockwell also purchased the Aero Design and Engineering Company from William and Rufus Travis Amis. Rockwell redesigned the company's Aero Commander aircraft, introducing its new design as the Rockwell Commander 112 and Commander 114. The company developed a desktop calculator based on a MOSFET chip for use by its engineers. In 1967 Rockwell set up its own manufacturing plant to produce them, starting North American Rockwell MicroElectronics Corp. (called NARMEC). This would later become Rockwell Semiconductor. One of its major successes came in the early 1990s when it introduced the first low-cost 14.4 kbit/s modem chipset, which was used in a huge number of modems. Collins radios were fitted to 80% of the airliners which were based in First World Countries. Collins designed and built the radios that communicated the Apollo moon landings and the high frequency radio network that allows worldwide communication with U.S. military aircraft. Rockwell's Rocketdyne division designed and built the third stage of the Minuteman intercontinental ballistic missile, and the Advanced Inertial Reference Sphere inertial navigation system that provided its navigation. It also built inertial navigation systems for the fleet of ballistic missile submarines. In addition to the manufacture of nuclear missiles and bombers, Rockwell also produced key components of the bombs they carried, including plutonium triggers at the Rocky Flats Plant in Colorado. Rockwell ran the weapons plant from 1975 to 1990. Rockwell built heavy-duty truck axles and drive-trains in the U.S., along with power windows, seats and locks. The Rockwell Tripmaster trip recording system for commercial vehicles was released along with the Logtrak module for DOT log recording for fleets who successfully petitioned the DOT for paper logbook exemptions. Rockwell also built yachts and business jets and owned large amounts of real estate. It was also involved in providing custom electronic intelligence equipment to the Imperial Iranian Air Force as part of Project Ibex and paid bribes to the Shah of Iran in order to secure contracts there. For a more extensive list, see Rockwell International had a major research laboratory complex in Thousand Oaks, Ventura County, California. It was founded and built by North American Aviation in 1962, as the North American Science Center. In 1973 it became the Rockwell International Science Center. The laboratory did independent contract research for the U.S. Government, and also provided research services for the company's business units. It was famous for its research in: advanced materials, particularly ceramics; for its infrared imagers; for its research in liquid-crystal displays; and for its high-speed electronics. The laboratory invented Metalorganic vapour-phase epitaxy (MOVPE), also commonly known as Metal Organic Chemical Vapor Deposition (MOCVD). It also achieved fame in selected areas of information science, notably human-computer interaction, augmented reality, multimedia systems, and diagnostics. Rockwell Science Center led the United States Army Research Laboratory's Advanced Displays Federated Laboratory Consortium in the late 1990s. In 2000, the infrared imaging division of the laboratory moved into a new building in Camarillo, California. After Rockwell International's breakup in 2001, the laboratory was spun off as a semi-autonomous company called Rockwell Scientific, half owned by Rockwell Collins and half owned by Rockwell Automation. In 2006, the main laboratory and infrared imaging division were sold to Teledyne Corporation. Teledyne made the laboratory complex in Thousand Oaks into its corporate headquarters. A reduced but active research and development operation continues there, under the name Teledyne Scientific & Imaging, LLC.
https://en.wikipedia.org/wiki?curid=26367
Richard I of England Richard I (8 September 1157 – 6 April 1199) was King of England from 1189 until his death. He also ruled as Duke of Normandy, Aquitaine and Gascony, Lord of Cyprus, Count of Poitiers, Anjou, Maine, and Nantes, and was overlord of Brittany at various times during the same period. He was the third of five sons of King Henry II of England and Duchess Eleanor of Aquitaine and seemed unlikely to become king, but all of his brothers except the youngest, John, predeceased their father. Richard is known as Richard Cœur de Lion (Norman French: "Le quor de lion") or Richard the Lionheart because of his reputation as a great military leader and warrior. The troubador Bertran de Born also called him Richard Oc-e-Non (Occitan for "Yes and No"), possibly from a reputation for terseness. By the age of 16, Richard had taken command of his own army, putting down rebellions in Poitou against his father. Richard was an important Christian commander during the Third Crusade, leading the campaign after the departure of Philip II of France and achieving considerable victories against his Muslim counterpart, Saladin, although he failed to retake Jerusalem. Richard probably spoke both French and Occitan. He was born in England, where he spent his childhood; before becoming king, however, he lived most of his adult life in the Duchy of Aquitaine, in the southwest of France. Following his accession, he spent very little time, perhaps as little as six months, in England. Most of his life as king was spent on Crusade, in captivity, or actively defending his lands in France. Rather than regarding his kingdom as a responsibility requiring his presence as ruler, he has been perceived as preferring to use it merely as a source of revenue to support his armies. Nevertheless, he was seen as a pious hero by his subjects. He remains one of the few kings of England remembered more commonly by his epithet than his regnal number, and is an enduring iconic figure both in England and in France. Richard was born on 8 September 1157, probably at Beaumont Palace, in Oxford, England, son of King Henry II of England and Duchess Eleanor of Aquitaine. He was a younger brother of Henry the Young King and Matilda, Duchess of Saxony. As a younger son of King Henry II, he was not expected to ascend to the throne. He was also an elder brother of Geoffrey II, Duke of Brittany; Queen Eleanor of Castile; Queen Joan of Sicily; and Count John of Mortain, who succeeded him as king. Richard was the younger maternal half-brother of Marie of France, Countess of Champagne, and Alix, Countess of Blois. Henry II and Eleanor's eldest son William IX, Count of Poitiers, died before Richard's birth. Richard is often depicted as having been the favourite son of his mother. His father was Angevin-Norman and great-grandson of William the Conqueror. Contemporary historian Ralph de Diceto traced his family's lineage through Matilda of Scotland to the Anglo-Saxon kings of England and Alfred the Great, and from there legend linked them to Noah and Woden. According to Angevin family tradition, there was even 'infernal blood' in their ancestry, with a claimed descent from the fairy, or female demon, Melusine. While his father visited his lands from Scotland to France, Richard probably spent his childhood in England. His first recorded visit to the European continent was in May 1165, when his mother took him to Normandy. His wet nurse was Hodierna of St Albans, whom he gave a generous pension after he became king. Little is known about Richard's education. Although he was born in Oxford and brought up in England up to his eighth year, it is not known to what extent he used or understood English; he was an educated man who composed poetry and wrote in Limousin ("") and also in French. During his captivity, English prejudice against foreigners was used in a calculated way by his brother John to help destroy the authority of Richard's chancellor, William Longchamp, who was a Norman. One of the specific charges laid against Longchamp, by John's supporter Hugh Nonant, was that he could not speak English. This indicates that by the late 12th century a knowledge of English was expected of those in positions of authority in England. Richard was said to be very attractive; his hair was between red and blond, and he was light-eyed with a pale complexion. According to Clifford Brewer, he was , though that is unverifiable since his remains have been lost since at least the French Revolution. John, his youngest brother, was known to be . The , a Latin prose narrative of the Third Crusade, states that: "He was tall, of elegant build; the colour of his hair was between red and gold; his limbs were supple and straight. He had long arms suited to wielding a sword. His long legs matched the rest of his body". From an early age, Richard showed significant political and military ability, becoming noted for his chivalry and courage as he fought to control the rebellious nobles of his own territory. Marriage alliances were common among medieval royalty: they led to political alliances and peace treaties and allowed families to stake claims of succession on each other's lands. In March 1159 it was arranged that Richard would marry one of the daughters of Ramon Berenguer IV, Count of Barcelona; however, these arrangements failed, and the marriage never took place. Henry the Young King was married to Margaret, daughter of Louis VII of France, on 2 November 1160. Despite this alliance between the Plantagenets and the Capetians, the dynasty on the French throne, the two houses were sometimes in conflict. In 1168, the intercession of Pope Alexander III was necessary to secure a truce between them. Henry II had conquered Brittany and taken control of Gisors and the Vexin, which had been part of Margaret's dowry. Early in the 1160s there had been suggestions Richard should marry Alys, Countess of the Vexin, fourth daughter of Louis VII; because of the rivalry between the kings of England and France, Louis obstructed the marriage. A peace treaty was secured in January 1169 and Richard's betrothal to Alys was confirmed. Henry II planned to divide his and Eleanor's territories among their three eldest surviving sons: Henry would become King of England and have control of Anjou, Maine, and Normandy; Richard would inherit Aquitaine and Poitiers from his mother; and Geoffrey would become Duke of Brittany through marriage with Constance, heir presumptive of Conan IV. At the ceremony where Richard's betrothal was confirmed, he paid homage to the King of France for Aquitaine, thus securing ties of vassalage between the two. After Henry II fell seriously ill in 1170, he enacted his plan to divide his kingdom, although he would retain overall authority over his sons and their territories. Young Henry was crowned as heir apparent in June 1170, and in 1171 Richard left for Aquitaine with his mother, and Henry II gave him the duchy of Aquitaine at the request of Eleanor. Richard and his mother embarked on a tour of Aquitaine in 1171 in an attempt to pacify the locals. Together they laid the foundation stone of St Augustine's Monastery in Limoges. In June 1172, at age 12, Richard was formally recognised as the duke of Aquitaine and count of Poitou when he was granted the lance and banner emblems of his office; the ceremony took place in Poitiers and was repeated in Limoges, where he wore the ring of St Valerie, who was the personification of Aquitaine. According to Ralph of Coggeshall, Henry the Young King instigated rebellion against Henry II; he wanted to reign independently over at least part of the territory his father had promised him, and to break away from his dependence on Henry II, who controlled the purse strings. There were rumors that Eleanor might have encouraged her sons to revolt against their father. Henry the Young King abandoned his father and left for the French court, seeking the protection of Louis VII; his younger brothers, Richard and Geoffrey, soon followed him, while the five-year-old John remained in England. Louis gave his support to the three brothers and even knighted Richard, tying them together through vassalage. Jordan Fantosme, a contemporary poet, described the rebellion as a "war without love". The brothers made an oath at the French court that they would not make terms with Henry II without the consent of Louis VII and the French barons. With the support of Louis, Henry the Young King attracted many barons to his cause through promises of land and money; one such baron was Philip I, Count of Flanders, who was promised £1,000 and several castles. The brothers also had supporters ready to rise up in England. Robert de Beaumont, 3rd Earl of Leicester, joined forces with Hugh Bigod, 1st Earl of Norfolk, Hugh de Kevelioc, 5th Earl of Chester, and William I of Scotland for a rebellion in Suffolk. The alliance with Louis was initially successful, and by July 1173 the rebels were besieging Aumale, Neuf-Marché, and Verneuil, and Hugh de Kevelioc had captured Dol in Brittany. Richard went to Poitou and raised the barons who were loyal to himself and his mother in rebellion against his father. Eleanor was captured, so Richard was left to lead his campaign against Henry II's supporters in Aquitaine on his own. He marched to take La Rochelle but was rejected by the inhabitants; he withdrew to the city of Saintes, which he established as a base of operations. In the meantime, Henry II had raised a very expensive army of more than 20,000 mercenaries with which to face the rebellion. He marched on Verneuil, and Louis retreated from his forces. The army proceeded to recapture Dol and subdued Brittany. At this point Henry II made an offer of peace to his sons; on the advice of Louis the offer was refused. Henry II's forces took Saintes by surprise and captured much of its garrison, although Richard was able to escape with a small group of soldiers. He took refuge in Château de Taillebourg for the rest of the war. Henry the Young King and the Count of Flanders planned to land in England to assist the rebellion led by the Earl of Leicester. Anticipating this, Henry II returned to England with 500 soldiers and his prisoners (including Eleanor and his sons' wives and fiancées), but on his arrival found out that the rebellion had already collapsed. William I of Scotland and Hugh Bigod were captured on 13 and 25 July respectively. Henry II returned to France and raised the siege of Rouen, where Louis VII had been joined by Henry the Young King after abandoning his plan to invade England. Louis was defeated and a peace treaty was signed in September 1174, the Treaty of Montlouis. When Henry II and Louis VII made a truce on 8 September 1174, its terms specifically excluded Richard. Abandoned by Louis and wary of facing his father's army in battle, Richard went to Henry II's court at Poitiers on 23 September and begged for forgiveness, weeping and falling at the feet of Henry, who gave Richard the kiss of peace. Several days later, Richard's brothers joined him in seeking reconciliation with their father. The terms the three brothers accepted were less generous than those they had been offered earlier in the conflict (when Richard was offered four castles in Aquitaine and half of the income from the duchy): Richard was given control of two castles in Poitou and half the income of Aquitaine; Henry the Young King was given two castles in Normandy; and Geoffrey was permitted half of Brittany. Eleanor remained Henry II's prisoner until his death, partly as insurance for Richard's good behaviour. After the conclusion of the war, the process of pacifying the provinces that had rebelled against Henry II began. The King travelled to Anjou for this purpose, and Geoffrey dealt with Brittany. In January 1175 Richard was dispatched to Aquitaine to punish the barons who had fought for him. The historian John Gillingham notes that the chronicle of Roger of Howden is the main source for Richard's activities in this period. According to the chronicle, most of the castles belonging to rebels were to be returned to the state they were in 15 days before the outbreak of war, while others were to be razed. Given that by this time it was common for castles to be built in stone, and that many barons had expanded or refortified their castles, this was not an easy task. Roger of Howden records the two-month siege of Castillon-sur-Agen; while the castle was "notoriously strong", Richard's siege engines battered the defenders into submission. On this campaign, Richard acquired the name "the Lion" or "the Lionheart" due to his noble, brave and fierce leadership.He is referred to as "this our lion" (') as early as 1187 in the ' of , while the byname "lionheart" (') is first recorded in Ambroise's ' in the context of the Accon campaign of 1191. Henry seemed unwilling to entrust any of his sons with resources that could be used against him. It was suspected that Henry had appropriated Alys, Richard's betrothed, the daughter of Louis VII of France by his second wife, as his mistress. This made a marriage between Richard and Alys technically impossible in the eyes of the Church, but Henry prevaricated: he regarded Alys's dowry, Vexin in the Île-de-France, as valuable. Richard was discouraged from renouncing Alys because she was the sister of King Philip II of France, a close ally. After his failure to overthrow his father, Richard concentrated on putting down internal revolts by the nobles of Aquitaine, especially in the territory of Gascony. The increasing cruelty of his rule led to a major revolt there in 1179. Hoping to dethrone Richard, the rebels sought the help of his brothers Henry and Geoffrey. The turning point came in the Charente Valley in the spring of 1179. The well-defended fortress of Taillebourg seemed impregnable. The castle was surrounded by a cliff on three sides and a town on the fourth side with a three-layer wall. Richard first destroyed and looted the farms and lands surrounding the fortress, leaving its defenders no reinforcements or lines of retreat. The garrison sallied out of the castle and attacked Richard; he was able to subdue the army and then followed the defenders inside the open gates, where he easily took over the castle in two days. Richard the Lionheart's victory at Taillebourg deterred many barons from thinking of rebelling and forced them to declare their loyalty to him. It also won Richard a reputation as a skilled military commander. In 1181–1182 Richard faced a revolt over the succession to the county of Angoulême. His opponents turned to Philip II of France for support, and the fighting spread through the Limousin and Périgord. The excessive cruelty of Richard's punitive campaigns aroused even more hostility. However, with support from his father and from the Young King, Richard the Lionheart eventually succeeded in bringing the Viscount Aimar V of Limoges and Count Elie of Périgord to terms. After Richard had subdued his rebellious barons he again challenged his father. From 1180 to 1183 the tension between Henry and Richard grew, as King Henry commanded Richard to pay homage to Henry the Young King, but Richard refused. Finally, in 1183 Henry the Young King and Geoffrey, Duke of Brittany, invaded Aquitaine in an attempt to subdue Richard. Richard's barons joined in the fray and turned against their duke. However, Richard and his army succeeded in holding back the invading armies, and they executed any prisoners. The conflict paused briefly in June 1183 when the Young King died. With the death of Henry the Young King, Richard became the eldest surviving son and therefore heir to the English crown. King Henry demanded that Richard give up Aquitaine (which he planned to give to his youngest son John as his inheritance). Richard refused, and conflict continued between them. Henry II soon gave John permission to invade Aquitaine. To strengthen his position, in 1187, Richard allied himself with 22-year-old Philip II, the son of Eleanor's ex-husband Louis VII by Adela of Champagne. Roger of Howden wrote: The King of England was struck with great astonishment, and wondered what [this alliance] could mean, and, taking precautions for the future, frequently sent messengers into France for the purpose of recalling his son Richard; who, pretending that he was peaceably inclined and ready to come to his father, made his way to Chinon, and, in spite of the person who had the custody thereof, carried off the greater part of his father's treasures, and fortified his castles in Poitou with the same, refusing to go to his father. Overall, Howden is chiefly concerned with the politics of the relationship between Richard and King Philip. Gillingham has addressed theories suggesting that this political relationship was also sexually intimate, which he posits probably stemmed from an official record announcing that, as a symbol of unity between the two countries, the kings of England and France had slept overnight in the same bed. Gillingham has characterized this as "an accepted political act, nothing sexual about it;... a bit like a modern-day photo opportunity". In exchange for Philip's help against his father, Richard promised to concede to him his rights to both Normandy and Anjou. Richard paid homage to Philip in November 1187. With news arriving of the Battle of Hattin, he took the cross at Tours in the company of other French nobles. In 1188 Henry II planned to concede Aquitaine to his youngest son John. But Richard objected. He felt that Aquitaine was his and that John was unfit to take over the land once belonging to his mother. This refusal is what finally made Henry II bring Queen Eleanor out of prison. He sent her to Aquitaine and demanded that Richard give up his lands to his mother, who would once again rule over those lands. The following year, Richard attempted to take the throne of England for himself by joining Philip's expedition against his father. On 4 July 1189, the forces of Richard and Philip defeated Henry's army at Ballans. Henry, with John's consent, agreed to name Richard his heir apparent. Two days later Henry II died in Chinon, and Richard the Lionheart succeeded him as King of England, Duke of Normandy, and Count of Anjou. Roger of Howden claimed that Henry's corpse bled from the nose in Richard's presence, which was assumed to be a sign that Richard had caused his death. Richard I was officially invested as Duke of Normandy on 20 July 1189 and crowned king in Westminster Abbey on 3 September 1189. Tradition barred all Jews and women from the investiture, but some Jewish leaders arrived to present gifts for the new king. According to Ralph of Diceto, Richard's courtiers stripped and flogged the Jews, then flung them out of court. When a rumour spread that Richard had ordered all Jews to be killed, the people of London attacked the Jewish population. Many Jewish homes were destroyed by arsonists, and several Jews were forcibly converted. Some sought sanctuary in the Tower of London, and others managed to escape. Among those killed was Jacob of Orléans, a respected Jewish scholar. Roger of Howden, in his "", claimed that the jealous and bigoted citizens started the rioting, and that Richard punished the perpetrators, allowing a forcibly converted Jew to return to his native religion. Baldwin of Forde, Archbishop of Canterbury, reacted by remarking, "If the King is not God's man, he had better be the devil's". Offended that he was not being obeyed and realising that the assaults could destabilise his realm on the eve of his departure on crusade, Richard ordered the execution of those responsible for the most egregious murders and persecutions, including rioters who had accidentally burned down Christian homes. He distributed a royal writ demanding that the Jews be left alone. The edict was only loosely enforced, however, and the following March further violence occurred, including a massacre at York. Richard had already taken the cross as Count of Poitou in 1187. His father and Philip II had done so at Gisors on 21 January 1188 after receiving news of the fall of Jerusalem to Saladin. After Richard became king, he and Philip agreed to go on the Third Crusade, since each feared that during his absence the other might usurp his territories. Richard swore an oath to renounce his past wickedness in order to show himself worthy to take the cross. He started to raise and equip a new crusader army. He spent most of his father's treasury (filled with money raised by the Saladin tithe), raised taxes, and even agreed to free King William I of Scotland from his oath of subservience to Richard in exchange for marks (£). To raise still more revenue he sold the right to hold official positions, lands, and other privileges to those interested in them. Those already appointed were forced to pay huge sums to retain their posts. William Longchamp, Bishop of Ely and the King's chancellor, made a show of bidding £ to remain as Chancellor. He was apparently outbid by a certain Reginald the Italian, but that bid was refused. Richard made some final arrangements on the continent. He reconfirmed his father's appointment of William Fitz Ralph to the important post of seneschal of Normandy. In Anjou, Stephen of Tours was replaced as seneschal and temporarily imprisoned for fiscal mismanagement. Payn de Rochefort, an Angevin knight, became seneschal of Anjou. In Poitou the ex-provost of Benon, Peter Bertin, was made seneschal, and finally, the household official Helie de La Celle was picked for the seneschalship in Gascony. After repositioning the part of his army he left behind to guard his French possessions, Richard finally set out on the crusade in summer 1190. (His delay was criticised by troubadours] such as Bertran de Born.) He appointed as regents Hugh de Puiset, Bishop of Durham, and William de Mandeville, 3rd Earl of Essex—who soon died and was replaced by William Longchamp. Richard's brother John was not satisfied by this decision and started scheming against William Longchamp. When Richard was raising funds for his crusade, he was said to declare, "I would have sold London if I could find a buyer". In September 1190 Richard and Philip arrived in Sicily. After the death of King William II of Sicily in 1189 his cousin Tancred had seized power, although the legal heir was William's aunt Constance, wife of Henry VI, Holy Roman Emperor. Tancred had imprisoned William's widow, Queen Joan, who was Richard's sister and did not give her the money she had inherited in William's will. When Richard arrived he demanded that his sister be released and given her inheritance; she was freed on 28 September, but without the inheritance. The presence of foreign troops also caused unrest: in October, the people of Messina revolted, demanding that the foreigners leave. Richard attacked Messina, capturing it on 4 October 1190. After looting and burning the city Richard established his base there, but this created tension between Richard and Philip Augustus. He remained there until Tancred finally agreed to sign a treaty on 4 March 1191. The treaty was signed by Richard, Philip, and Tancred. Its main terms were: The two kings stayed on in Sicily for a while, but this resulted in increasing tensions between them and their men, with Philip Augustus plotting with Tancred against Richard. The two kings finally met to clear the air and reached an agreement, including the end of Richard's betrothal to Philip's sister Alys. In April 1191 Richard left Messina for Acre, but a storm dispersed his large fleet. After some searching, it was discovered that the ship carrying his sister Joan and his new fiancée, Berengaria of Navarre, was anchored on the south coast of Cyprus, along with the wrecks of several other vessels, including the treasure ship. Survivors of the wrecks had been taken prisoner by the island's ruler, Isaac Komnenos. On 1 May 1191 Richard's fleet arrived in the port of Lemesos on Cyprus. He ordered Isaac to release the prisoners and treasure. Isaac refused, so Richard landed his troops and took Limassol. Various princes of the Holy Land arrived in Limassol at the same time, in particular Guy of Lusignan. All declared their support for Richard provided that he support Guy against his rival, Conrad of Montferrat. The local magnates abandoned Isaac, who considered making peace with Richard, joining him on the crusade, and offering his daughter in marriage to the person named by Richard. Isaac changed his mind, however, and tried to escape. Richard's troops, led by Guy de Lusignan, conquered the whole island by 1 June. Isaac surrendered and was confined with silver chains because Richard had promised that he would not place him in irons. Richard named Richard de Camville and Robert of Thornham as governors. He later sold the island to the master of Knights Templar, Robert de Sablé, and it was subsequently acquired, in 1192, by Guy of Lusignan and became a stable feudal kingdom. The rapid conquest of the island by Richard was of strategic importance. The island occupies a key strategic position on the maritime lanes to the Holy Land, whose occupation by the Christians could not continue without support from the sea. Cyprus remained a Christian stronghold until the battle of Lepanto (1571). Richard's exploit was well publicised and contributed to his reputation, and he also derived significant financial gains from the conquest of the island. Richard left Cyprus for Acre on 5 June with his allies. Before leaving Cyprus on crusade, Richard married Berengaria, the first-born daughter of King Sancho VI of Navarre. Richard first grew close to her at a tournament held in her native Navarre. The wedding was held in Limassol on 12 May 1191 at the Chapel of St George and was attended by Richard's sister Joan, whom he had brought from Sicily. The marriage was celebrated with great pomp and splendour, many feasts and entertainments, and public parades and celebrations followed commemorating the event. When Richard married Berengaria he was still officially betrothed to Alys, and he pushed for the match in order to obtain the Kingdom of Navarre as a fief, as Aquitaine had been for his father. Further, Eleanor championed the match, as Navarre bordered Aquitaine, thereby securing the southern border of her ancestral lands. Richard took his new wife on crusade with him briefly, though they returned separately. Berengaria had almost as much difficulty in making the journey home as her husband did, and she did not see England until after his death. After his release from German captivity, Richard showed some regret for his earlier conduct, but he was not reunited with his wife. The marriage remained childless. King Richard landed at Acre on 8 June 1191. He gave his support to his Poitevin vassal Guy of Lusignan, who had brought troops to help him in Cyprus. Guy was the widower of his father's cousin Sibylla of Jerusalem and was trying to retain the kingship of Jerusalem, despite his wife's death during the Siege of Acre the previous year. Guy's claim was challenged by Conrad of Montferrat, second husband of Sibylla's half-sister, Isabella: Conrad, whose defence of Tyre had saved the kingdom in 1187, was supported by Philip of France, son of his first cousin Louis VII of France, and by another cousin, Leopold V, Duke of Austria. Richard also allied with Humphrey IV of Toron, Isabella's first husband, from whom she had been forcibly divorced in 1190. Humphrey was loyal to Guy and spoke Arabic fluently, so Richard used him as a translator and negotiator. Richard and his forces aided in the capture of Acre, despite Richard's serious illness. At one point, while sick from scurvy, he is said to have picked off guards on the walls with a crossbow, while being carried on a stretcher. Eventually, Conrad of Montferrat concluded the surrender negotiations with Saladin's forces inside Acre and raised the banners of the kings in the city. Richard quarrelled with Leopold of Austria over the deposition of Isaac Komnenos (related to Leopold's Byzantine mother) and his position within the crusade. Leopold's banner had been raised alongside the English and French standards. This was interpreted as arrogance by both Richard and Philip, as Leopold was a vassal of the Holy Roman Emperor (although he was the highest-ranking surviving leader of the imperial forces). Richard's men tore the flag down and threw it in the moat of Acre. Leopold left the crusade immediately. Philip also left soon afterwards, in poor health and after further disputes with Richard over the status of Cyprus (Philip demanded half the island) and the kingship of Jerusalem. Richard, suddenly, found himself without allies. Richard had kept 2,700 Muslim prisoners as hostages against Saladin fulfilling all the terms of the surrender of the lands around Acre. Philip, before leaving, had entrusted his prisoners to Conrad, but Richard forced him to hand them over to him. Richard feared his forces being bottled up in Acre as he believed his campaign could not advance with the prisoners in train. He, therefore, ordered all the prisoners executed. He then moved south, defeating Saladin's forces at the Battle of Arsuf north of Jaffa on 7 September 1191. Saladin attempted to harass Richard's army into breaking its formation in order to defeat it in detail. Richard maintained his army's defensive formation, however, until the Hospitallers broke ranks to charge the right wing of Saladin's forces. Richard then ordered a general counterattack, which won the battle. Arsuf was an important victory. The Muslim army was not destroyed, despite the considerable casualties it suffered, but it did rout; this was considered shameful by the Muslims and boosted the morale of the Crusaders. In November 1191, following the fall of Jaffa, the Crusader army advanced inland towards Jerusalem. The army then marched to Beit Nuba, only 12 miles from Jerusalem. Muslim morale in Jerusalem was so low that the arrival of the Crusaders would probably have caused the city to fall quickly. However, the weather was appallingly bad, cold with heavy rain and hailstorms; this, combined with the fear that the Crusader army, if it besieged Jerusalem, might be trapped by a relieving force, led to the decision to retreat back to the coast. Richard attempted to negotiate with Saladin, but this was unsuccessful. In the first half of 1192, he and his troops refortified Ascalon. An election forced Richard to accept Conrad of Montferrat as King of Jerusalem, and he sold Cyprus to his defeated protégé, Guy. Only days later, on 28 April 1192, Conrad was stabbed to death by Assassins before he could be crowned. Eight days later Richard's own nephew Henry II of Champagne was married to the widowed Isabella, although she was carrying Conrad's child. The murder has never been conclusively solved, and Richard's contemporaries widely suspected his involvement. The crusader army made another advance on Jerusalem, and in June 1192 it came within sight of the city before being forced to retreat once again, this time because of dissension amongst its leaders. In particular, Richard and the majority of the army council wanted to force Saladin to relinquish Jerusalem by attacking the basis of his power through an invasion of Egypt. The leader of the French contingent, Hugh III, Duke of Burgundy, however, was adamant that a direct attack on Jerusalem should be made. This split the Crusader army into two factions, and neither was strong enough to achieve its objective. Richard stated that he would accompany any attack on Jerusalem but only as a simple soldier; he refused to lead the army. Without a united command the army had little choice but to retreat back to the coast. There commenced a period of minor skirmishes with Saladin's forces, punctuated by another defeat in the field for the Ayyubid army at the Battle of Jaffa. Baha' al-Din, a contemporary Muslim soldier and biographer of Saladin, recorded a tribute to Richard's martial prowess at this battle: "I have been assured ... that on that day the king of England, lance in hand, rode along the whole length of our army from right to left, and not one of our soldiers left the ranks to attack him. The Sultan was wroth thereat and left the battlefield in anger...". Both sides realised that their respective positions were growing untenable. Richard knew that both Philip and his own brother John were starting to plot against him, and the morale of Saladin's army had been badly eroded by repeated defeats. However, Saladin insisted on the razing of Ascalon's fortifications, which Richard's men had rebuilt, and a few other points. Richard made one last attempt to strengthen his bargaining position by attempting to invade Egypt—Saladin's chief supply-base—but failed. In the end, time ran out for Richard. He realised that his return could be postponed no longer since both Philip and John were taking advantage of his absence. He and Saladin finally came to a settlement on 2 September 1192. The terms provided for the destruction of Ascalon's fortifications, allowed Christian pilgrims and merchants access to Jerusalem, and initiated a three-year truce. Richard, being ill with scurvy, left for England on October 9, 1192. Bad weather forced Richard's ship to put in at Corfu, in the lands of Byzantine Emperor Isaac II Angelos, who objected to Richard's annexation of Cyprus, formerly Byzantine territory. Disguised as a Knight Templar, Richard sailed from Corfu with four attendants, but his ship was wrecked near Aquileia, forcing Richard and his party into a dangerous land route through central Europe. On his way to the territory of his brother-in-law Henry the Lion, Richard was captured shortly before Christmas 1192 near Vienna by Leopold of Austria, who accused Richard of arranging the murder of his cousin Conrad of Montferrat. Moreover, Richard had personally offended Leopold by casting down his standard from the walls of Acre. Leopold kept Richard prisoner at Dürnstein Castle under the care of Leopold's Hadmar of Kuenring. His mishap was soon known to England, but the regents were for some weeks uncertain of his whereabouts. While in prison, Richard wrote ' or ' ("No man who is imprisoned"), which is addressed to his half-sister Marie. He wrote the song, in French and Occitan versions, to express his feelings of abandonment by his people and his sister. The detention of a crusader was contrary to public law, and on these grounds Pope Celestine III excommunicated Duke Leopold. On 28 March 1193 Richard was brought to Speyer and handed over to Holy Roman Emperor Henry VI, who imprisoned him in Trifels Castle. Henry VI was aggrieved by the support the Plantagenets had given to the family of Henry the Lion and by Richard's recognition of Tancred in Sicily. Henry VI needed money to raise an army and assert his rights over southern Italy and continued to hold Richard for ransom. In response, Pope Celestine III excommunicated Henry VI, as he had Duke Leopold, for the continued wrongful imprisonment of Richard. Richard famously refused to show deference to the Emperor and declared to him, "". Despite his complaints, the conditions of his captivity were not severe. The Emperor demanded that marks (100,000 pounds of silver) be delivered to him before he would release the King, the same amount raised by the Saladin tithe only a few years earlier, and 2–3 times the annual income for the English Crown under Richard. Richard's mother, Eleanor, worked to raise the ransom. Both clergy and laymen were taxed for a quarter of the value of their property, the gold and silver treasures of the churches were confiscated, and money was raised from the scutage and the carucage taxes. At the same time, John, Richard's brother, and King Philip of France offered marks for Henry VI to hold Richard prisoner until Michaelmas 1194. Henry turned down the offer. The money to rescue the King was transferred to Germany by the Emperor's ambassadors, but "at the king's peril" (had it been lost along the way, Richard would have been held responsible), and finally, on 4 February 1194 Richard was released. Philip sent a message to John: "Look to yourself; the devil is loose". In Richard's absence, his brother John revolted with the aid of Philip; amongst Philip's conquests in the period of Richard's imprisonment was Normandy. Richard forgave John when they met again and named him as his heir in place of their nephew, Arthur. At Winchester, on 11 March 1194, Richard was crowned a second time to nullify the shame of his captivity. Richard began his reconquest of Normandy. The fall of the Château de Gisors to the French in 1193 opened a gap in the Norman defences. The search began for a fresh site for a new castle to defend the duchy of Normandy and act as a base from which Richard could launch his campaign to take back the Vexin from French control. A naturally defensible position was identified perched high above the River Seine, an important transport route, in the manor of Andeli. Under the terms of the Treaty of Louviers (December 1195) between Richard and Philip II, neither king was allowed to fortify the site; despite this, Richard intended to build the vast Château Gaillard. Richard tried to obtain the manor through negotiation. Walter de Coutances, Archbishop of Rouen, was reluctant to sell the manor as it was one of the diocese's most profitable, and other lands belonging to the diocese had recently been damaged by war. When Philip besieged Aumale in Normandy, Richard grew tired of waiting and seized the manor, although the act was opposed by the Catholic Church. The archbishop issued an interdict against performing church services in the duchy of Normandy; Roger of Howden detailed "unburied bodies of the dead lying in the streets and square of the cities of Normandy". The interdict was still in force when work began on the castle, but Pope Celestine III repealed it in April 1197 after Richard made gifts of land to the archbishop and the diocese of Rouen, including two manors and the prosperous port of Dieppe. Royal expenditure on castles declined from the levels spent under Henry II, attributed to a concentration of resources on Richard's war with the king of France. However, the work at Château Gaillard was some of the most expensive of its time and cost an estimated £15,000 to £20,000 between 1196 and 1198. This was more than double Richard's spending on castles in England, an estimated £7,000. Unprecedented in its speed of construction, the castle was mostly complete in two years when most construction on such a scale would have taken the best part of a decade. According to William of Newburgh, in May 1198 Richard and the labourers working on the castle were drenched in a "rain of blood". While some of his advisers thought the rain was an evil omen, Richard was undeterred. As no master-mason is mentioned in the otherwise detailed records of the castle's construction, military historian Allen Brown has suggested that Richard himself was the overall architect; this is supported by the interest Richard showed in the work through his frequent presence. In his final years, the castle became Richard's favourite residence, and writs and charters were written at Château Gaillard bearing """ (at the Fair Castle of the Rock). Château Gaillard was ahead of its time, featuring innovations that would be adopted in castle architecture nearly a century later. Allen Brown described Château Gaillard as "one of the finest castles in Europe", and military historian Sir Charles Oman wrote that it was considered "the masterpiece of its time. The reputation of its builder, Cœur de Lion, as a great military engineer might stand firm on this single structure. He was no mere copyist of the models he had seen in the East, but introduced many original details of his own invention into the stronghold". Determined to resist Philip's designs on contested Angevin lands such as the Vexin and Berry, Richard poured all his military expertise and vast resources into the war on the French King. He organised an alliance against Philip, including Baldwin IX of Flanders, Renaud, Count of Boulogne, and his father-in-law King Sancho VI of Navarre, who raided Philip's lands from the south. Most importantly, he managed to secure the Welf inheritance in Saxony for his nephew, Henry the Lion's son, who was elected Otto IV of Germany in 1198. Partly as a result of these and other intrigues, Richard won several victories over Philip. At Fréteval in 1194, just after Richard's return to France from captivity and money-raising in England, Philip fled, leaving his entire archive of financial audits and documents to be captured by Richard. At the Battle of Gisors (sometimes called Courcelles) in 1198, Richard took "—"God and my Right"—as his motto (still used by the British monarchy today), echoing his earlier boast to Emperor Henry that his rank acknowledged no superior but God. In March 1199, Richard was in Limousin suppressing a revolt by Viscount Aimar V of Limoges. Although it was Lent, he "devastated the Viscount's land with fire and sword". He besieged the puny, virtually unarmed castle of Châlus-Chabrol. Some chroniclers claimed that this was because a local peasant had uncovered a treasure trove of Roman gold. On 26 March 1199, Richard was hit in the shoulder by a crossbow, and the wound turned gangrenous. Richard asked to have the crossbowman brought before him; called alternatively Pierre (or Peter) Basile, John Sabroz, Dudo, and Bertrand de Gourdon (from the town of Gourdon) by chroniclers, the man turned out (according to some sources, but not all) to be a boy. He said Richard had killed his father and two brothers, and that he had killed Richard in revenge. He expected to be executed, but as a final act of mercy Richard forgave him, saying "Live on, and by my bounty behold the light of day", before he ordered the boy to be freed and sent away with 100 shillings. It is unclear whether the pardon was upheld following his death. Richard then set his affairs in order, bequeathing all his territory to his brother John and his jewels to his nephew Otto. Richard died on 6 April 1199 in the arms of his mother, and thus "ended his earthly day." Because of the nature of Richard's death, it was later referred to as "the Lion by the Ant was slain". According to one chronicler, Richard's last act of chivalry proved fruitless when the infamous mercenary captain Mercadier had the boy flayed alive and hanged as soon as Richard died. Richard's heart was buried at Rouen in Normandy, his entrails in Châlus (where he died), and the rest of his body at the feet of his father at Fontevraud Abbey in Anjou. In 2012, scientists analysed the remains of Richard's heart and found that it had been embalmed with various substances, including frankincense, a symbolically important substance because it had been present both at the birth and embalming of the Christ. Henry Sandford, Bishop of Rochester (1226–1235), announced that he had seen a vision of Richard ascending to Heaven in March 1232 (along with Stephen Langton, the former Archbishop of Canterbury), the King having presumably spent 33 years in purgatory as expiation for his sins. Richard produced no legitimate heirs and acknowledged only one illegitimate son, Philip of Cognac. As a result, he was succeeded by his brother John as king. However, his French territories (with the exception of Rouen) initially rejected John as a successor, preferring his nephew Arthur, as how royal inheritance rules applied to the situation at the time of Richard’s death was unclear. The lack of any direct heirs from Richard was the first step in the dissolution of the Angevin Empire. Contemporaries considered Richard as both a king and a knight famed for personal martial prowess; this was, apparently, the first such instance of this combination. He was known as a valiant, competent military leader and individual fighter who was courageous and generous. At the same time, he was considered prone to the sins of lust, pride, greed, and above all excessive cruelty. Ralph of Coggeshall, summarising Richard's career, deplores that the King was one of "the immense cohort of sinners". He was criticised by clergy chroniclers for having taxed the clergy both for the Crusade and for his ransom, whereas the church and the clergy were usually exempt from taxes. Richard was a patron and a protector of the trouvères and troubadours of his entourage; he was also a poet himself. He was interested in writing and music, and two poems are attributed to him. The first one is a sirventes in Old French, "Dalfin je us voill desrenier", and the second one is a lament that he wrote during his imprisonment at Dürnstein Castle, "Ja nus hons pris", with a version in Old Occitan and a version in Old French. In the historiography of the second half of the 20th century, much interest was shown in Richard's sexuality, in particular whether there was evidence of homosexuality. The topic had not been raised by Victorian or Edwardian historians, a fact which was itself denounced as a "conspiracy of silence" by John Harvey (1948). The argument primarily drew on accounts of Richard's behaviour, as well as of his confessions and penitences, and of his childless marriage. Richard did have at least one illegitimate child, and there are reports on his sexual relations with local women during his campaigns. Historians remain divided on the question of Richard's sexuality. Harvey argued in favour of his homosexuality but has been disputed by other historians, most notably John Gillingham (1994), who argues that Richard was probably heterosexual. Flori (1999) again argued in favour of Richard's homosexuality, based on Richard's two public confessions and penitences (in 1191 and 1195) which, according to Flori, "must have" referred to the sin of sodomy. Flori, however, concedes that contemporary accounts of Richard taking women by force exist, concluding that he probably had sexual relations with both men and women. Flori and Gillingham nevertheless agree that accounts of bed-sharing do not support the suggestion that Richard had a sexual relationship with King Philip II, as had been suggested by other modern authors. The second Great Seal of Richard I (1198) shows him bearing a shield depicting "three lions passant-guardant". This is the first instance of the appearance of this blazon, which later became established as the Royal Arms of England. It is likely, therefore, that Richard introduced this heraldic design. In his earlier Great Seal of 1189, he had used either one "lion rampant" or two "lions rampants combatants", which arms he may have adopted from his father. Richard is also credited with having originated the English crest of a "lion statant" (now "statant-guardant"). The coat of three lions continues to represent England on several coins of the pound sterling, forms the basis of several emblems of English national sports teams (such as the England national football team, and the team's "Three Lions" anthem), and endures as one of the most recognisable national symbols of England. Around the middle of the 13th century, various legends developed that, after Richard's capture, his minstrel Blondel travelled Europe from castle to castle, loudly singing a song known only to the two of them (they had composed it together). Eventually, he came to the place where Richard was being held, and Richard heard the song and answered with the appropriate refrain, thus revealing where the King was incarcerated. The story was the basis of André Ernest Modeste Grétry's opera "" and seems to be the inspiration for the opening to Richard Thorpe's film version of "Ivanhoe". It seems unconnected to the real Jean 'Blondel' de Nesle, an aristocratic . It also does not correspond to the historical reality, since the King's jailers did not hide the fact; on the contrary, they publicised it. At some time around the 16th century, tales of Robin Hood started to mention him as a contemporary and supporter of King Richard the Lionheart, Robin being driven to outlawry, during the misrule of Richard's evil brother John, while Richard was away at the Third Crusade. Richard's reputation over the years has "fluctuated wildly", according to historian John Gillingham. While contemporary sources emphasize his stern and unforgiving nature and his excessive cruelty, his image had already been romanticized a few decades after his death, with the new views on Richard depicting him as generous-hearted '. Richard left an indelible imprint on the imagination extending to the present, in large part because of his military exploits, and his popular image tended to be dominated by the positive qualities of chivalry and military competence. This is reflected in Steven Runciman's final verdict of Richard I: "he was a bad son, a bad husband, and a bad king, but a gallant and splendid soldier" ("History of the Crusades" Vol. III). Meanwhile, Muslim writers during the Crusades period and after wrote of him: "Never have we had to face a bolder or more subtle opponent". Victorian England was divided on Richard: many admired him as a crusader and man of God, erecting an heroic statue to him outside the Houses of Parliament. The late-Victorian scholar William Stubbs, on the other hand, thought him "a bad son, a bad husband, a selfish ruler, and a vicious man". During his ten years' reign, he was in England for no more than six months, and was totally absent for the last five years. Stubbs argued that: He was a bad king: his great exploits, his military skill, his splendour and extravagance, his poetical tastes, his adventurous spirit, do not serve to cloak his entire want of sympathy, or even consideration, for his people. He was no Englishman, but it does not follow that he gave to Normandy, Anjou, or Aquitaine the love or care that he denied to his kingdom. His ambition was that of a mere warrior: he would fight for anything whatever, but he would sell everything that was worth fighting for. The glory that he sought was that of victory rather than conquest. In World War I, when British troops commanded by General Edmund Allenby captured Jerusalem, the British press printed cartoons of Richard looking down from the heavens with the caption reading, "At last my dream has come true". General Allenby protested against his campaign being presented as a latter-day Crusade, however, stating "The importance of Jerusalem lay in its strategic importance, there was no religious impulse in this campaign".
https://en.wikipedia.org/wiki?curid=26368
Robert Jordan James Oliver Rigney Jr. (October 17, 1948 – September 16, 2007), better known by his pen name Robert Jordan, was an American author of epic fantasy. He is best known for the "Wheel of Time" series (finished by Brandon Sanderson upon Jordan's death) which comprises 14 books and a prequel novel. He is one of several writers to have written original Conan the Barbarian novels; his are highly acclaimed to this day. Rigney also wrote historical fiction under his pseudonym Reagan O'Neal, a western as Jackson O'Reilly, and dance criticism as Chang Lung. Additionally, he ghostwrote an "international thriller" that is still believed to have been written by someone else. Jordan was born in Charleston, South Carolina. He served two tours in Vietnam (from 1968 to 1970) with the United States Army as a helicopter gunner. He was awarded the Distinguished Flying Cross with oak leaf cluster, the Bronze Star with "V" and oak leaf cluster, and two Vietnamese Gallantry Crosses with palm. After returning from Vietnam he attended The Citadel, where he received an undergraduate degree in physics; after graduating he was employed by the United States Navy as a nuclear engineer. He began writing in 1977. Jordan was a history buff and enjoyed hunting, fishing, sailing, poker, chess, pool, and pipe-collecting. He described himself as a "High Church" Episcopalian and received communion more than once a week. He lived with his wife, Harriet McDougal, who works as a book editor (currently with Tor Books; she was also Jordan's editor) in a house built in 1797. On March 23, 2006, Jordan disclosed that he had been diagnosed with cardiac amyloidosis and that with treatment, his median life expectancy was four years. In a separate blog post, he encouraged his fans not to worry about him and stated that he intended to lead a long and creative life. He began chemotherapy at Mayo Clinic in early April 2006. He was enrolled in a study of Revlimid, which had recently been approved for multiple myeloma but not yet tested on primary amyloidosis. Jordan died on September 16, 2007, and his funeral service was held on September 19, 2007. He was cremated and his ashes buried in the churchyard of St. James Church in Goose Creek, outside Charleston, South Carolina. Jordan's papers can be found in the special collections of the College of Charleston.
https://en.wikipedia.org/wiki?curid=26370
Rift Valley fever Rift Valley fever (RVF) is a viral disease that can cause mild to severe symptoms. The mild symptoms may include: fever, muscle pains, and headaches which often last for up to a week. The severe symptoms may include: loss of sight beginning three weeks after the infection, infections of the brain causing severe headaches and confusion, and bleeding together with liver problems which may occur within the first few days. Those who have bleeding have a chance of death as high as 50%. The disease is caused by the RVF virus, which is of the "Phlebovirus" type. It is spread by either touching infected animal blood, breathing in the air around an infected animal being butchered, drinking raw milk from an infected animal, or the bite of infected mosquitoes. Animals such as cows, sheep, goats, and camels may be affected. In these animals it is spread mostly by mosquitoes. It does not appear that one person can infect another person. The disease is diagnosed by finding antibodies against the virus or the virus itself in the blood. Prevention of the disease in humans is accomplished by vaccinating animals against the disease. This must be done before an outbreak occurs because if it is done during an outbreak it may worsen the situation. Stopping the movement of animals during an outbreak may also be useful, as may decreasing mosquito numbers and avoiding their bites. There is a human vaccine; however, as of 2010 it is not widely available. There is no specific treatment and medical efforts are supportive. Outbreaks of the disease have only occurred in Africa and Arabia. Outbreaks usually occur during periods of increased rain which increase the number of mosquitoes. The disease was first reported among livestock in Rift Valley of Kenya in the early 1900s, and the virus was first isolated in 1931. In humans, the virus can cause several syndromes. Usually, sufferers have either no symptoms or only a mild illness with fever, headache, muscle pains, and liver abnormalities. In a small percentage of cases (< 2%), the illness can progress to hemorrhagic fever syndrome, meningoencephalitis (inflammation of the brain and tissues lining the brain), or affect the eye. Patients who become ill usually experience fever, generalised weakness, back pain, dizziness, and weight loss at the onset of the illness. Typically, people recover within two to seven days after onset. About 1% of people with the disease die of it. In livestock, the fatality level is significantly higher. Pregnant livestock infected with RVF abort virtually 100% of foetuses. An epizootic (animal disease epidemic) of RVF is usually first indicated by a wave of unexplained abortions. Other signs in livestock include vomiting and diarrhoea, respiratory disease, fever, lethargy, anorexia and sudden death in young animals. The virus belongs to the "Bunyavirales" order. This is an order of enveloped negative single stranded RNA viruses. All Bunyaviruses have an outer lipid envelope with two glycoproteins—G(N) and G(C)—required for cell entry. They deliver their genome into the host-cell cytoplasm by fusing their envelope with an endosomal membrane. The virus' G(C) protein has a class II membrane fusion protein architecture similar to that found in flaviviruses and alphaviruses. This structural similarity suggests that there may be a common origin for these viral families. The virus' 11.5 kb tripartite genome is composed of single-stranded RNA. As a "Phlebovirus," it has an ambisense genome. Its L and M segments are negative-sense, but its S segment is ambisense. These three genome segments code for six major proteins: L protein (viral polymerase), the two glycoproteins G(N) and G(C), the nucleocapsid N protein, and the nonstructural NSs and NSm proteins. The virus is transmitted through mosquito vectors, as well as through contact with the tissue of infected animals. Two species—"Culex tritaeniorhynchus" and "Aedes vexans"—are known to transmit the virus. Other potential vectors include "Aedes caspius", "Aedes mcintosh", "Aedes ochraceus," "Culex pipiens", "Culex antennatus", "Culex perexiguus", "Culex zombaensis" and "Culex quinquefasciatus". Contact with infected tissue is considered to be the main source of human infections. The virus has been isolated from two bat species: the Peter's epauletted fruit bat ("Micropteropus pusillus") and the aba roundleaf bat ("Hipposideros abae"), which are believed to be reservoirs for the virus. Although many components of the RVFV's RNA play an important role in the virus’ pathology, the nonstructural protein encoded on the S segment (NSs) is the only component that has been found to directly affect the host. NSs is hostile and combative against the hosts interferon (IFNs) antiviral response. IFNs are essential in order for the immune system to fight off viral infections in a host. This inhibitory mechanism is believed to be due to a number of reasons, the first being, competitive inhibition of the formation of the transcription factor. On this transcription factor, NSs interacts with and binds to a subunit that is needed for RNA polymerase I and II. This interaction cause competitive inhibition with another transcription factor component and prevents the assembly process of the transcription factor complex, which results in the suppression of the host antiviral response. Transcription suppression is believed to be another mechanism of this inhibitory process. This occurs when an area of NSs interacts with and binds to the host's protein, SAP30 and forms a complex. This complex causes histone acetylation to regress, which is needed for transcriptional activation of the IFN promoter. This causes IFN expression to be obstructed. Lastly, NSs has also been known to affect regular activity of double-stranded RNA-dependent protein kinase R.. This protein is involved in cellular antiviral responses in the host. When RVFV is able to enter the hosts DNA, NSs forms a filamentous structure in the nucleus. This allows the virus to interact with specific areas of the hosts DNA that relates to segregation defects and induction of chromosome continuity. This increases host infectivity and decreases the host's antiviral response. Diagnosis relies on viral isolation from tissues, or serological testing with an ELISA. Other methods of diagnosis include Nucleic Acid Testing (NAT), cell culture, and IgM antibody assays. As of September 2016, the Kenya Medical Research Institute (KEMRI) has developed a product called Immunoline, designed to diagnose the disease in humans much faster than in previous methods. A person's chances of becoming infected can be reduced by taking measures to decrease contact with blood, body fluids, or tissues of infected animals and protection against mosquitoes and other bloodsucking insects. Use of mosquito repellents and bed nets are two effective methods. For persons working with animals in RVF-endemic areas, wearing protective equipment to avoid any exposure to blood or tissues of animals that may potentially be infected is an important protective measure. Potentially, establishing environmental monitoring and case surveillance systems may aid in the prediction and control of future RVF outbreaks. No vaccines are currently available for humans. While a vaccines have been developed for humans, it has only been used experimentally for scientific personnel in high-risk environments. Trials of a number of vaccines, such as NDBR-103 and TSI-GSD 200, are ongoing. Different types of vaccines for veterinary use are available. The killed vaccines are not practical in routine animal field vaccination because of the need of multiple injections. Live vaccines require a single injection but are known to cause birth defects and abortions in sheep and induce only low-level protection in cattle. The live-attenuated vaccine, MP-12, has demonstrated promising results in laboratory trials in domesticated animals, but more research is needed before the vaccine can be used in the field. The live-attenuated clone 13 vaccine was recently registered and used in South Africa. Alternative vaccines using molecular recombinant constructs are in development and show promising results. A vaccine has been conditionally approved for use in animals in the US. It has been shown that knockout of the NSs and NSm nonstructural proteins of this virus produces an effective vaccine in sheep as well. RVF outbreaks occur across sub-Saharan Africa, with outbreaks occurring elsewhere infrequently. In Egypt in 1977–78, an estimated 200,000 people were infected and there were at least 594 deaths. In Kenya in 1998, the virus killed more than 400 people. In September 2000, an outbreak was confirmed in Saudi Arabia and Yemen. On 19 October 2011, a case of Rift Valley fever contracted in Zimbabwe was reported in a Caucasian female traveler who returned to France after a 26-day stay in Marondera, Mashonaland East Province during July and August, 2011 but later classified as "not confirmed." Outbreaks of this disease usually correspond with the warm phases of the EI Niño/Southern Oscillation. During this time there is an increase in rainfall, flooding and greenness of vegetation index, which leads to an increase in mosquito vectors. RVFV can be transmitted vertically in mosquitos, meaning that the virus can be passed from the mother to her offspring. During dry conditions, the virus can remain viable for a number of years in the egg. Mosquitos lay their eggs in water, where they eventually hatch. As water is essential for mosquito eggs to hatch, rainfall and flooding cause an increase in the mosquito population and an increased potential for the virus. In November 2006, a Rift Valley fever outbreak started in Kenya. The cases were from the North Eastern Province and Coast Province of Kenya, which had received heavy rain, causing floods and creating breeding grounds for mosquitoes, which spread the virus of the fever from infected livestock to humans. By 7 January 2007, about 75 people had died and another 183 were infected. The outbreak forced the closure of livestock markets in the North Eastern Province, affecting the economy of the region. The outbreak was subsequently reported to have moved into Maragua and Kirinyaga districts of Central Province of Kenya. On 20 January 2007, the outbreak was reported to have crossed into Somalia from Kenya and killed 14 people in the Lower Jubba region. As of 23 January 2007, cases had started to crop up at the Kenyan capital, Nairobi. Businesses were suffering large losses, as customers were shunning the common meat joints for the popular "nyama choma" (roast meat), as it was believed to be spreading the fever. In December 2006 and again in January 2007, Taiwan International Health Action (Taiwan IHA) began operating missions in Kenya consisting of medical experts assisting in training laboratory and health facility personnel, and included donations of supplies, such as mosquito sprays. The United States Centers for Disease Control also set up an assistance mission and laboratory in Kenya. By the end of January, 2007, some 148 people had died since the outbreak began in December. On 14 March 2007, the Kenyan government declared RVF cases to be diminished after spending an estimated $2.5 million in vaccine and deployment costs. It also lifted the ban on cattle movement in the affected areas. The final death toll in this outbreak was more than 150 people. However, on 8 June 2018, the Ministry of Health in Kenya declared another outbreak of RVF. As of 2 November 2007, 125 cases, including 60 deaths, had been reported from more than 10 localities of White Nile, Sinnar, and Gezira states in Sudan. Young adult males were predominantly affected. More than 25 human samples have been found positive for RVF by PCR or ELISA. As of 8 April 2010, the Ministry of Health South Africa had reported 87 human cases infected with Rift Valley fever (RVF), including two deaths in Free State, Eastern Cape and Northern Cape provinces. Most of these cases reported direct contact with RVFV-infected livestock and or were linked to farms with confirmed animal cases of RVF. The human cases were among farmers, veterinarians and farm workers. All cases were confirmed with RVF by test conducted at the National Institute of Communicable Diseases (NICD) in Johannesburg, South Africa. An outbreak of Rift Valley fever virus (RVFV) infection affected sheep, goats, cattle and wildlife on farms within Free State, Eastern Cape, Northern Cape, Western Cape, Mpumalanga, North West, and Gauteng provinces. As of 29 March 2010, about 78 farms reported laboratory-confirmed animal cases, with extensive livestock deaths. Before the 2010 outbreak, sporadic cases of RVFV infection in animals had been documented in South Africa. The last major outbreak of the disease in humans occurred between 1974 and 1976, where an estimated 10,000 to 20,000 cases were recorded. On March 2016, a male butcher from Kabale District in western Uganda reported to a local hospital with symptoms of headache, fever, fatigue and bleeding, subsequently testing positive for Rift Valley Fever. CDC sent epidemiologists to the District to assist the Ugandan Ministry of Health with the epidemiologic investigation of this small, localized outbreak of 3 confirmed and 2 probable cases. Working with the Uganda Virus Research Institute (UVRI) and the Uganda Ministry of Health, the CDC team conducted a serologic study in animals and humans and also assessed residents’ knowledge, attitudes, and practices related to Rift Valley Fever. The team collected samples from cows, goats and sheep, and interviewed and tested 650 district residents. A coordinated educational campaign targeting the general population, farmers, herders, and butchers was initiated and informational posters were created targeting these groups. As of 16 June 2018, an outbreak of Rift Valley fever is ongoing in northern Kenya, with 26 suspected human cases including 6 deaths in Wajir County (24 cases) and Marsabit County (2 cases); 7 cases have been confirmed. There have also been numerous deaths and abortions in camels, goats and other livestock across a wider area of the country. As of 3 May 2019, an outbreak is ongoing in the French Mayotte Islands, part of the Comoro group off Mozambique. The first human case showed symptoms on 22 November 2018, and there have been a total of 129 confirmed human cases, as well as more than a hundred foci in livestock. The outbreak has led to restrictions on the sale of uncooked milk, as well as the sale and export of cattle and uncooked meat. WHO noted that mosquito transmission should decrease as the rainy season ends in April, although Cyclone Kenneth has been associated with increased rain. Rift Valley fever was one of more than a dozen agents that the United States researched as potential biological weapons before the nation suspended its biological weapons program in 1969. The disease is one of several identified by WHO as a likely cause of a future epidemic in a new plan developed after the Ebola epidemic for urgent research and development toward new diagnostic tests, vaccines and medicines.
https://en.wikipedia.org/wiki?curid=26378
Rogue state Rogue state or outlaw state is a term applied by some international theorists to states they consider threatening to the world's peace. This means being seen to meet certain criteria, such as being ruled by authoritarian or totalitarian governments that severely restrict human rights, sponsoring terrorism and seeking to proliferate weapons of mass destruction. The term is used most by the United States (though the US State Department officially stopped using the term in 2000), and in his speech at the United Nations (UN) in 2017, Donald Trump reiterated this phrase. However, it has been applied by other countries as well. As early as July 1985, President Ronald Reagan stated that "we are not going to tolerate … attacks from outlaw states by the strangest collection of misfits, looney tunes, and squalid criminals since the advent of the Third Reich," but it fell to the Clinton administration to elaborate on this concept. In the 1994 issue of "Foreign Affairs", U.S. National Security Advisor Anthony Lake labelled five nations as "rogue states": North Korea, Cuba, Iran, Libya under Muammar Gaddafi, and Ba'athist Iraq. He described these regimes as "recalcitrant and outlaw states that not only choose to remain outside the family [of democratic nations] but also assault its basic values". In theory, to be classified as a "rogue state", a state had to do the following: seek to obtain weapons of mass destruction, support terrorism, and severely abuse its own citizens. While four of the listed countries met all these conditions, Cuba, though known from repressing it citizens and its vocal criticism of the United States, was put on the list solely because of the political influence of the Cuban-American community and specifically that of the Cuban American National Foundation (pre-Jorge Mas Santos), whereas Syria and Pakistan avoided being added to the list because the United States hoped that Damascus could play a constructive role in the Arab-Israeli peace process, and because Washington had long maintained close relations with Islamabad—a vestige of the Cold War. Three other nations, the Federal Republic of Yugoslavia, Sudan and the Islamic Emirate of Afghanistan, were treated as "rogue states" as well. The US State Department at times labelled Yugoslavia as a "rogue state" because its leader, Slobodan Milošević, had been accused of violating the rights of his nation's citizens, including but not limited to attempted genocide in Croatia and orchestrating the Srebrenica massacre in eastern Bosnia. The United States employed several tools to isolate and punish "rogue states". Tough unilateral economic sanctions, often at congressional behest, were imposed on or tightened against Iran, Libya, Cuba, Sudan, and Afghanistan. After the conclusion of the Gulf War in 1991, the United States selectively used airpower against Iraq for years during the Iraqi no-fly zones to force them in complying with various United Nations Security Council resolutions regarding disarmament (i.e., Resolution 687) and human rights (i.e., Resolution 688). Cruise missiles were fired at Afghanistan and Sudan in retaliation for terrorist attacks against U.S. embassies in Kenya and Tanzania in August 1998. In March 1999, NATO launched a massive air-bombing campaign against Yugoslavia in response to the Yugoslav Army's crackdown on ethnic Albanian separatists in the province of Kosovo. In the last six months of the Clinton administration, U.S. Secretary of State Madeleine Albright announced that the term "rogue state" would be abolished in June 2000, in favour of the term "states of concern", as three of the nations listed as "rogue states" (Libya, Iran, and North Korea) no longer met the conditions established to define a "rogue state". Libya was removed from the State Sponsors of Terrorism list in 2006 after achieving success through diplomacy. Relations with Libya also became more mutual following the eight month Libyan Civil War in 2011, which resulted in the National Transitional Council ousting longtime Libyan leader Muammar Gaddafi from power. In 2015, after the US reopened its embassy in Cuba and restarted diplomatic relations with the Cuban government, Cuba was removed from the list of State sponsors of terrorism and was no longer referred to as a "rogue state". More recently, the Donald Trump administration labelled Venezuela a "rogue state" due to its gross human rights violations, economic collapse and rampant excess deaths, anti-American stances and its reported involvement in international drug trafficking. During the 2017 UN general assembly, UN ambassador Nikki Haley called Venezuela a global threat and a "dangerous narco-state". Some figures of the Venezuelan government, like Vice-president Tareck el Aissami and minister of defense Vladimir Padrino López, were permanently banned from entering US territory, due to their involvement with human rights abuses and drug cartels. Later in the year, the US government banned all high ranking Venezuelan government officials from entering US territory. Currently, due to the 2019 Venezuelan presidential crisis, Nicolas Maduro's government (which controls Venezuela "de facto") is not recognized as legitimate by the United States or any other state in the Western Hemisphere, with the exceptions of Cuba, Dominica, Nicaragua, Saint Kitts and Nevis, Saint Vincent and the Grenadines, and Suriname. On 19 June 2020, US Secretary of State Mike Pompeo criticised China's "rogue attitude" in its neighbourhood with regard to the border confrontation between China and India. Besides, Mike Pompeo labelled Chinese Communist Party and its Party leader Xi Jinping as the actors that "greenlighted a brutal campaign of repression against Chinese Muslims, a human rights violation on a scale we haven’t seen since World War II." Besides, he said that China was responsible for “pushing disinformation and malicious cyber campaigns so as to drive a wedge between the US and Europe”, and that Beijing had “lied about the coronavirus and let it spread to the rest of the world, while pressuring the WHO to assist its cover-up campaign. Hundreds of thousands of people have died, and the global economy has been decimated.” In the aftermath of the September 11 attacks, the Bush administration returned to using a similar term. The concept of "rogue states" was replaced by the Bush administration with the "Axis of Evil" concept (gathering Iraq, Iran, and North Korea). U.S. President George W. Bush first spoke of this "Axis of Evil" during his January 2002 State of the Union Address. More terms, such as "Beyond the Axis of Evil" and "Outposts of Tyranny", would follow suit. As the U.S. government remains the most active proponent of the expression "rogue state", the term has received much criticism from those who disagree with U.S. foreign policy. Both the concepts of "rogue states" and the "Axis of Evil" have been criticized by certain scholars, including philosopher Jacques Derrida and linguist Noam Chomsky, who considered it more or less a justification of imperialism and a useful word for propaganda. Some critics charge that "rogue state" merely means any state that is generally hostile to the U.S., or even one that opposes the U.S. without necessarily posing a wider threat. Others, such as author William Blum, have written that the term is also applicable to the U.S. and Israel. In his "", Blum makes the case that the United States defines itself as a rogue state through its foreign policy. In 23 February 1999, Turkish President Süleyman Demirel described Greece as a "rogue state" because of its alleged support to PKK. Demirel said that: "Greece serves as a sanctuary for members of the PKK seeking shelter and provides training facilities and logistics to the terrorists." On June 28, 2012, after the shooting down of a Turkish warplane by the Syrian Army during the Syrian Civil War, Turkish Prime Minister Recep Tayyip Erdoğan declared Syria to be a "rogue state". Commentator Robert Ellis, writing in the British newspaper "The Independent" in 2016, wrote that Turkey under President Recep Tayyip Erdogan risks "being regarded as a rogue state" due to its increasingly authoritarian government, the deterioration of the human rights in the country, the Turkish government's involvement in Syria and its alleged support of terrorist groups.
https://en.wikipedia.org/wiki?curid=26383
Rebol Rebol ( ; historically REBOL) is a cross-platform data exchange language and a multi-paradigm dynamic programming language designed by Carl Sassenrath for network communications and distributed computing. It introduces the concept of dialecting: small, optimized, domain-specific languages for code and data, which is also the most notable property of the language according to its designer Carl Sassenrath: Douglas Crockford, known for his involvement in the development of JavaScript, has described Rebol as "a more modern language, but with some very similar ideas to Lisp, in that it's all built upon a representation of data which is then executable as programs" and as one of JSON's influences. Originally, the language and its official implementation were proprietary and closed source, developed by REBOL Technologies. Following discussion with Lawrence Rosen, the Rebol version 3 interpreter was released under the Apache 2.0 license on December 12, 2012. Older versions are only available in binary form, and no source release for them is planned. Rebol has been used to program Internet applications (both client- and server-side), database applications, utilities, and multimedia applications. Rebol was initially an acronym for Relative Expression Based Object Language written in all caps. To align with modern trends in language naming represented, e.g. by the change replacing historical name "LISP" by "Lisp", programmers ceased the practice of writing "REBOL" in all caps. Sassenrath eventually put the naming question to the community debate on his blog. In subsequent writing, Sassenrath adopted the convention of writing the language name as "Rebol". First released in 1997, Rebol was designed over a 20-year period by Carl Sassenrath, the architect and primary developer of AmigaOS, based on his study of denotational semantics and using concepts from the programming languages Lisp, Forth, Logo, and Self. One of the Rebol design principles is "to do simple things in simple ways". In the following example the "Visual interface dialect" is used to describe a simple Hello world program with a graphical user interface: view layout [text "Hello world!" button "Quit" [quit]] This is how a similar example looks in R3-GUI: view [text "Hello world!" button "Quit" on-action [quit]] Rebol domain-specific languages, called "dialects", are micro-languages optimized for a specific purpose. Dialects can be used to define business rules, graphical user interfaces or sequences of screens during the installation of a program. Users can define their own dialects, reusing any existing Rebol word and giving it a specific meaning in that dialect. Dialects are interpreted by functions processing Rebol blocks (or parsing strings) in a specific way. An example of Rebol's dialecting abilities can be seen with the word codice_1. In the "data exchange dialect" codice_1 is just a word not having any specific meaning. In the "do dialect", codice_1 is a global variable referring to a native function passing back a function result value. In the "visual interface dialect (VID)", codice_1 is a keyword causing the layout engine to simulate a carriage return, moving the "rendering pen" down to the beginning of the next line. A Rebol interpreter with graphical abilities must understand and interpret many dialects. The table below lists the most important ones in order of significance. Rebol syntax is free-form, not requiring specific positioning. However, indentation is recommended to better convey the structure of the text to human readers. Syntactic properties of different dialects may differ. The common platform for all Rebol dialects is the "data exchange dialect"; other dialects are usually derived from it. In addition to being the common platform for all dialects, the "data exchange dialect" is directly used to represent data and metadata, populate data structures, send data over Internet, and save them in data storage. In contrast to programming languages like C, the "data exchange dialect" does not consist of declarations, statements, expressions or keywords. A valid "data exchange dialect" text stream is a tree data structure consisting of blocks (the root block is implicit, subblocks are delimited by square brackets), parens (delimited by round brackets), strings (delimited by double quotes or curly brackets suitable for multi-line strings; caret notation is used for unprintable characters), URLs, e-mail addresses, files, paths or other composite values. Unlike ALGOL blocks, Rebol blocks are composite values similar to quoted s-expressions in Lisp. The fact that code is written in the form of Rebol blocks makes the language homoiconic. Blocks as well as parens may contain other composite values (a block may contain subblocks, parens, strings, ...) or scalar values like words, set-words (words suffixed by the colon), get-words (words prefixed by the colon), lit-words (words prefixed by the apostrophe), numbers, money, characters, etc., separated by whitespace. Note that special characters are allowed in words, so codice_5 is a word unlike codice_6, which is a sequence of three words separated by spaces. Comments may appear following the semicolon until the end of the line. Multi-line comments or comments not ignored by the lexical parser can be written using "ordinary" datatypes like multi-line strings. Blocks containing domain-specific language can be submitted as arguments to specific "evaluator" functions. The most frequently used evaluator is the codice_7 function. It is used by default to interpret the text input to the interpreter console. The "do dialect" interpreted by the codice_7 function, is an expression-oriented sublanguage of the "data exchange dialect". The main semantic unit of the language is the expression. In contrast to imperative programming languages descending from ALGOL, the "do dialect" has neither keywords, nor statements. Words are used as case-insensitive variables. Like in all dynamically typed languages, variables don't have an associated type, type is associated with values. The result, i.e. the evaluation of a word is returned, when a word is encountered by the codice_7 function. The set-word form of a word can be used for assignment. While not having statements, assignment, together with functions with side-effects can be used for imperative programming. Subblocks of the root block evaluate to themselves. This property is used to handle data blocks, for structured programming by submitting blocks as arguments to control functions like codice_10, codice_11, codice_12, etc., and for dialecting, when a block is passed to a specific interpreter function. A specific problem worth noting is that composite values, assigned to variables, are not copied. To make a copy, the value must be passed to the codice_13 function. The codice_7 function normally follows a prefix style of evaluation, where a function processes the arguments that follow it. However, infix evaluation using infix operators exists too. Infix evaluation takes precedence over the prefix evaluation. For example, returns 1, since the infix addition takes precedence over the computation of the absolute value. When evaluating infix expressions, the order of evaluation is left to right, no operator takes precedence over another. For example, 2 + 3 * 4 returns 20, while an evaluation giving precedence to multiplication would yield 14. All operators have prefix versions. codice_15 usually evaluates arguments before passing them to a function. So, the below expression: first reads the Wikipedia Rebol page and then passes the result to the codice_16 function. Parentheses can be used to change the order of evaluation. Using prefix notation, the usage of parentheses in expressions can be avoided. The simple precedence rules are both an advantage: as well as a disadvantage: The codice_17 function is preferably used to specify, validate, transform and interpret dialects. It does so by matching "parse expressions" at run time. "Parse expressions" are written in the "parse dialect", which, like the "do dialect", is an expression-oriented sublanguage of the "data exchange dialect". Unlike the "do dialect", the "parse dialect" uses keywords representing operators and the most important nonterminals, infix parsing operators don't have prefix equivalents and use precedence rules ("sequence" has higher precedence than "choice"). Actions can be included to be taken during the parsing process as well and the codice_17 function can be used to process blocks or strings. At the "string parsing" level codice_17 must handle the "low level" parsing, taking into account characters and delimiters. "Block parsing" is higher level, handling the scanning at the level of Rebol values. The parse dialect belongs to the family of grammars represented by the top-down parsing language or the parsing expression grammar (PEG). The main similarity is the presence of the "sequence" and "choice" operators all the family members have. Parse dialect syntax and the similarities between the parse dialect and the PEG are illustrated by this transliteration of a PEG example that parses an arithmetic expression: Digit: charset [#"0" - #"9"] Value: [some Digit | "(" Expr ")"] Product: [Value any "/"] Value Sum: [Product any "-"] Product Expr: Sum parse/all "12+13" Expr The official Rebol 2.7.8 implementation is available in several editions ("/Core", "/View", "/Command", "/SDK" and "/IOS"). Both "/Core" and "/View" editions are freely redistributable software. The runtime environment is stored in a single executable file. "Rebol/Core" 2.7.8, the console edition, is about 300 KB and "Rebol/View" 2.7.8, the graphical user interface edition, is about 650 KB in size. "Rebol/View" provides platform-independent graphics and sound access, and comes with its own windowing toolkit and extensible set of styles (GUI widgets). Extended editions, such as "Rebol/Command" 2.7.8 or "Rebol/SDK" 2.7.8 require a paid license; they add features like ODBC data access, and the option to create standalone executable files.
https://en.wikipedia.org/wiki?curid=26384
Red Hat Red Hat, Inc. is an American multinational software company that provides open source software products to enterprises. Founded in 1993, Red Hat has its corporate headquarters in Raleigh, North Carolina, with other offices worldwide. It became a subsidiary of IBM on July 9, 2019. Red Hat has become associated to a large extent with its enterprise operating system Red Hat Enterprise Linux. With the acquisition of open-source enterprise middleware vendor JBoss, Red Hat also offers Red Hat Virtualization (RHV), an enterprise virtualization product. Red Hat provides storage, operating system platforms, middleware, applications, management products, and support, training, and consulting services. Red Hat creates, maintains, and contributes to many free software projects. It has acquired several proprietary software product codebases through corporate mergers and acquisitions and has released such software under open source licenses. , Red Hat is the second largest corporate contributor to the Linux kernel version 4.14 after Intel. On October 28, 2018, IBM announced its intent to acquire Red Hat for $34 billion. The acquisition closed on July 9, 2019. Red Hat's lead financial adviser in the transaction was Guggenheim Securities. In 1993, Bob Young incorporated the ACC Corporation, a catalog business that sold Linux and Unix software accessories. In 1994, Marc Ewing created his own Linux distribution, which he named Red Hat Linux (associated to the time Ewing wore a red Cornell University lacrosse hat, given to him by his grandfather, while attending Carnegie Mellon University). Ewing released the software in October, and it became known as the Halloween release. Young bought Ewing's business in 1995, and the two merged to become Red Hat Software, with Young serving as chief executive officer (CEO). Red Hat went public on August 11, 1999, achieving the eighth-biggest first-day gain in the history of Wall Street. Matthew Szulik succeeded Bob Young as CEO in December of that year. Bob Young went on to found the online print on demand and self-publishing company, Lulu in 2002. On November 15, 1999, Red Hat acquired Cygnus Solutions. Cygnus provided commercial support for free software and housed maintainers of GNU software products such as the GNU Debugger and GNU Binutils. One of the founders of Cygnus, Michael Tiemann, became the chief technical officer of Red Hat and the vice president of open-source affairs. Later Red Hat acquired WireSpeed, C2Net and Hell's Kitchen Systems. In February 2000, "InfoWorld" awarded Red Hat its fourth consecutive "Operating System Product of the Year" award for Red Hat Linux 6.1. Red Hat acquired Planning Technologies, Inc in 2001 and AOL's iPlanet directory and certificate-server software in 2004. Red Hat moved its headquarters from Durham to North Carolina State University's Centennial Campus in Raleigh, North Carolina in February 2002. In the following month Red Hat introduced Red Hat Linux Advanced Server, later renamed Red Hat Enterprise Linux (RHEL). Dell, IBM, HP and Oracle Corporation announced their support of the platform. In December 2005, "CIO Insight" magazine conducted its annual "Vendor Value Survey", in which Red Hat ranked #1 in value for the second year in a row. Red Hat stock became part of the NASDAQ-100 on December 19, 2005. Red Hat acquired open-source middleware provider JBoss on June 5, 2006, and JBoss became a division of Red Hat. On September 18, 2006, Red Hat released the Red Hat Application Stack, which integrated the JBoss technology and which was certified by other well-known software vendors. On December 12, 2006, Red Hat stock moved from trading on NASDAQ (RHAT) to the New York Stock Exchange (RHT). In 2007 Red Hat acquired MetaMatrix and made an agreement with Exadel to distribute its software. On March 15, 2007, Red Hat released Red Hat Enterprise Linux 5, and in June acquired Mobicents. On March 13, 2008, Red Hat acquired Amentra, a provider of systems integration services for service-oriented architecture, business process management, systems development and enterprise data services. On July 27, 2009, Red Hat replaced CIT Group in Standard and Poor's 500 stock index, a diversified index of 500 leading companies of the U.S. economy. This was reported as a major milestone for Linux. On December 15, 2009, it was reported that Red Hat will pay to settle a class action lawsuit related to the restatement of financial results from July 2004. The suit had been pending in U.S. District Court for the Eastern District of North Carolina. Red Hat reached the proposed settlement agreement and recorded a one-time charge of for the quarter that ended Nov. 30. On January 10, 2011, Red Hat announced that it would expand its headquarters in two phases, adding 540 employees to the Raleigh operation, and investing over . The state of North Carolina is offering up to in incentives. The second phase involves "expansion into new technologies such as software virtualization and technology cloud offerings". On August 25, 2011, Red Hat announced it would move about 600 employees from the N.C. State Centennial Campus to the Two Progress Plaza building. A ribbon cutting ceremony was held June 24, 2013, in the re-branded Red Hat Headquarters. In 2012, Red Hat became the first one-billion dollar open-source company, reaching in annual revenue during its fiscal year. Red Hat passed the $2 billion benchmark in 2015. the company's annual revenue was nearly $3 billion. On October 16, 2015, Red Hat announced its acquisition of IT automation startup Ansible, rumored for an estimated US$100 million. On June 2017, Red Hat announced Red Hat Hyperconverged Infrastructure (RHHI) 1.0 software product In May 2018, Red Hat acquired CoreOS. On October 28, 2018, IBM announced its intent to acquire Red Hat for US$34 billion, in one of its largest-ever acquisitions. The company will operate out of IBM's Hybrid Cloud division. Six months later, on May 3, 2019, the US Department of Justice concluded its review of IBM's proposed Red Hat acquisition, and according to Steven J. Vaughan-Nichols "essentially approved the IBM/Red Hat deal". The acquisition was closed on July 9, 2019. Red Hat sponsors the Fedora Project, a community-supported free software project that aims to promote the rapid progress of free and open-source software and content. Fedora aims for rapid innovation using open processes and public forums. The Fedora Project Board, which comprises community leaders and representatives of Red Hat, leads the project and steers the direction of the project and of Fedora, the Linux distribution it develops. Red Hat employees work with the code alongside community members, and many innovations within the Fedora Project make their way into new releases of Red Hat Enterprise Linux. Red Hat operates on a business model based on open-source software, development within a community, professional quality assurance, and subscription-based customer support. They produce open-source code so that more programmers can make adaptations and improvements. Red Hat sells subscriptions for the support, training, and integration services that help customers in using their open-source software products. Customers pay one set price for unlimited access to services such as Red Hat Network and up to 24/7 support. In September 2014, however, CEO Jim Whitehurst announced that Red Hat was "in the midst of a major shift from client-server to cloud-mobile". Rich Bynum, a member of Red Hat's legal team, attributes Linux's success and rapid development partially to open-source business models, including Red Hat's. Red Hat engineers worked with the One Laptop per Child initiative (a non-profit organization established by members of the MIT Media Lab) to design and produce an inexpensive laptop and try to provide every child in the world with access to open communication, open knowledge, and open learning. The XO-4 laptop, the machine of this project, runs a slimmed-down version of Fedora 17 as its operating system. Red Hat is the largest contributor to the GNOME desktop environment. It has several employees working full-time on Evolution, the official personal information manager for GNOME. Dogtail, an open-source automated graphical user interface (GUI) test framework initially developed by Red Hat, consists of free software released under the GNU General Public License (GPL) and is written in Python. It allows developers to build and test their applications. Red Hat announced the release of Dogtail at the 2006 Red Hat Summit. Red Hat MRG is a clustering product intended for integrated high-performance computing (HPC). The acronym MRG stands for "Messaging Realtime Grid". Red Hat Enterprise MRG replaces the Red Hat Enterprise Linux RHEL, a Linux distribution developed by Red Hat, kernel in order to provide extra support for real-time computing, together with middleware support for message brokerage and scheduling workload to local or remote virtual machines, grid computing, and cloud computing. , Red Hat works with the Condor High-Throughput Computing System community and also provides support for the software. The Tuna performance-monitoring tool runs in the MRG environment. Red Hat produces the online publication Opensource.com since January 20, 2010. The site highlights ways open-source principles apply in domains other than software development. The site tracks the application of open-source philosophy to business, education, government, law, health, and life. The company originally produced a newsletter called Under the Brim. Wide Open magazine first appeared in March 2004, as a means for Red Hat to share technical content with subscribers on a regular basis. The Under the Brim newsletter and Wide Open magazine merged in November 2004, to become "Red Hat Magazine". In January 2010, "Red Hat Magazine" became Opensource.com. In 2007, Red Hat announced that it had reached an agreement with some free software and open-source (FOSS) companies that allowed it to make a distribution portal called Red Hat Exchange, reselling FOSS software with the original branding intact. However, by 2010, Red Hat had abandoned the Exchange program to focus their efforts more on their Open Source Channel Alliance which began in April 2009. Red Hat Single Sign On is a software product to allow single sign-on with Identity Management and Access Management aimed at modern applications and services. There is an ongoing Open source project alongside Red Hat SSO, that is Keycloak. Keycloak is basically the community version from Red Hat SSO. Red Hat Single Sign On 7.3 is the latest version available. Red Hat Subscription Management (RHSM) combines content delivery with subscription management. Red Hat is the largest contributor to the CEPH Storage (SDS) project : Block, File & Object Storage which runs on industry-standard x86 servers Red Hat operates OpenShift, a cloud computing platform as a service, supporting applications written in Node.js, PHP, Perl, Python, Ruby, JavaEE and more. On July 31, 2018, Red Hat announced the release of Istio 1.0, a microservices management program used in tandem with the Kubernetes platform. The software purports to provide "traffic management, service identity and security, policy enforcement and telemetry" services in order to streamline Kubernetes use under the various Fedora-based operating systems. Red Hat's Brian Redbeard Harring described Istio as "aiming to be a control plane, similar to the Kubernetes control plane, for configuring a series of proxy servers that get injected between application components". Red Hat markets a version of OpenStack which helps manage a data center in the manner of cloud computing. Red Hat CloudForms provides management of virtual machines, instances and containers based on VMware vSphere, Red Hat Virtualization, Microsoft Hyper-V, OpenStack, Amazon EC2, Google Cloud Platform, Microsoft Azure, and Red Hat OpenShift. CloudForms is based on the ManageIQ project that Red Hat open sourced. Code in ManageIQ is from the over acquisition of ManageIQ in 2012. Red Hat contributes, with several software developers, to LibreOffice, a free and open-source office suite. Red Hat has some employees working full-time on other free and open-source software projects that are not Red Hat products, such as two full-time employees working on the free software "radeon" (David Airlie and Jerome Glisse) and one full-time employee working on the free software "nouveau" graphic drivers. Another such project is AeroGear, an open-source project that brings security and development expertise to cross-platform enterprise mobile development. Red Hat also organises "Open Source Day" events where multiple partners show their open-source technologies. Subscribers have access to: Over and above Red Hat's major products and acquisitions, Red Hat programmers have produced software programming-tools and utilities to supplement standard Unix and Linux software. Some of these Red Hat "products" have found their way from specifically Red Hat operating environments via open-source channels to a wider community. Such utilities include: The Red Hat website lists the organization's major involvements in free and open-source software projects. Community projects under the aegis of Red Hat include: Red Hat Czech s.r.o. is a research and development arm of Red Hat, based in Brno, Czech Republic. The subsidiary was formed in 2006 and has 1,180 employees (2019). Red Hat chose to enter the Czech Republic in 2006 over other locations due to the country's embrace of open-source. The subsidiary expanded in 2017 to a second location in the Brno Technology Park to accommodate an additional 350 employees. In 2016, Red Hat Czech reported revenue of CZK 1,002 million (FY 2016), and net income of CZK 123 million (FY 2016), with assets of CZK 420 million (FY 2016)|CZK 325 million (FY 2015). The group was named the "Most progressive employer of the year" in the Czech Republic in 2010, and the "Best Employer in the Czech Republic" for large scale companies in 2011 by Aon. In 2000, Red Hat created the subsidiary Red Hat India to deliver Red Hat software, support, and services to Indian customers. Colin Tenwick, former vice president and general manager of Red Hat EMEA, said Red Hat India was opened "in response to the rapid adoption of Red Hat Linux in the subcontinent. Demand for open-source solutions from the Indian markets is rising and Red Hat wants to play a major role in this region." Red Hat India has worked with local companies to enable adoption of open-source technology in both government and education. In 2006, Red Hat India had a distribution network of more than 70 channel partners spanning 27 cities across India. Red Hat India's channel partners included MarkCraft Solutions, Ashtech Infotech Pvt Ltd, Efensys Technologies, Embee Software, Allied Digital Services, and Softcell Technologies. Distributors include Integra Micro Systems and Ingram Micro. Red Hat's first major acquisition involved Delix Computer GmbH-Linux Div, the Linux-based operating-system division of Delix Computer, a German computer company, on July 30, 1999. Red Hat acquired Cygnus Solutions, a company that provided commercial support for free software, on January 11, 2000 – it was the company's largest acquisition, for . Michael Tiemann, co-founder of Cygnus, served as the chief technical officer of Red Hat after the acquisition. Red Hat made the most acquisitions in 2000 with five: Cygnus Solutions, Bluecurve, Wirespeed Communications, Hell's Kitchen Systems, and C2Net. On June 5, 2006, Red Hat acquired open-source middleware provider JBoss for and integrated it as its own division of Red Hat. On December 14, 1998, Red Hat made its first divestment, when Intel and Netscape acquired undisclosed minority stakes in the company. The next year, on March 9, 1999, Compaq, IBM, Dell and Novell each acquired undisclosed minority stakes in Red Hat.
https://en.wikipedia.org/wiki?curid=26386
Reno, Nevada Reno ( ) is a city in the northwest section of the U.S. state of Nevada, approximately from Lake Tahoe, known as "The Biggest Little City in the World". Known for its casino and tourism industry, Reno is the county seat and largest city of Washoe County and sits in a high desert river valley at the foot of the Sierra Nevada. Its downtown area (along with Sparks) occupies a valley informally known as the Truckee Meadows, which due to large-scale investments from Seattle and Bay Area companies such as Amazon, Tesla, Panasonic, Microsoft, Apple and Google has become a new major technology hub in the United States. The city is named after Union Major General Jesse L. Reno, who was killed in action during the American Civil War at the Battle of South Mountain on Fox's Gap. Reno is part of the Reno–Sparks metropolitan area, the second most-populous metropolitan area in Nevada after the Las Vegas Valley. Known as Greater Reno, it includes of Washoe, Storey, and Lyon counties as well as the state capital, Carson City. Archaeological finds place the eastern border for the prehistoric Martis people in the Reno area. As early as the mid 1850s a few pioneers settled in the Truckee Meadows, a relatively fertile valley through which the Truckee River made its way from Lake Tahoe to Pyramid Lake. In addition to subsistence farming, these early residents could pick up business from travelers along the California Trail, which followed the Truckee westward, before branching off towards Donner Lake, where the formidable obstacle of the Sierra Nevada began. Gold was discovered in the vicinity of Virginia City in 1850, and a modest mining community developed, but the discovery of silver in 1859 at the Comstock Lode led to a mining rush, and thousands of emigrants left their homes, bound for the West, hoping to find a fortune. To provide the necessary connection between Virginia City and the California Trail, Charles W. Fuller built a log toll bridge across the Truckee River in 1859. A small community that would service travelers soon grew near the bridge. After two years, Fuller sold the bridge to Myron C. Lake, who continued to develop the community by adding a grist mill, kiln, and livery stable to the hotel and eating house. He renamed it "Lake's Crossing". Most of what is present-day western Nevada was formed as the Nevada Territory from part of Utah Territory in 1861. By January 1863, the Central Pacific Railroad (CPRR) had begun laying tracks east from Sacramento, California, eventually connecting with the Union Pacific Railroad at Promontory, Utah, to form the First Transcontinental Railroad. Lake deeded land to the CPRR in exchange for its promise to build a depot at Lake's Crossing. In 1864, Washoe County was consolidated with Roop County, and Lake's Crossing became the county's largest town. Lake had earned himself the title "founder of Reno". Once the railroad station was established, the town of Reno officially came into being on May 9, 1868. CPRR construction superintendent Charles Crocker named the community after Major General Jesse Lee Reno, a Union officer killed in the American Civil War at the Battle of South Mountain. In 1871, Reno became the county seat of the newly expanded Washoe County, replacing the county seat in Washoe City. However, political power in Nevada remained with the mining communities, first Virginia City and later Tonopah and Goldfield. The extension of the Virginia and Truckee Railroad to Reno in 1872 provided a boost to the new city's economy. In the following decades, Reno continued to grow and prosper as a business and agricultural center and became the principal settlement on the transcontinental railroad between Sacramento and Salt Lake City. As the mining boom waned early in the 20th century, Nevada's centers of political and business activity shifted to the non-mining communities, especially Reno and Las Vegas, and today the former mining metropolises stand as little more than ghost towns. Despite this, Nevada is still the third-largest gold producer in the world, after South Africa and Australia; the state yielded 6.9 percent of the world's supply in 2005 world gold production. The "Reno Arch" was erected on Virginia Street in 1926 to promote the upcoming Transcontinental Highways Exposition of 1927. The arch included the words "Nevada's Transcontinental Highways Exposition" and the dates of the exposition. After the exposition, the Reno City Council decided to keep the arch as a permanent downtown gateway, and Mayor E.E. Roberts asked the citizens of Reno to suggest a slogan for the arch. No acceptable slogan was received until a $100 prize was offered, and G.A. Burns of Sacramento was declared the winner on March 14, 1929, with "Reno, The Biggest Little City in the World". Reno took a leap when the state of Nevada legalized open-gambling on March 19, 1931, along with the passage of even more liberal divorce laws than places like Hot Springs, Arkansas, offered. No other state offered what Nevada had in the 1930s, and casinos like the Bank Club and Palace were popular. The new Nevada divorce laws, passed circa 1927, allowed people to divorce after six weeks of residency instead of six months. People wishing to divorce stayed in hotels, houses, and/or dude ranches. Most divorcees left Nevada when their divorces were finalized, while some stayed. Within a few years, the Bank Club, owned by George Wingfield, Bill Graham, and Jim McKay, was the state's largest employer and the largest casino in the world. Wingfield owned most of the buildings in town that housed gaming and took a percentage of the profits, along with his rent. Ernie Pyle once wrote in one of his columns, "All the people you saw on the streets in Reno were obviously there to get divorces." In Ayn Rand's novel "The Fountainhead", published in 1943, the New York-based female protagonist tells a friend, "I am going to Reno," which is taken as a different way of saying "I am going to divorce my husband." Among others, the Belgian-French writer Georges Simenon, at the time living in the U.S., came to Reno in 1950 in order to divorce his first wife. The divorce business eventually died after circa 1970, as the other states fell in line by passing their own laws easing the requirements for divorce, but gambling continued as a major Reno industry. While gaming pioneers like "Pappy" and Harold Smith of Harold's Club and Bill Harrah of the soon-to-dominate Harrah's casino set up shop in the 1930s, the war years of the 1940s cemented Reno as the place to play for two decades. Beginning in the 1950s, the need for economic diversification beyond gaming fueled a movement for more lenient business taxation. A disaster occurred on the afternoon of February 5, 1957, when an explosion ripped through the heart of downtown. At 1:03 pm, two explosions, caused by natural gas leaking into the maze of pipes and ditches under the city, and an ensuing fire destroyed five buildings in the vicinity of Sierra and First streets along the Truckee River. The disaster killed two people and injured forty-nine. The first explosion hit under the block of shops on the west side of Sierra Street (now the site of the Century Riverside), the second, across Sierra Street, now the site of the Palladio. The presence of a main east–west rail line, the emerging interstate highway system, favorable state tax climate, and relatively inexpensive land created good conditions for warehousing and distribution of goods. In the 1980s, Indian gaming rules were relaxed, and starting in 2000, Californian Native casinos began to cut into casino revenues. Major new construction projects have been completed in the Reno and Sparks areas. A few new luxury communities were recently built in Truckee, California, approximately west of Reno on Interstate 80. Reno also is an outdoor recreation destination, due to its close proximity to the Sierra Nevada, Lake Tahoe, and numerous ski resorts in the region. In 2018, the city officially changed its flag after a local contest was held. Wetlands are an important part of the Reno/Tahoe area. They act as a natural filter for the solids that come out of the water treatment plant. Plant roots absorb nutrients from the water and naturally filter it. Wetlands are home to over 75% of the species in the Great Basin. However, the area's wetlands are at risk of being destroyed due to development around the city. While developers build on top of the wetlands they fill them with dirt, destroying the habitat they create for the plants and animals. Washoe County has devised a plan that will help protect these ecosystems: mitigation. In the future, when developers try to build over a wetland, they will be responsible for creating another wetland near Washoe Lake. The Truckee River is Reno's primary source of drinking water. It supplies Reno with of water a day during the summer, and of water per day in the winter. Before the water goes to the homes around the Reno area, it must go to one of two water treatment plants, Chalk Bluff or Glendale Water Treatment Plant. To help save water, golf courses in Reno have been using treated effluent water rather than treated water from one of Reno's water plants. The Reno-Sparks wastewater treatment plant discharges tertiary treated effluent to the Truckee River. In the 1990s this capacity was increased from 20 to 30 million U.S. gallons (70 to 110 million liters) per day. While treated, the effluent contains suspended solids, nitrogen, and phosphorus, aggravating water quality concerns of the river and its receiving waters of Pyramid Lake. Local agencies working with the Environmental Protection Agency have developed several watershed management strategies to accommodate this expanded discharge. To accomplish this successful outcome, the DSSAM Model was developed and calibrated for the Truckee River to analyze the most cost-effective available management strategy set. The resulting management strategies included measures such as land use controls in the Lake Tahoe basin, urban runoff controls in Reno and Sparks, and best management practices for wastewater discharge. The Reno area is often subject to wildfires that cause property damage and sometimes loss of life. In August 1960, the Donner Ridge fire resulted in a loss of electricity to the city for four days. In November 2011, arcing from powerlines caused a fire in Caughlin in southwest Reno that destroyed 26 homes and killed one man. Just two months later, in January 2012, a fire in Washoe Drive sparked by fireplace ashes destroyed 29 homes and killed one woman. Around 10,000 residents were evacuated, and a state of emergency was declared. The fires came at the end of Reno's longest recorded dry spell. Reno is just east of the Sierra Nevada on the western edge of the Great Basin at an elevation of about above sea level. Numerous faults exist throughout the region. Most of these are normal (vertical motion) faults associated with the uplift of the various mountain ranges, including the Sierra Nevada. In February 2008, an earthquake swarm began to occur, lasting for several months, and with the largest quake registering at 4.9 on the Richter magnitude scale, although some geologic estimates put it at 5.0. The earthquakes were centered on the Somersett community in western Reno near Mogul and Verdi. Many homes in these areas were damaged. Reno sits in the rain shadow of the Sierra Nevada mountain range. Annual rainfall averages . Reno features a semi-arid climate (Köppen: "BSk)" due to its low evapotranspiration. The city experiences cool to cold winters, and hot summers. Annual precipitation has ranged from in 1947 to in 1983. The most precipitation in one month was in December 1955 and the most precipitation in 24 hours was on January 21, 1943. Winter has snowfall which is usually light to moderate but can be heavy some days, averaging annually. Snowfall varies with the lowest amounts (roughly 19–23 inches annually) at the lowest part of the valley at and east of the Reno–Tahoe International Airport at , while the foothills of the Carson Range to the west ranging from in elevation just a few miles west of downtown can receive up to two to three times as much annual snowfall. The mountains of the Virginia Range to the east can receive more summer thunderstorms and precipitation, and around twice as much annual snowfall above . However, snowfall increases in the Virginia Range are less dramatic as elevation climbs than in the Carson Range to the west, because the Virginia Range is well within the rain shadow of the Sierra Nevada and Carson Range. The most snowfall in the city in one year was in 1971, and the most snowfall in one month was in March 1952. Most rainfall occurs in winter and spring. The city has 300 days of sunshine per year. Summer thunderstorms can occur between April and October. The eastern side of town and the mountains east of Reno tend to be prone to thunderstorms more often, and these storms may be severe because an afternoon downslope west wind, called a "Washoe Zephyr", can develop in the Sierra Nevada, causing air to be pulled down in the Sierra Nevada and Reno, destroying or preventing thunderstorms, but the same wind can push air upward against the Virginia Range and other mountain ranges east of Reno, creating powerful thunderstorms. The monthly daily average temperature ranges from in December to in July, with the diurnal temperature variation reaching in summer, still lower than much of the high desert to the east. There are 3.9 days of + highs, 58 days of + highs, and 2.5 nights with sub- lows annually; the temperature does not rise above freezing on only 5.1 days. The all-time record high temperature is , which occurred on July 10 and 11, 2002, and again on July 5, 2007. The all-time record low temperature is , which occurred on January 21, 1916. In addition, the region is windy throughout the year; observers such as Mark Twain have commented about the "Washoe Zephyr", northwestern Nevada's distinctive wind. As of the census of 2010, there were 225,221 people, 90,924 households, and 51,112 families residing in the city. The population density was 2,186.6 per square mile (844.2/km²). There were 102,582 housing units at an average density of 995.9 per square mile (384.5/km²). The city's racial makeup was 74.2% White, 2.9% African American, 1.3% Native American, 6.3% Asian, 0.7% Pacific Islander, 10.5% some other race, and 4.2% from two or more races. Hispanic or Latino of any race were 24.3% of the population. Non-Hispanic Whites were 62.5% of the population in 2010, down from 88.5% in 1980. At the 2010 census, there were 90,924 households, of which 29.8% had children under the age of 18 living with them, 38.4% were headed by married couples living together, 11.8% had a female householder with no husband present, and 43.8% were non-families. 32.1% of all households were made up of individuals, and 9.7% were someone living alone who was 65 years of age or older. The average household size was 2.43, and the average family size was 3.10. In the city, the 2010 population was spread out with 22.8% under the age of 18, 12.5% from 18 to 24, 28.2% from 25 to 44, 24.9% from 45 to 64, and 11.7% who were 65 years of age or older. The median age was 34.6 years. For every 100 females, there were 103.4 males. For every 100 females age 18 and over, there were 102.7 males. In 2011 the city's estimated median household income was $44,846, and the median family income was $53,896. Males had a median income of $42,120 versus $31,362 for females. The city's per capita income was $25,041. About 9.6% of families and 14.4% of the population were below the poverty line, including 15.1% of those under age 18 and 12.8% of those age 65 or over. The population was 180,480 at the 2000 census; in 2010, its population had risen to 225,221, making it the third-largest city in the state after Las Vegas and Henderson, and the largest outside Clark County. Reno lies north of the Nevada state capital, Carson City, and northeast of Lake Tahoe in a shrub-steppe environment. Reno shares its eastern border with the city of Sparks and is the larger of the principal cities of the Reno–Sparks, Nevada Metropolitan Statistical Area (MSA), a metropolitan area that covers Storey and Washoe counties. The MSA had a combined population of 425,417 at the 2010 census. The MSA is combined with the Fernley Micropolitan Statistical Area and the Carson City MSA to form the Reno-Carson City-Fernley Combined Statistical Area, which had a population of 477,397 at the 2010 census. Before the late 1950s, Reno was the gambling capital of the United States, but Las Vegas' rapid growth, American Airlines' 2000 buyout of Reno Air, and the growth of Native American gaming in California have reduced its business. Older casinos were torn down (Mapes Hotel, Fitzgerald's Nevada Club, Primadonna, Horseshoe Club, Harold's Club, Palace Club), or smaller casinos like the Comstock, Sundowner, Golden Phoenix, Kings Inn, Money Tree, Virginian, and Riverboat were either closed or were converted into residential units. Because of its location, Reno has traditionally drawn the majority of its California tourists and gamblers from the San Francisco Bay Area and Sacramento, while Las Vegas has historically served more tourists from Southern California and the Phoenix area. Several local large hotel casinos have shown significant growth and have moved gaming further away from the downtown core. These larger hotel casinos are the Atlantis, the Peppermill and the Grand Sierra Resort. The Peppermill was chosen as the most outstanding Reno gaming/hotel property by "Casino Player" and "Nevada" magazines. In 2005, the Peppermill Reno began a $300 million Tuscan-themed expansion. Reno holds several events throughout the year to draw tourists to the area. They include Hot August Nights (a classic car convention), Street Vibrations (a motorcycle fan gathering and rally), The Great Reno Balloon Race, a Cinco de Mayo celebration, bowling tournaments (held in the National Bowling Stadium), and the Reno Air Races. Several large commercial developments were constructed during the mid-2000s boom, such as The Summit in 2007 and Legends at Sparks Marina in 2008. Reno is the location of the corporate headquarters for several companies, including Braeburn Capital, Hamilton, Server Technology, EE Technologies, and Port of Subs. International Game Technology, Bally Technologies and GameTech have a development and manufacturing presence. Since the turn of the 21st century, greater Reno saw an influx of technology companies entering the area, following major initiatives and investments by investors from Seattle & the Bay Area. After the Great Recession, the state placed an increased focus on economic development. Thousands of new jobs were created. Tesla's Gigafactory is at the Tahoe Reno Industrial Center, and it is the largest building in the world. The Gigafactory purportedly covers 5.8 million square feet. The arrival of several data centers at the Tahoe Reno Industrial Center is further diversifying a region that was best known for distribution and logistics outside gaming and tourism. Switch 's new SuperNAP campus at the Tahoe Reno Industrial Center is shaping up to be the largest data center in the world once completed. Apple is expanding its data center at the adjacent Reno Technology Park and recently built a warehouse on land in downtown Reno. Rackspace is also building a $422 million data center next to Apple. The greater Reno area also hosts distribution facilities for Amazon, Walmart, Petsmart and Zulily. According to Reno's 2016 Comprehensive Annual Financial Report, the top employers in the city are: Reno has several museums. The Nevada Museum of Art is the only American Alliance of Museums (AAM) accredited art museum in Nevada. The National Automobile Museum contains 200 cars that were from the collection of William F. Harrah, including Elvis Presley's 1973 Cadillac Eldorado. Reno also hosts a number of music venues, such as the Nevada Opera, the Pioneer Center for the Performing Arts, the Reno Philharmonic Orchestra, and the Reno Pops Orchestra. The Reno Youth Symphony Orchestra (YSO), affiliated with the Reno Philharmonic, gives talented youth the opportunity to play advanced music and perform nationwide. In 2016 they had the honor of performing at Carnegie Hall. Reno has a public library, a branch of the Washoe County Library System. The Downtown branch of the Washoe County Library was listed on the National Register of Historic Places in 2013. Reno is home to the Reno Aces, the minor league baseball Triple-A affiliate of the Arizona Diamondbacks, playing in Greater Nevada Field, a downtown ballpark opened in 2009. Reno has hosted multiple professional baseball teams in the past, most under the Reno Silver Sox name. The Reno Astros, a former professional, unaffiliated baseball team, played at Moana Stadium until 2009. In basketball, the Reno Bighorns of the NBA G League played at the Reno Events Center from 2008 to 2018. They were primarily an affiliate of the Sacramento Kings throughout its existence. The Sacramento Kings bought the team in 2016 and moved the franchise to become the Stockton Kings in 2018. Reno is host to both amateur and professional combat sporting events such as mixed martial arts and boxing. The "Fight of the Century" between Jack Johnson and James J. Jeffries was held in Reno in 1910. Boxer Ray Mancini fought four of his last five fights in Reno against Bobby Chacon, Livingstone Bramble, Hector Camacho, and Greg Haugen. Reno expected to be the future home of an ECHL ice hockey team, named the Reno Raiders, but construction on a suitable arena never began. The franchise was dormant since 1998, when it was named the Reno Rage, and earlier the Reno Renegades, and played in the now-defunct West Coast Hockey League (WCHL). In 2016, Reno was removed from the ECHL's Future Markets page. The Reno–Tahoe Open is northern Nevada's only PGA Tour event, held at Montrêux Golf & Country Club in Reno. As part of the FedEx Cup, the tournament follows 132 PGA Tour professionals competing for a share of the event's $3 million purse. The Reno-Tahoe Open Foundation has donated more than $1.8 million to local charities. Reno has a college sports scene, with the Nevada Wolf Pack appearing in football bowl games and an Associated Press and Coaches Poll Top Ten ranking in basketball in 2018. In 2004, the city completed a $1.5 million whitewater park on the Truckee River in downtown Reno which hosts whitewater events throughout the year. The course runs Class 2 and 3 rapids with year-round public access. The north channel features more aggressive rapids, drop pools and "holes" for rodeo kayak-type maneuvers. The milder south channel is set up as a kayak slalom course and a beginner area. Reno is home to two roller derby teams, the Battle Born Derby Demons and the Reno Roller Girls. The Battle Born Derby Demons compete on flat tracks locally and nationally. They are the only derby team locally to compete in a national Derby league. Reno is the home of the National Bowling Stadium, which hosts the United States Bowling Congress (USBC) Open Championships every three years. Reno is home to a variety of recreation activities including both seasonal and year-round. In the summer, Reno locals can be found near three major bodies of water: Lake Tahoe, the Truckee River, and Pyramid Lake. The Truckee River originates at Lake Tahoe and flows west to east through the center of downtown Reno before terminating at Pyramid Lake to the north. The river is a major part of Artown, held in the summer at Wingfield Park. Washoe Lake is a popular kite and windsurfing location because of its high wind speeds during the summer. Skiing and snowboarding are among the most popular winter sports and draw many tourists. There are 18 ski resorts (8 major resorts) as close as and as far as from the Reno–Tahoe International Airport, including Northstar California, Sierra-at-Tahoe, Alpine Meadows, Squaw Valley, Sugar Bowl, Diamond Peak, Heavenly Mountain, and Mount Rose. Other popular Reno winter activities include snowshoeing, ice skating, and snowmobiling. There are many bike paths to ride in the summer time. Lake Tahoe hosts international bike competitions each summer. The Reno Air Races, also known as the National Championship Air Races, are held each September at the Reno Stead Airport. Reno has a democratic municipal government. The city council is the core of the government, with seven members. Five of these council people represent districts of Reno, and are vetted in the primary by the citizens of each district. In general, the top two vote earners in each ward make the ballot for the citywide election. The other two council members are the at-large member, who represents the entire city, and the mayor, who is elected by the people of the city. The council has several duties, including setting priorities for the city, promoting communication with the public, planning development, and redevelopment. There is an elected city attorney who is responsible for civil and criminal cases. The City Attorney represents the city government in court, and prosecutes misdemeanors. The city's charter calls for a council-manager form of government, meaning the council appoints only two positions, the city manager, who implements and enforces the policies and programs the council approves, and the city clerk. The city manager is in charge of the budget and workforce for all city programs. The city clerk, who records the proceedings of the council, makes appointments for the council, and makes sure efficient copying and printing services are available. In 2010, there was a ballot question asking whether the Reno city government and the Washoe County government should explore the idea of becoming one combined governmental body. Fifty-four percent of voters approved of the ballot measure to make an inquiry into consolidating the governments. The city of Reno is protected by the Reno Fire Department (RFD) manning 14 fire stations. The Reno Fire Department (RFD) provides all-risk emergency service to the City of Reno residents. All-risk emergency service is the national model of municipal fire departments, providing the services needed in the most efficient way possible. The department provides paramedic-level service to the citizens and visitors of Reno. This is the highest level of emergency medical care available in the field. In addition to responding to fires, whether they occur in structures, vegetation/brush or vehicles, the fire department also provides rescue capabilities for almost any type of emergency situation. This includes quick and efficient emergency medical care for the citizens; a hazardous materials team capable of identifying unknown materials and controlling a release disaster; and preparedness and management of large-scale incidents. Maintaining this level of service requires nearly constant training of personnel. This training maintains both the skills needed to operate safely in emergency environments and the physical fitness necessary to reduce the likelihood and severity of injuries. The minimum annual-training requirement to maintain firefighting and medical skills is 240 hours per year. Special teams and company-level drills add significantly to that number of hours. Public education is provided by the Washoe County School District. Reno has many charter schools, which include Academy for Career Education, serving grades 10–12, opened 2002; Alpine Academy Charter High School, serving grades 9–12, opened 2009; Bailey Charter Elementary School, serving grades K-6, opened 2001; Coral Academy of Science, serving grades K-12; Davidson Academy, serving grades 6–12, opened 2006; High Desert Montessori School, serving grades PreK-7, opened 2002; I Can Do Anything Charter School, serving grades 9–12, opened 2000; Rainshadow Community Charter High School, serving grades 9–12, opened 2003; Sierra Nevada Academy Charter School, serving grades PreK-8, opened 1999; and TEAM A (Together Everyone Achieves More Academy), serving grades 9–12, opened 2004. Reno has a few private elementary schools such as Legacy Christian School, Excel Christian School, St. Nicholas Orthodox Academy, Lamplight Christian School, and Nevada Sage Waldorf School as well as private high schools, the largest of which are Bishop Manogue High School and Sage Ridge School. Reno was historically served by the Victory Highway and a branch of the Lincoln Highway. After the formation of the U.S. Numbered Highways system, U.S. Route 40 was routed along 4th Street through downtown Reno, before being replaced by Interstate 80. The primary north–south highway through Reno is U.S. Route 395/Interstate 580. The Regional Transportation Commission of Washoe County (RTC) has a bus system that provides intracity buses, intercity buses to Carson City, and an on-demand shuttle service for disabled persons. The system has its main terminal on 4th Street in downtown Reno and secondary terminals in Sparks and at Meadowood Mall in south Reno. Numerous shuttle and excursion services are offered connecting the Reno–Tahoe International Airport to various destinations: Greyhound stops at a downtown terminal. Megabus stopped at the Silver Legacy Reno, but has since discontinued service to Reno. Reno was historically a stopover along the First Transcontinental Railroad; the modern Overland Route continues to run through Reno. Reno was additionally the southern terminus of the Nevada–California–Oregon Railway (NCO) and the northern terminus of the Virginia and Truckee Railroad. Using the NCO depot and right of way, the Western Pacific Railroad also provided rail service to Reno. In the early 20th century, Reno also had a modest streetcar system. Downtown Reno has two historic train depots, the inactive Nevada-California-Oregon Railroad Depot and the still active Amtrak depot, originally built by the Southern Pacific Railroad. Amtrak provides daily passenger service to Reno via the "California Zephyr" and multiple Amtrak Thruway Motorcoaches connecting to trains departing from Sacramento. The city is served by Reno–Tahoe International Airport, with general aviation traffic handled by Reno Stead Airport. Reno–Tahoe International Airport is the second busiest commercial airport in the state of Nevada after McCarran International Airport in Las Vegas. Reno was the hub and headquarters of the defunct airline Reno Air. The Truckee Meadows Water Authority provides potable water for the city. The Truckee River is the primary water source, with purification occurring at two plants, Chalk Bluff and Glendale. The Chalk Bluff plant's main intakes are west of Reno in Verdi, with the water flowing through a series of flumes and ditches to the plant. Alternative intakes are below the plant along the banks of the Truckee River itself. The Glendale plant is alongside the river, and is fed by a rock and concrete rubble diversion dam a short distance upstream. Sewage treatment for most of the Truckee Meadows region takes place at the Truckee Meadows Water Reclamation Facility at the eastern edge of the valley. Treated effluent returns to the Truckee River by way of Steamboat Creek. NV Energy, formerly Sierra Pacific, provides electric power and natural gas. Power comes from multiple sources, including Tracy-Clark Station to the east, and the Steamboat Springs binary cycle power plants at the southern end of town. The young adult author Ellen Hopkins has written a series of novels called "Crank" set in Reno. Also, many of the short stories included in Claire Vaye Watkins' collection "Battleborn" are set in the city. American songwriter Richard Fariña composed a song named "Reno Nevada"; it was first released on Richard & Mimi Fariña's debut album "Celebrations For A Grey Day" in 1965. The song was covered by Fairport Convention in 1968 and by Iain Matthews in 1971. Thomas Dolby composed a song named "Road to Reno" as part of his "A Map of the Floating City" album, released in 2011. The comedy TV series "Reno 911!" takes place in Reno. The videogame "Fallout 2" features the fictional city of New Reno, based upon the real-world location. Movies filmed in Reno include: Music videos filmed in Reno include: Reno has eight sister cities: Friendship cities
https://en.wikipedia.org/wiki?curid=26388
Riemann integral In the branch of mathematics known as real analysis, the Riemann integral, created by Bernhard Riemann, was the first rigorous definition of the integral of a function on an interval. It was presented to the faculty at the University of Göttingen in 1854, but not published in a journal until 1868. For many functions and practical applications, the Riemann integral can be evaluated by the fundamental theorem of calculus or approximated by numerical integration. The Riemann integral is unsuitable for many theoretical purposes. Some of the technical deficiencies in Riemann integration can be remedied with the Riemann–Stieltjes integral, and most disappear with the Lebesgue integral, though the latter does not have a satisfactory treatment of improper integrals. The gauge integral is a generalisation of the Lebesgue integral that is at once closer to the Riemann integral. These more general theories allow for the integration of more "jagged" or "highly oscillating" functions whose Riemann integral does not exist; but the theories give the same value as the Riemann integral when it does exist. Let be a non-negative real-valued function on the interval , and let be the region of the plane under the graph of the function and above the interval (see the figure on the top right). We are interested in measuring the area of . Once we have measured it, we will denote the area by: The basic idea of the Riemann integral is to use very simple approximations for the area of . By taking better and better approximations, we can say that "in the limit" we get exactly the area of under the curve. Note that where can be both positive and negative, the definition of is modified so that the integral corresponds to the "signed area" under the graph of : that is, the area above the -axis minus the area below the -axis. A partition of an interval is a finite sequence of numbers of the form Each is called a sub-interval of the partition. The mesh or norm of a partition is defined to be the length of the longest sub-interval, that is, A tagged partition of an interval is a partition together with a finite sequence of numbers subject to the conditions that for each , . In other words, it is a partition together with a distinguished point of every sub-interval. The mesh of a tagged partition is the same as that of an ordinary partition. Suppose that two partitions and are both partitions of the interval . We say that is a refinement of if for each integer , with , there exists an integer such that and such that for some with . Said more simply, a refinement of a tagged partition breaks up some of the sub-intervals and adds tags to the partition where necessary, thus it "refines" the accuracy of the partition. We can define a partial order on the set of all tagged partitions by saying that one tagged partition is greater than or equal to another if the former is a refinement of the latter. Let be a real-valued function defined on the interval . The "Riemann sum" of with respect to the tagged partition together with is Each term in the sum is the product of the value of the function at a given point and the length of an interval. Consequently, each term represents the (signed) area of a rectangle with height and width . The Riemann sum is the (signed) area of all the rectangles. Closely related concepts are the "lower and upper Darboux sums". These are similar to Riemann sums, but the tags are replaced by the infimum and supremum (respectively) of on each sub-interval: If is continuous, then the lower and upper Darboux sums for an untagged partition are equal to the Riemann sum for that partition, where the tags are chosen to be the minimum or maximum (respectively) of on each subinterval. (When is discontinuous on a subinterval, there may not be a tag that achieves the infimum or supremum on that subinterval.) The Darboux integral, which is similar to the Riemann integral but based on Darboux sums, is equivalent to the Riemann integral. Loosely speaking, the Riemann integral is the limit of the Riemann sums of a function as the partitions get finer. If the limit exists then the function is said to be integrable (or more specifically Riemann-integrable). The Riemann sum can be made as close as desired to the Riemann integral by making the partition fine enough. One important requirement is that the mesh of the partitions must become smaller and smaller, so that in the limit, it is zero. If this were not so, then we would not be getting a good approximation to the function on certain subintervals. In fact, this is enough to define an integral. To be specific, we say that the Riemann integral of equals if the following condition holds: For all , there exists such that for any tagged partition and whose mesh is less than , we have Unfortunately, this definition is very difficult to use. It would help to develop an equivalent definition of the Riemann integral which is easier to work with. We develop this definition now, with a proof of equivalence following. Our new definition says that the Riemann integral of equals if the following condition holds: For all , there exists a tagged partition and such that for any tagged partition and which is a refinement of and , we have Both of these mean that eventually, the Riemann sum of with respect to any partition gets trapped close to . Since this is true no matter how close we demand the sums be trapped, we say that the Riemann sums converge to . These definitions are actually a special case of a more general concept, a net. As we stated earlier, these two definitions are equivalent. In other words, works in the first definition if and only if works in the second definition. To show that the first definition implies the second, start with an , and choose a that satisfies the condition. Choose any tagged partition whose mesh is less than . Its Riemann sum is within of , and any refinement of this partition will also have mesh less than , so the Riemann sum of the refinement will also be within of . To show that the second definition implies the first, it is easiest to use the Darboux integral. First, one shows that the second definition is equivalent to the definition of the Darboux integral; for this see the article on Darboux integration. Now we will show that a Darboux integrable function satisfies the first definition. Fix , and choose a partition such that the lower and upper Darboux sums with respect to this partition are within of the value of the Darboux integral. Let If , then is the zero function, which is clearly both Darboux and Riemann integrable with integral zero. Therefore, we will assume that . If , then we choose such that If , then we choose to be less than one. Choose a tagged partition and with mesh smaller than . We must show that the Riemann sum is within of . To see this, choose an interval . If this interval is contained within some , then where and are respectively, the infimum and the supremum of "f" on . If all intervals had this property, then this would conclude the proof, because each term in the Riemann sum would be bounded by a corresponding term in the Darboux sums, and we chose the Darboux sums to be near . This is the case when , so the proof is finished in that case. Therefore, we may assume that . In this case, it is possible that one of the is not contained in any . Instead, it may stretch across two of the intervals determined by . (It cannot meet three intervals because is assumed to be smaller than the length of any one interval.) In symbols, it may happen that (We may assume that all the inequalities are strict because otherwise we are in the previous case by our assumption on the length of .) This can happen at most times. To handle this case, we will estimate the difference between the Riemann sum and the Darboux sum by subdividing the partition at . The term in the Riemann sum splits into two terms: Suppose, without loss of generality, that . Then so this term is bounded by the corresponding term in the Darboux sum for . To bound the other term, notice that It follows that, for some (indeed any) , Since this happens at most times, the distance between the Riemann sum and a Darboux sum is at most . Therefore, the distance between the Riemann sum and is at most . Let formula_17 be the function which takes the value 1 at every point. Any Riemann sum of on will have the value 1, therefore the Riemann integral of on is 1. Let formula_18 be the indicator function of the rational numbers in ; that is, formula_19 takes the value 1 on rational numbers and 0 on irrational numbers. This function does not have a Riemann integral. To prove this, we will show how to construct tagged partitions whose Riemann sums get arbitrarily close to both zero and one. To start, let and be a tagged partition (each is between and ). Choose . The have already been chosen, and we can't change the value of at those points. But if we cut the partition into tiny pieces around each , we can minimize the effect of the . Then, by carefully choosing the new tags, we can make the value of the Riemann sum turn out to be within of either zero or one. Our first step is to cut up the partition. There are of the , and we want their total effect to be less than . If we confine each of them to an interval of length less than , then the contribution of each to the Riemann sum will be at least and at most . This makes the total sum at least zero and at most . So let be a positive number less than . If it happens that two of the are within of each other, choose smaller. If it happens that some is within of some , and is not equal to , choose smaller. Since there are only finitely many and , we can always choose sufficiently small. Now we add two cuts to the partition for each . One of the cuts will be at , and the other will be at . If one of these leaves the interval [0, 1], then we leave it out. will be the tag corresponding to the subinterval If is directly on top of one of the , then we let be the tag for both intervals: We still have to choose tags for the other subintervals. We will choose them in two different ways. The first way is to always choose a rational point, so that the Riemann sum is as large as possible. This will make the value of the Riemann sum at least . The second way is to always choose an irrational point, so that the Riemann sum is as small as possible. This will make the value of the Riemann sum at most . Since we started from an arbitrary partition and ended up as close as we wanted to either zero or one, it is false to say that we are eventually trapped near some number , so this function is not Riemann integrable. However, it is Lebesgue integrable. In the Lebesgue sense its integral is zero, since the function is zero almost everywhere. But this is a fact that is beyond the reach of the Riemann integral. There are even worse examples. formula_19 is equivalent (that is, equal almost everywhere) to a Riemann integrable function, but there are non-Riemann integrable bounded functions which are not equivalent to any Riemann integrable function. For example, let be the Smith–Volterra–Cantor set, and let be its indicator function. Because is not Jordan measurable, is not Riemann integrable. Moreover, no function equivalent to is Riemann integrable: , like , must be zero on a dense set, so as in the previous example, any Riemann sum of has a refinement which is within of 0 for any positive number . But if the Riemann integral of exists, then it must equal the Lebesgue integral of , which is . Therefore, is not Riemann integrable. It is popular to define the Riemann integral as the Darboux integral. This is because the Darboux integral is technically simpler and because a function is Riemann-integrable if and only if it is Darboux-integrable. Some calculus books do not use general tagged partitions, but limit themselves to specific types of tagged partitions. If the type of partition is limited too much, some non-integrable functions may appear to be integrable. One popular restriction is the use of "left-hand" and "right-hand" Riemann sums. In a left-hand Riemann sum, for all , and in a right-hand Riemann sum, for all . Alone this restriction does not impose a problem: we can refine any partition in a way that makes it a left-hand or right-hand sum by subdividing it at each . In more formal language, the set of all left-hand Riemann sums and the set of all right-hand Riemann sums is cofinal in the set of all tagged partitions. Another popular restriction is the use of regular subdivisions of an interval. For example, the th regular subdivision of consists of the intervals Again, alone this restriction does not impose a problem, but the reasoning required to see this fact is more difficult than in the case of left-hand and right-hand Riemann sums. However, combining these restrictions, so that one uses only left-hand or right-hand Riemann sums on regularly divided intervals, is dangerous. If a function is known in advance to be Riemann integrable, then this technique will give the correct value of the integral. But under these conditions the indicator function formula_19 will appear to be integrable on with integral equal to one: Every endpoint of every subinterval will be a rational number, so the function will always be evaluated at rational numbers, and hence it will appear to always equal one. The problem with this definition becomes apparent when we try to split the integral into two pieces. The following equation ought to hold: If we use regular subdivisions and left-hand or right-hand Riemann sums, then the two terms on the left are equal to zero, since every endpoint except 0 and 1 will be irrational, but as we have seen the term on the right will equal 1. As defined above, the Riemann integral avoids this problem by refusing to integrate formula_26 The Lebesgue integral is defined in such a way that all these integrals are 0. The Riemann integral is a linear transformation; that is, if and are Riemann-integrable on and and are constants, then Because the Riemann integral of a function is a number, this makes the Riemann integral a linear functional on the vector space of Riemann-integrable functions. A bounded function on a compact interval is Riemann integrable if and only if it is continuous almost everywhere (the set of its points of discontinuity has measure zero, in the sense of Lebesgue measure). This is known as the or Lebesgue's criterion for Riemann integrability or the Riemann–Lebesgue theorem. The criterion has "nothing to do" with the Lebesgue integral. It is due to Lebesgue and uses his measure zero, but makes use of neither Lebesgue's general measure or integral. The integrability condition can be proven in various ways, one of which is sketched below. In particular, any set that is at most countable has Lebesgue measure zero, and thus a bounded function (on a compact interval) with only finitely or countably many discontinuities is Riemann integrable. An indicator function of a bounded set is Riemann-integrable if and only if the set is Jordan measurable. The Riemann integral can be interpreted measure-theoretically as the integral with respect to the Jordan measure. If a real-valued function is monotone on the interval it is Riemann-integrable, since its set of discontinuities is at most countable, and therefore of Lebesgue measure zero. If a real-valued function on is Riemann-integrable, it is Lebesgue-integrable. That is, Riemann-integrability is a "stronger" (meaning more difficult to satisfy) condition than Lebesgue-integrability. If is a uniformly convergent sequence on with limit , then Riemann integrability of all implies Riemann integrability of , and However, the Lebesgue monotone convergence theorem (on a monotone pointwise limit) does not hold. In Riemann integration, taking limits under the integral sign is far more difficult to logically justify than in Lebesgue integration. It is easy to extend the Riemann integral to functions with values in the Euclidean vector space formula_29 for any . The integral is defined component-wise; in other words, if then In particular, since the complex numbers are a real vector space, this allows the integration of complex valued functions. The Riemann integral is only defined on bounded intervals, and it does not extend well to unbounded intervals. The simplest possible extension is to define such an integral as a limit, in other words, as an improper integral: This definition carries with it some subtleties, such as the fact that it is not always equivalent to compute the Cauchy principal value For example, consider the function which is 0 at , 1 for , and −1 for . By symmetry, always, regardless of . But there are many ways for the interval of integration to expand to fill the real line, and other ways can produce different results; in other words, the multivariate limit does not always exist. We can compute In general, this improper Riemann integral is undefined. Even standardizing a way for the interval to approach the real line does not work because it leads to disturbingly counterintuitive results. If we agree (for instance) that the improper integral should always be then the integral of the translation is −2, so this definition is not invariant under shifts, a highly undesirable property. In fact, not only does this function not have an improper Riemann integral, its Lebesgue integral is also undefined (it equals ). Unfortunately, the improper Riemann integral is not powerful enough. The most severe problem is that there are no widely applicable theorems for commuting improper Riemann integrals with limits of functions. In applications such as Fourier series it is important to be able to approximate the integral of a function using integrals of approximations to the function. For proper Riemann integrals, a standard theorem states that if is a sequence of functions that converge uniformly to on a compact set , then On non-compact intervals such as the real line, this is false. For example, take to be on and zero elsewhere. For all we have: The sequence converges uniformly to the zero function, and clearly the integral of the zero function is zero. Consequently, This demonstrates that for integrals on unbounded intervals, uniform convergence of a function is not strong enough to allow passing a limit through an integral sign. This makes the Riemann integral unworkable in applications (even though the Riemann integral assigns both sides the correct value), because there is no other general criterion for exchanging a limit and a Riemann integral, and without such a criterion it is difficult to approximate integrals by approximating their integrands. A better route is to abandon the Riemann integral for the Lebesgue integral. The definition of the Lebesgue integral is not obviously a generalization of the Riemann integral, but it is not hard to prove that every Riemann-integrable function is Lebesgue-integrable and that the values of the two integrals agree whenever they are both defined. Moreover, a function defined on a bounded interval is Riemann-integrable if and only if it is bounded and the set of points where is discontinuous has Lebesgue measure zero. An integral which is in fact a direct generalization of the Riemann integral is the Henstock–Kurzweil integral. Another way of generalizing the Riemann integral is to replace the factors in the definition of a Riemann sum by something else; roughly speaking, this gives the interval of integration a different notion of length. This is the approach taken by the Riemann–Stieltjes integral. In multivariable calculus, the Riemann integrals for functions from formula_39 are multiple integrals.
https://en.wikipedia.org/wiki?curid=26390
Run-length encoding Run-length encoding (RLE) is a form of lossless data compression in which "runs" of data (sequences in which the same data value occurs in many consecutive data elements) are stored as a single data value and count, rather than as the original run. This is most useful on data that contains many such runs. Consider, for example, simple graphic images such as icons, line drawings, Conway’s Game of Life, and animations. It is not useful with files that don't have many runs as it could greatly increase the file size. RLE may also be used to refer to an early graphics file format supported by CompuServe for compressing black and white images, but was widely supplanted by their later Graphics Interchange Format (GIF). RLE also refers to a little-used image format in Windows 3.x, with the extension codice_1, which is a Run Length Encoded Bitmap, used to compress the Windows 3.x startup screen. For example, consider a screen containing plain black text on a solid white background. There will be many long runs of white pixels in the blank space, and many short runs of black pixels within the text. A hypothetical scan line, with B representing a black pixel and W representing white, might read as follows: With a run-length encoding (RLE) data compression algorithm applied to the above hypothetical scan line, it can be rendered as follows: This can be interpreted as a sequence of twelve Ws, one B, twelve Ws, three Bs, etc., The run-length code represents the original 67 characters in only 18. While the actual format used for the storage of images is generally binary rather than ASCII characters like this, the principle remains the same. Even binary data files can be compressed with this method; file format specifications often dictate repeated bytes in files as padding space. However, newer compression methods such as DEFLATE often use LZ77-based algorithms, a generalization of run-length encoding that can take advantage of runs of strings of characters (such as codice_4). Run-length encoding can be expressed in multiple ways to accommodate data properties as well as additional compression algorithms. For instance, one popular method encodes run lengths for runs of two or more characters only, using an "escape" symbol to identify runs, or using the character itself as the escape, so that any time a character appears twice it denotes a run. On the previous example, this would give the following: This would be interpreted as a run of twelve Ws, a B, a run of twelve Ws, a run of three Bs, etc. In data where runs are less frequent, this can significantly improve the compression rate. One other matter is the application of additional compression algorithms. Even with the runs extracted, the frequencies of different characters may be large, allowing for further compression; however, if the run lengths are written in the file in the locations where the runs occurred, the presence of these numbers interrupts the normal flow and makes it harder to compress. To overcome this, some run-length encoders separate the data and escape symbols from the run lengths, so that the two can be handled independently. For the example data, this would result in two outputs, the string "codice_6" and the numbers (codice_7). Run-length encoding (RLE) schemes were employed in the transmission of analog television signals as far back as 1967. In 1983, run-length encoding was patented by Hitachi. RLE is particularly well suited to palette-based bitmap images such as computer icons, and was a popular image compression method on early online services such as CompuServe before the advent of more sophisticated formats such as GIF. It does not work well on continuous-tone images such as photographs, although JPEG uses it on the coefficients that remain after transforming and quantizing image blocks. Common formats for run-length encoded data include Truevision TGA, PackBits, PCX and ILBM. The International Telecommunication Union also describes a standard to encode run-length-colour for fax machines, known as T.45. The standard, which is combined with other techniques into Modified Huffman coding, is relatively efficient because most faxed documents are generally white space, with occasional interruptions of black.
https://en.wikipedia.org/wiki?curid=26392
Red–black tree In computer science, a red–black tree is a kind of self-balancing binary search tree. Each node stores an extra bit representing color, used to ensure that the tree remains approximately balanced during insertions and deletions. When the tree is modified, the new tree is rearranged and repainted to restore the coloring properties that constrain how unbalanced the tree can become in the worst case. The properties are designed such that this rearranging and recoloring can be performed efficiently. The re-balancing is not perfect, but guarantees searching in time, where "n" is the number of nodes of the tree. The insertion and deletion operations, along with the tree rearrangement and recoloring, are also performed in time. Tracking the color of each node requires only 1 bit of information per node because there are only two colors. The tree does not contain any other data specific to its being a red–black tree so its memory footprint is almost identical to a classic (uncolored) binary search tree. In many cases, the additional bit of information can be stored at no additional memory cost. In 1972, Rudolf Bayer invented a data structure that was a special order-4 case of a B-tree. These trees maintained all paths from root to leaf with the same number of nodes, creating perfectly balanced trees. However, they were not binary search trees. Bayer called them a "symmetric binary B-tree" in his paper and later they became popular as 2-3-4 trees or just 2-4 trees. In a 1978 paper, "A Dichromatic Framework for Balanced Trees", Leonidas J. Guibas and Robert Sedgewick derived the red-black tree from the symmetric binary B-tree. The color "red" was chosen because it was the best-looking color produced by the color laser printer available to the authors while working at Xerox PARC. Another response from Guibas states that it was because of the red and black pens available to them to draw the trees. In 1993, Arne Andersson introduced the idea of right leaning tree to simplify insert and delete operations. In 1999, Chris Okasaki showed how to make the insert operation purely functional. Its balance function needed to take care of only 4 unbalanced cases and one default balanced case. The original algorithm used 8 unbalanced cases, but reduced that to 6 unbalanced cases. Sedgewick showed that the insert operation can be implemented in just 46 lines of Java code. In 2008, Sedgewick proposed the left-leaning red–black tree, leveraging Andersson's idea that simplified algorithms. Sedgewick originally allowed nodes whose two children are red, making his trees more like 2-3-4 trees, but later this restriction was added, making new trees more like 2-3 trees. Sedgewick implemented the insert algorithm in just 33 lines, significantly shortening his original 46 lines of code. A red–black tree is a special type of binary tree, used in computer science to organize pieces of comparable data, such as text fragments or numbers. The leaf nodes of red–black trees do not contain data. These leaves need not be explicit in computer memory—a null child pointer (like NIL in the figure "An example of a red–black tree" below) can encode the fact that this child is a leaf. However, in the description of this figure, the leaves are considered to be explicit nodes—a view which may simplify the description and the understanding of some algorithms for operating on red–black trees. Now, in order to save a marginal amount of execution time (see there), these NIL-leaves may be implemented as sentinel nodes (instead of null pointers). On the other hand, in order to save (main) memory, a single sentinel node (instead of many individuals) may perform the role of all leaf nodes: all references (pointers) from internal nodes to leaf nodes then point to this unique sentinel node. Red–black trees, like all binary search trees, allow efficient in-order traversal (that is: in the order Left–Root–Right) of their elements. The search-time results from the traversal from root to leaf, and therefore a balanced tree of "n" nodes, having the least possible tree height, results in search time. In addition to the requirements imposed on a binary search tree the following must be satisfied by a The only constraint on the children of black nodes is (5). In particular, a black node (like a leaf node) can have a black parent; for example, every perfect binary tree that consists only of black nodes is a red-black tree. The black depth of a node is defined as the number of black nodes from the root to that node (i.e. the number of black ancestors). The black height of a red–black tree is the number of black nodes in any path from the root to the leaves, which, by property 5, is constant (alternatively, it could be defined as the black depth of any leaf node). These constraints enforce a critical property of red–black trees: "the path from the root to the farthest leaf is no more than twice as long as the path from the root to the nearest leaf". The result is that the tree is roughly height-balanced. Since operations such as inserting, deleting, and finding values require worst-case time proportional to the height of the tree, this theoretical upper bound on the height allows red–black trees to be efficient in the worst case, unlike ordinary binary search trees. To see why this is guaranteed, consider a red–black tree having "b" black nodes in property 5. The shortest path from the root to any leaf has "b" black nodes and 0 red nodes. We can insert at most one red node between each to black nodes (property 4). Therefore, the longest possible path (ignoring the NIL leaf nodes) has of "2*b "nodes, alternating black and red. A red–black tree is similar in structure to a B-tree of order 4, where each node can contain between 1 and 3 values and (accordingly) between 2 and 4 child pointers. In such a B-tree, each node will contain only one value matching the value in a black node of the red–black tree, with an optional value before and/or after it in the same node, both matching an equivalent red node of the red–black tree. One way to see this equivalence is to "move up" the red nodes in a graphical representation of the red–black tree, so that they align horizontally with their parent black node, by creating together a horizontal cluster. In the B-tree, or in the modified graphical representation of the red–black tree, all leaf nodes are at the same depth. The red–black tree is then structurally equivalent to a B-tree of order 4, with a minimum fill factor of 33% of values per cluster with a maximum capacity of 3 values. This B-tree type is still more general than a red–black tree though, as it allows ambiguity in a red–black tree conversion—multiple red–black trees can be produced from an equivalent B-tree of order 4. If a B-tree cluster contains only 1 value, it is the minimum, black, and has two child pointers. If a cluster contains 3 values, then the central value will be black and each value stored on its sides will be red. If the cluster contains two values, however, either one can become the black node in the red–black tree (and the other one will be red). So the order-4 B-tree does not maintain which of the values contained in each cluster is the root black tree for the whole cluster and the parent of the other values in the same cluster. Despite this, the operations on red–black trees are more economical in time because you don't have to maintain the vector of values. It may be costly if values are stored directly in each node rather than being stored by reference. B-tree nodes, however, are more economical in space because you don't need to store the color attribute for each node. Instead, you have to know which slot in the cluster vector is used. If values are stored by reference, e.g. objects, null references can be used and so the cluster can be represented by a vector containing 3 slots for value pointers plus 4 slots for child references in the tree. In that case, the B-tree can be more compact in memory, improving data locality. The same analogy can be made with B-trees with larger orders that can be structurally equivalent to a colored binary tree: you just need more colors. Suppose that you add blue, then the blue–red–black tree defined like red–black trees but with the additional constraint that no two successive nodes in the hierarchy will be blue and all blue nodes will be children of a red node, then it becomes equivalent to a B-tree whose clusters will have at most 7 values in the following colors: blue, red, blue, black, blue, red, blue (For each cluster, there will be at most 1 black node, 2 red nodes, and 4 blue nodes). For moderate volumes of values, insertions and deletions in a colored binary tree are faster compared to B-trees because colored trees don't attempt to maximize the fill factor of each horizontal cluster of nodes (only the minimum fill factor is guaranteed in colored binary trees, limiting the number of splits or junctions of clusters). B-trees will be faster for performing rotations (because rotations will frequently occur within the same cluster rather than with multiple separate nodes in a colored binary tree). For storing large volumes, however, B-trees will be much faster as they will be more compact by grouping several children in the same cluster where they can be accessed locally. All optimizations possible in B-trees to increase the average fill factors of clusters are possible in the equivalent multicolored binary tree. Notably, maximizing the average fill factor in a structurally equivalent B-tree is the same as reducing the total height of the multicolored tree, by increasing the number of non-black nodes. The worst case occurs when all nodes in a colored binary tree are black, the best case occurs when only a third of them are black (and the other two thirds are red nodes). Red–black trees offer worst-case guarantees for insertion time, deletion time, and search time. Not only does this make them valuable in time-sensitive applications such as real-time applications, but it makes them valuable building blocks in other data structures which provide worst-case guarantees; for example, many data structures used in computational geometry can be based on red–black trees, and the Completely Fair Scheduler used in current Linux kernels and epoll system call implementation uses red–black trees. The AVL tree is another structure supporting search, insertion, and removal. AVL trees can be colored red-black, thus are a subset of RB trees. Worst-case height is 0.720 times the worst-case height of RB trees, so AVL trees are more rigidly balanced. The performance measurements of Ben Pfaff with realistic test cases in 79 runs find AVL to RB ratios between 0.677 and 1.077, median at 0.947, and geometric mean 0.910. WAVL trees have a performance in between those two. Red–black trees are also particularly valuable in functional programming, where they are one of the most common persistent data structures, used to construct associative arrays and sets which can retain previous versions after mutations. The persistent version of red–black trees requires space for each insertion or deletion, in addition to time. For every 2-4 tree, there are corresponding red–black trees with data elements in the same order. The insertion and deletion operations on 2-4 trees are also equivalent to color-flipping and rotations in red–black trees. This makes 2-4 trees an important tool for understanding the logic behind red–black trees, and this is why many introductory algorithm texts introduce 2-4 trees just before red–black trees, even though 2-4 trees are not often used in practice. In 2008, Sedgewick introduced a simpler version of the red–black tree called the left-leaning red–black tree by eliminating a previously unspecified degree of freedom in the implementation. The LLRB maintains an additional invariant that all red links must lean left except during inserts and deletes. Red–black trees can be made isometric to either 2-3 trees, or 2-4 trees, for any sequence of operations. The 2-4 tree isometry was described in 1978 by Sedgewick. With 2-4 trees, the isometry is resolved by a "color flip," corresponding to a split, in which the red color of two children nodes leaves the children and moves to the parent node. The original description of the tango tree, a type of tree optimized for fast searches, specifically uses red–black trees as part of its data structure. In the version 8 of Java, the HashMap has been modified such that instead of using a LinkedList to store different elements with colliding hashcodes, a red-black tree is used. This results in the improvement of time complexity of searching such an element from to . Read-only operations on a red–black tree require no modification from those used for binary search trees, because every red–black tree is a special case of a simple binary search tree. However, the immediate result of an insertion or removal may violate the properties of a red–black tree. Restoring the red–black properties requires a small number ( or amortized) of color changes (which are very quick in practice) and no more than three tree rotations (two for insertion). Although insert and delete operations are complicated, their times remain . If the example implementation below is not suitable, other implementations with explanations may be found in Ben Pfaff's annotated C library GNU libavl (v2.0.3 as of June 2019). The details of the insert and removal operations will be demonstrated with example C++ code. The example code may call upon the helper functions below to find the parent, sibling, uncle and grandparent nodes and to rotate a node left or right: // Basic type definitions: enum color_t { BLACK, RED }; struct Node { // Helper functions: Node* GetParent(Node* n) { Node* GetGrandParent(Node* n) { Node* GetSibling(Node* n) { Node* GetUncle(Node* n) { void RotateLeft(Node* n) { void RotateRight(Node* n) { Insertion begins by adding the node in a very similar manner as a standard binary search tree insertion and by coloring it red. The big difference is that in the binary search tree a new node is added as a leaf, whereas leaves contain no information in the red–black tree, so instead the new node replaces an existing leaf and then has two black leaves of its own added. Node* Insert(Node* root, Node* n) { void InsertRecurse(Node* root, Node* n) { What happens next depends on the color of other nearby nodes. There are several cases of red–black tree insertion to handle: void InsertRepairTree(Node* n) { Note that: Case 1: The current node N is at the root of the tree. In this case, it is repainted black to satisfy property 2 (the root is black). Since this adds one black node to every path at once, property 5 (all paths from any given node to its leaf nodes contain the same number of black nodes) is not violated. void InsertCase1(Node* n) { Case 2: The current node's parent P is black, so property 4 (both children of every red node are black) holds. Property 5 (all paths from any given node to its leaf nodes contain the same number of black nodes) is not threatened, because the new node N has two black leaf children, but because N is red, the paths through each of its children have the same number of black nodes as the path through the leaf it replaced, which was black, and so this property remains satisfied. So the tree remains valid. void InsertCase2(Node* n) { void InsertCase3(Node* n) { void InsertCase4(Node* n) { void InsertCase4Step2(Node* n) { Note that inserting is actually in-place, since all the calls above use tail recursion. In the algorithm above, all cases are called only once, except in Case 3 where it can recurse back to Case 1 with the grandparent node, which is the only case where an iterative implementation will effectively loop. Because the problem of repair in that case is escalated two levels higher each time, it takes maximally iterations to repair the tree (where is the height of the tree). Because the probability for escalation decreases exponentially with each iteration the average insertion cost is practically constant. In a regular binary search tree when deleting a node with two non-leaf children, we find either the maximum element in its left subtree (which is the in-order predecessor) or the minimum element in its right subtree (which is the in-order successor) and move its value into the node being deleted (as shown here). We then delete the node we copied the value from, which must have fewer than two non-leaf children. (Non-leaf children, rather than all children, are specified here because unlike normal binary search trees, red–black trees can have leaf nodes anywhere, which are actually the sentinel Nil, so that all nodes are either internal nodes with two children or leaf nodes with, by definition, zero children. In effect, internal nodes having two leaf children in a red–black tree are like the leaf nodes in a regular binary search tree.) Because merely copying a value does not violate any red–black properties, this reduces to the problem of deleting a node with at most one non-leaf child. Once we have solved that problem, the solution applies equally to the case where the node we originally want to delete has at most one non-leaf child as to the case just considered where it has two non-leaf children. Therefore, for the remainder of this discussion we address the deletion of a node with at most one non-leaf child. We use the label M to denote the node to be deleted; C will denote a selected child of M, which we will also call "its child". If M does have a non-leaf child, call that its child, C; otherwise, choose either leaf as its child, C. If M is a red node, we simply replace it with its child C, which must be black by property 4. (This can only occur when M has two leaf children, because if the red node M had a black non-leaf child on one side but just a leaf child on the other side, then the count of black nodes on both sides would be different, thus the tree would violate property 5.) All paths through the deleted node will simply pass through one fewer red node, and both the deleted node's parent and child must be black, so property 3 (all leaves are black) and property 4 (both children of every red node are black) still hold. Another simple case is when M is black and C is red. Simply removing a black node could break Properties 4 (“Both children of every red node are black”) and 5 (“All paths from any given node to its leaf nodes contain the same number of black nodes”), but if we repaint C black, both of these properties are preserved. The complex case is when both M and C are black. (This can only occur when deleting a black node which has two leaf children, because if the black node M had a black non-leaf child on one side but just a leaf child on the other side, then the count of black nodes on both sides would be different, thus the tree would have been an invalid red–black tree by violation of property 5.) We begin by replacing M with its child C – we recall that in this case "its child C" is either child of M, both being leaves. We will "relabel" this child C (in its new position) N, and its sibling (its new parent's other child) S. (S was previously the sibling of M.) In the diagrams below, we will also use P for N's new parent (M's old parent), SL for S's left child, and SR for S's right child (S cannot be a leaf because if M and C were black, then P's one subtree which included M counted two black-height and thus P's other subtree which includes S must also count two black-height, which cannot be the case if S is a leaf node). We can perform the steps outlined above with the following code, where the function codice_1 substitutes codice_2 into codice_3's place in the tree. For convenience, code in this section will assume that null leaves are represented by actual node objects rather than NULL (the code in the "Insertion" section works with either representation). void ReplaceNode(Node* n, Node* child) { void DeleteOneChild(Node* n) { If both N and its original parent are black, then deleting this original parent causes paths which proceed through N to have one fewer black node than paths that do not. As this violates property 5 (all paths from any given node to its leaf nodes contain the same number of black nodes), the tree must be rebalanced. There are several cases to consider: Case 1: N is the new root. In this case, we are done. We removed one black node from every path, and the new root is black, so the properties are preserved. void DeleteCase1(Node* n) { void DeleteCase2(Node* n) { void DeleteCase3(Node* n) { void DeleteCase4(Node* n) { void DeleteCase5(Node* n) { void DeleteCase6(Node* n) { Again, the function calls all use tail recursion, so the algorithm is in-place. In the algorithm above, all cases are chained in order, except in delete case 3 where it can recurse to case 1 back to the parent node: this is the only case where an iterative implementation will effectively loop. No more than loops back to case 1 will occur (where is the height of the tree). And because the probability for escalation decreases exponentially with each iteration the average removal cost is constant. Additionally, no tail recursion ever occurs on a child node, so the tail recursion loop can only move from a child back to its successive ancestors. If a rotation occurs in case 2 (which is the only possibility of rotation within the loop of cases 1–3), then the parent of the node N becomes red after the rotation and we will exit the loop. Therefore, at most one rotation will occur within this loop. Since no more than two additional rotations will occur after exiting the loop, at most three rotations occur in total. A red black tree which contains "n" internal nodes has a height of . Definitions: Lemma: A subtree rooted at node "v" has at least formula_1 internal nodes. Proof of Lemma (by induction height): Basis: h("v") = 0 If "v" has a height of zero then it must be "null", therefore bh("v") = 0. So: Inductive Step: "v" such that h("v") = k, has at least formula_1 internal nodes implies that formula_4 such that h(formula_4) = k+1 has at least formula_6 internal nodes. Since formula_4 has h(formula_4) > 0 it is an internal node. As such it has two children each of which have a black-height of either bh(formula_4) or bh(formula_4)-1 (depending on whether the child is red or black, respectively). By the inductive hypothesis each child has at least formula_11 internal nodes, so formula_4 has at least: internal nodes. Using this lemma we can now show that the height of the tree is logarithmic. Since at least half of the nodes on any path from the root to a leaf are black (property 4 of a red–black tree), the black-height of the root is at least h(root)/2. By the lemma we get: Therefore, the height of the root is . In addition to the single-element insert, delete and lookup operations, several set operations have been defined on red-black trees: union, intersection and set difference. Then fast "bulk" operations on insertions or deletions can be implemented based on these set functions. These set operations rely on two helper operations, "Split" and "Join". With the new operations, the implementation of red-black trees can be more efficient and highly-parallelizable. This implementation allows a red root. The join algorithm is as follows: Here formula_16 of a node formula_17 means twice the black height of a black node, and the twice the black height of a red node. expose(v)=(l,⟨k,c⟩,r) means to extract a tree node formula_17's left child formula_19, the key of the node formula_20, the color of the node formula_21 and the right child formula_22. Node(l,⟨k,c⟩,r) means to create a node of left child formula_19, key formula_20, color formula_21 and right child formula_22. The split algorithm is as follows: The union of two red-black trees and representing sets and , is a red-black tree that represents . The following recursive function computes this union: Here, "Split" is presumed to return two trees: one holding the keys less its input key, one holding the greater keys. (The algorithm is non-destructive, but an in-place destructive version exists as well.) The algorithm for intersection or difference is similar, but requires the "Join2" helper routine that is the same as "Join" but without the middle key. Based on the new functions for union, intersection or difference, either one key or multiple keys can be inserted to or deleted from the red-black tree. Since "Split" calls "Join" but does not deal with the balancing criteria of red-black trees directly, such an implementation is usually called the "join-based" implementation. The complexity of each of union, intersection and difference is formula_27 for two red-black trees of sizes formula_28 and formula_29. This complexity is optimal in terms of the number of comparisons. More importantly, since the recursive calls to union, intersection or difference are independent of each other, they can be executed in parallel with a parallel depth formula_30. When formula_31, the join-based implementation has the same computational directed acyclic graph (DAG) as single-element insertion and deletion if the root of the larger tree is used to split the smaller tree. Parallel algorithms for constructing red–black trees from sorted lists of items can run in constant time or time, depending on the computer model, if the number of processors available is asymptotically proportional to the number of items where . Fast search, insertion, and deletion parallel algorithms are also known. The join-based algorithms for red-black trees are parallel for bulk operations, including union, intersection, construction, filter, map-reduce, and so on. Basic operations like insertion, removal or update can be parallelized by defining operations that process bulks of multiple elements. It is also possible to process bulks with several basic operations, for example bulks may contain elements to insert and also elements to remove from the tree. The algorithms for bulk operations aren't just applicable to the red-black tree, but can be adapted to other sorted sequence data structures as well, like the 2-3 tree, 2-3-4 tree and (a,b)-tree. In the following different algorithms for bulk insert will be explained, but the same algorithms can also be applied to removal and update. Bulk insert is an operation that inserts each element of a sequence formula_32 into a tree formula_33. This approach can be applied to every sorted sequence data structure that supports efficient join- and split-operations. The general idea is to split formula_32 and formula_33 in multiple parts and perform the insertions on these parts in parallel. Note that in Step 3 the constraints for splitting formula_32 assure that in Step 5 the trees can be joined again and the resulting sequence is sorted. The pseudo code shows a simple divide-and-conquer implementation of the join-based algorithm for bulk-insert. Both recursive calls can be executed in parallel. The join operation used here differs from the version explained in this article, instead join2 is used which misses the second parameter k. Sorting formula_32 is not considered in this analysis. This can be improved by using parallel algorithms for splitting and joining. In this case the execution time is formula_52. Another method of parallelizing bulk operations is to use a pipelining approach. This can be done by breaking the task of processing a basic operation up into a sequence of subtasks. For multiple basic operations the subtasks can be processed in parallel by assigning each subtask to a separate processor. Sorting formula_32 is not considered in this analysis. Also, formula_83 is assumed to be smaller than formula_84, otherwise it would be more sufficient to construct the resulting tree from scratch. A red-black-tree was referenced correctly in an episode of Missing (Canadian TV series) as noted by Robert Sedgewick in one of his lectures:
https://en.wikipedia.org/wiki?curid=26397
Retroactive continuity Retroactive continuity, or retcon for short, is a literary device in which established diegetic 'facts' in the plot of a fictional work are adjusted, ignored, or contradicted by a subsequently published work which breaks continuity with the former. There are various motivations for applying retroactive continuity, including: Retcons are used by authors to increase their creative freedom, on the assumption that the changes are unimportant to the audience compared to the new story which can be told. For instance, by retroactively setting a prior story in a parallel universe, departed popular characters can be reintroduced. More subtly, a minor plot point might be retroactively expunged (for instance, the heroine leaving home without any food), removing an obstacle to further storytelling (that she should be getting hungry). Retcons are common in pulp fiction, and especially in comic books published by long-established publishers such as DC and Marvel. The long history of popular titles and the number of writers who contribute stories can often create situations that demand clarification or revision. Retcons also often appear in manga, soap operas, serial dramas, movie sequels, cartoons, professional wrestling angles, video games, radio series, and other forms of serial fiction. The first published use of the phrase "retroactive continuity" is found in theologian E. Frank Tupper's 1973 book "The Theology of Wolfhart Pannenberg": "Pannenberg's conception of retroactive continuity ultimately means that history flows fundamentally from the future into the past, that the future is not basically a product of the past." The first known printed use of "retroactive continuity" referring to the altering of history in a fictional work is in "All-Star Squadron" #18 (February 1983) from DC Comics. The series was set on DC's Earth-Two, an alternate universe in which Golden Age comic characters age in real time. "All-Star Squadron" was set during World War II on Earth-Two; as it was in the past of an alternate universe, all its events had repercussions on the contemporary continuity of the DC multiverse. Each issue changed the history of the fictional world in which it was set. In the letters column, a reader remarked that the comic "must make you [the creators] feel at times as if you're painting yourself into a corner", and, "Your matching of Golden Age comics history with new plotlines has been an artistic (and I hope financial!) success." Writer Roy Thomas responded, "we like to think that an enthusiastic ALL-STAR booster at one of Adam Malin's Creation Conventions in San Diego came up with the best name for it a few months back: 'Retroactive Continuity'. Has kind of a ring to it, don't you think?" The term then took firm root in the consciousness of fans of American superhero comics. At some point, "retroactive continuity" was shortened to "retcon", reportedly by Damian Cugley in 1988 on Usenet. Hard evidence of Cugley's abbreviation has yet to surface, though in a Usenet posting on August 18, 1990, Cugley posted a reply in which he identified himself as "the originator of the word "retcon"". Cugley used the neologism to describe a development in the comic book "Saga of the Swamp Thing", which reinterprets the events of the title character's origin by revealing facts that previously were not part of the narrative and were not intended by earlier writers. Retcons sometimes do not contradict previously established facts but instead fill in missing background details, usually to support current plot points. Thomas referred to "retroactive continuity" in this sense, as a purely additive process that did not undo any previous work; such additions were common in "All-Star Squadron". Kurt Busiek took a similar approach with "Untold Tales of Spider-Man", a series which told stories that specifically fit between issues of the original "The Amazing Spider-Man" series, sometimes explaining discontinuities between those earlier stories. John Byrne used a similar structure with "". Possibly the earliest Marvel Comics example of new stories placed between long-established stories was the 1977-8 magazine "The Rampaging Hulk". In "The Godfather Part II", the character Frank Pentangeli is introduced as an old friend of the family though he is not referenced in the first movie; similarly Don Altobello is one of the "old time" Dons, though he is not mentioned until "The Godfather Part III". Neither addition affects the plot line of the previous films. The addition, in later seasons, of an attic to the family's home in "Full House" stands as a similar additive example. A similar concept is that of secret history, in which the events of a story occur within the bounds of already-established events (especially real-world ones), revealing different interpretations of the events. Some of Tim Powers' novels use secret history, such as "Last Call", which suggests that Bugsy Siegel's actions were due to his being a modern-day Fisher King. Alan Moore's additional information about the Swamp Thing's origins – revealing that Swamp Thing was not actually scientist Alec Holland converted into a plant, but actually a plant that had absorbed Holland's body and consciousness so that it merely thought it was Holland – did not contradict or change any of the events depicted in the character's previous appearances, but instead changed the reader's interpretation of them. Such additions and reinterpretations are very common in "Doctor Who". In the "Star Wars" franchise, one of the most cited plot holes is that the Galactic Empire's superweapon, the Death Star, has a glaring, poorly defended weak point: an exhaust port. However, the movie "Rogue One: A Star Wars Story", which acts as a direct prequel to "A New Hope", presents the design flaw as deliberate: Galen Erso, the head engineer of the Death Star project, designed the reactors to be unstable, thus needing an exhaust port, and gave the plans to the Rebel Alliance in order to sabotage the Empire and to secretly show the Rebels how to destroy the Death Star. Retcons sometimes add information that seemingly contradicts previous information. This frequently takes the form of a character who was shown to have died but is later revealed to have somehow survived. This is a common practice in horror films, which may end with the death of a monster that goes on to appear in one or more sequels. The technique is so common in superhero comics that the term "comic book death" has been coined for it. An early example of this type of retcon is the return of Sherlock Holmes, whom writer Arthur Conan Doyle apparently killed off in "The Final Problem" in 1893, only to bring him back, in large part because of readers' responses, with "The Empty House" in 1903. Another example is the character Phil Coulson, who was killed off in the Marvel Studios film "The Avengers" (2012), but is later shown to have survived in the television series "Agents of S.H.I.E.L.D". In many of his detective novels, Rex Stout implies that his character Nero Wolfe was born in Montenegro, giving some details of his early life in the Balkans around World War I. But in 1939's "Over My Dead Body", Wolfe tells an FBI agent that he was born in the United States. Stout revealed the reason for the change in a letter obtained by his authorized biographer, John McAleer: "In the original draft of "Over My Dead Body" Nero was a Montenegrin by birth, and it all fitted previous hints as to his background; but violent protests from "The American Magazine", supported by Farrar & Rinehart, caused his cradle to be transported five thousand miles." In the 1940s and 1950s, Isaac Asimov placed the planet Trantor, capital of the Galactic Empire, at the "center of the galaxy", but later astronomical research indicated that the actual Galactic Center might be occupied by a supermassive black hole, making human life there impossible; in later works, Asimov adjusted his galaxy and Trantor's location in it. When E.E. "Doc" Smith wrote the original "The Skylark of Space", space flight was a completely theoretical proposition. However, the last book of the series, "Skylark DuQuesne", was written in 1963, when the United States and the Soviet Union were involved in the space race. Smith adjusted the past of his series accordingly, mentioning an American base and a Soviet one being established on the Moon prior to the protagonist Seaton discovering faster-than-light flight. Alan Moore's retcons often involve false memories. He has used this technique in the "Marvelman" series, "Swamp Thing" and "". Retconning can bring back characters who were initially killed off. An example of this occurs on the CBS comedy "Two and a Half Men". The character Charlie Harper (Charlie Sheen) was killed in a train accident and briefly returns as a ghost (played by Kathy Bates) in the ninth season. Despite numerous instances that confirm his demise including the fact that a mutilated body was in his coffin, the season twelve series finale "Of Course He's Dead", it is revealed that Charlie survived the ordeal and has been held captive against his will for over four years. It is mentioned that there was no body only random body parts. It is also common in soap operas. On "The Bold and the Beautiful", Taylor Forrester (Hunter Tylo) was shown to flatline and have a funeral. When Tylo reprised the character in 2005, a retcon explained that Taylor had actually gone into a coma. The TV series "Dallas" annulled its entire Season 9 as being just the dream of another character, Pam Ewing. Writers did this to offer a supposedly plausible reason for the major character of Bobby Ewing, who had died onscreen at the end of Season 8, to be still alive when actor Patrick Duffy wanted to return to the series. This season is sometimes referred to as the "Dream Season" and was referred to humorously in later TV series such as "Family Guy". In "", the character of Spock dies in the film. This was done because actor Leonard Nimoy would not agree to appear in the film unless the character was killed off. After the film was finished, Nimoy told producers he enjoyed making the movie and wanted to return if a third "Star Trek" film was made. As a result, an epilogue was added to the film, hinting at the possibility of Spock's resurrection, described as a result of the Genesis Device, a technology designed to reorganize dead matter into habitable worlds. Also in "Star Trek II", the characters of Khan (Ricardo Montalbán) and Pavel Chekov (Walter Koenig) remember each other from Khan's appearance in the first season "" television series episode "Space Seed", 15 years earlier; The character of Chekov did not appear in "Star Trek" until the second season; It was later explained that Chekov was not a member of the "Enterprise" bridge crew at that time and was assigned elsewhere on the ship. In "Beverly Hills 90210", Kelly Taylor starts out in the first season initially not knowing Dylan McKay. Later when Brenda Walsh is sent to Paris with Donna Martin, Kelly Taylor has suddenly known him since kindergarten. This apparent retooling of the storyline contradicted years of prior character development and was apparently done to prepare viewers for Shannen Doherty's exit from the series. Marvel Comics' Beyonder was originally stated to be omnipotent and the most powerful being in Marvel Universe, coming from the Beyond Realm. However, after his creator, Jim Shooter, left Marvel, writer-editor Tom DeFalco re-tooled the Beyonder, diminishing his power greatly: he was no longer omnipotent, as certain other cosmic entities were retroactively vastly upgraded to transcend the scale of infinity on which the character worked. Even after this, Beyonder was still one of the most powerful beings in Marvel, with several characters exceeding him. In 2003, in the title of DC Comics' Teen Titans, Geoff Johns changed the entire genetic code of Kon-El (the modern version of Superboy) from a genetically altered human clone that was designed to be as Kryptonian as possible into a hybrid clone of both Superman and Lex Luthor. This change contradicted years of continuity and various facts that proved that Kon-El was human and in the process mostly ignored his unique ability of tactile telekinesis that made his powers very different from those of Superman. In "Boy Meets World", both Shawn Hunter and Topanga Lawrence have siblings in the first season but later in the series are retconned to be only children (though Shawn's half-brother Jack is introduced in later seasons). The ages of the characters of "Boy Meets World" are altered notably where Cory is age 11 in 6th grade during season 1 to age 13 and 7th grade in season 2. This happens again in high school skipping another grade. The age gap between Cory and Eric also narrows from 4 years apart in age to 2 years apart in age. Another prominent retcon during that series was that of the character of Morgan Matthews, who was portrayed by two actresses during the show's run; in the series finale of the sequel series "Girl Meets World", the continuity suggests that they were two separate people who both appeared together for the only time during that episode. Unpopular or embarrassing stories are sometimes later ignored by publishers, and effectively erased from a series's continuity. Later stories may contradict the previous ones or explicitly establish that they never happened. A notable example of subtractive retconning is the "X-Men" film series. The 2014 film "" features the character Wolverine traveling in time to 1973 to prevent an assassination that, if carried out, would lead to the creation of a new weapons system called the Sentinels that threatens the existence of mutants — and potentially, all of humanity. The events of this film retconned those of two previous films in the series: The film, a loose adaptation of "The Dark Phoenix Saga" from the "Uncanny X-Men" comic book series, featured the unceremonious deaths of select key characters (including Cyclops), triggering a mixed fan response from those who felt the deaths were sudden and under-explored. "Days of Future Past" erased these deaths and the characters are shown to have survived at the end of the film. "The Dark Phoenix Saga" would subsequently be adapted a second time in "Dark Phoenix" (2019). The film featured the character Deadpool, who is created by the character Colonel William Stryker as an amalgamation of several other mutants. He is referred to as "the Deadpool" due to having the other mutants' abilities "pooled" together. This version of the character was poorly received by fans due to his appearance and origin, both of which were considered unfaithful to the comics version of the character. The events of "Days of Future Past" allowed a more faithful version of the character to be adapted in "Deadpool" (2016) and its 2018 sequel, both of which were positively received by fans and critics. An unpopular retcon may itself be retconned away, as happened with John Byrne's "". Film sequels often employ temporal compression to maintain a sense of urgency in each installment, despite more time having elapsed diegetically ("in-universe") between one installment and another than has elapsed in the real world. For example, despite the implied contemporaneity in each of the films of "The Omen" series, the lead character ages some 15 or 20 years across three films released over a period of less than 6 years. Retroactive continuity is similar to, but not the same as, plot inconsistencies introduced accidentally or through lack of concern for continuity; retconning, by comparison, is done deliberately. For example, the ongoing continuity contradictions on episodic TV series such as "The Simpsons" (in which the timeline of the family's history must be continually shifted forward to explain they are not getting any older) reflects intentionally lost continuity, not genuine retcons. However, in series with generally tight continuity, retcons are sometimes created after the fact to explain continuity errors, such was the case in "The Flintstones", where Wilma Flintstone was mistakenly given two separate maiden names, "Pebble" and "Slaghoople", over the course of the series. Upon discovering the discrepancy, the producers settled on "Slaghoople" and retconned it into later series in the franchise. Retconning is also generally distinct from replacing the actor who plays a part in an ongoing series, which is more commonly an example of "loose continuity" rather than retroactively changing past continuity. The different appearance of the character is either ignored (as was done with the characters of Darrin Stephens and Gladys Kravitz on the television series "Bewitched") or explained within the series, such as with "regeneration" in "Doctor Who", or the Oracle in "The Matrix Revolutions". Sometimes, there are referential, inside jokes on actor changes in the show, such as with "My Wife & Kids" and "Roseanne", where there was a change of actresses playing a role (characters Claire Kyle and Becky Conner, respectively). In the latter, another character observes that children can change as they reach adulthood, remarking that when Becky came back from college (played by a new actress), they could not even recognize her. When the actor playing Rory was replaced in "Mrs Brown's Boys", the new actor first appeared on set bandaged up, supposedly following cosmetic surgery on his face. When the bandages were removed, the other characters claimed not to notice any difference. A similar set up gag was used with the character of Herr Flick in the BBC sitcom "Allo Allo". Retconning also differs from direct revision. For example, when George Lucas re-edited the original "Star Wars" trilogy, he made changes directly to the source material, rather than introducing new source material that contradicted the contents of previous material. Retconning is not the same as a reboot or reimagining which completely discards the original timeline, such as in "Battlestar Galactica". However, there have been partial reboots of franchises where the core of the franchise is still canonical but the expanded universe is relegated to a secondary continuity which, while not completely invalid, is subject to revision and critical review. "Robotech" is an example of this. With the release of the 2006 sequel film "Robotech The Shadow Chronicles", Harmony Gold established that only the original 1985 animated series and the 2006 sequel film are considered canonical relegating the aborted "", comics, and novels from the 80s and 90s to secondary continuity and, if elements are used from them, they are subject to selective revision and updating as appropriate to mesh with future canonical productions and to prevent conflict with the original animated series. While the Jack McKinney "Robotech" novel "End of the Circle" is evidently no longer canon, the prequel comic "" establishes that the general storyline of The Sentinels still occurred in some fashion, but various elements, including the timeline, specific unfolding of events, and some characterizations are different from what was previously depicted in earlier comics and novels. In such cases, the franchise producer may state that there is no intention to address the changes through remakes or direct retellings of such stories. It is essentially left to the viewer's imagination as to how differently the revised story unfolded.
https://en.wikipedia.org/wiki?curid=26400
Risk management Risk management is the identification, evaluation, and prioritization of risks (defined in ISO 31000 as "the effect of uncertainty on objectives") followed by coordinated and economical application of resources to minimize, monitor, and control the probability or impact of unfortunate events or to maximize the realization of opportunities. Risks can come from various sources including uncertainty in financial markets, threats from project failures (at any phase in design, development, production, or sustaining of life-cycles), legal liabilities, credit risk, accidents, natural causes and disasters, deliberate attack from an adversary, or events of uncertain or unpredictable root-cause. There are two types of events i.e. negative events can be classified as risks while positive events are classified as opportunities. Risk management standards have been developed by various institutions, including the Project Management Institute, the National Institute of Standards and Technology, actuarial societies, and ISO standards. Methods, definitions and goals vary widely according to whether the risk management method is in the context of project management, security, engineering, industrial processes, financial portfolios, actuarial assessments, or public health and safety. Strategies to manage threats (uncertainties with negative consequences) typically include avoiding the threat, reducing the negative effect or probability of the threat, transferring all or part of the threat to another party, and even retaining some or all of the potential or actual consequences of a particular threat. The opposite of these strategies can be used to respond to opportunities (uncertain future states with benefits). Certain risk management standards have been criticized for having no measurable improvement on risk, whereas the confidence in estimates and decisions seems to increase. For example, one study found that one in six IT projects were "black swans" with gigantic overruns (cost overruns averaged 200%, and schedule overruns 70%). A widely used vocabulary for risk management is defined by "ISO Guide 73:2009", "Risk management. Vocabulary." In ideal risk management, a prioritization process is followed whereby the risks with the greatest loss (or impact) and the greatest probability of occurring are handled first. Risks with lower probability of occurrence and lower loss are handled in descending order. In practice the process of assessing overall risk can be difficult, and balancing resources used to mitigate between risks with a high probability of occurrence but lower loss, versus a risk with high loss but lower probability of occurrence can often be mishandled. Intangible risk management identifies a new type of a risk that has a 100% probability of occurring but is ignored by the organization due to a lack of identification ability. For example, when deficient knowledge is applied to a situation, a knowledge risk materializes. Relationship risk appears when ineffective collaboration occurs. Process-engagement risk may be an issue when ineffective operational procedures are applied. These risks directly reduce the productivity of knowledge workers, decrease cost-effectiveness, profitability, service, quality, reputation, brand value, and earnings quality. Intangible risk management allows risk management to create immediate value from the identification and reduction of risks that reduce productivity. Opportunity cost represents a unique challenge for risk managers. It can be difficult to determine when to put resources toward risk management and when to use those resources elsewhere. Again, ideal risk management minimizes spending (or manpower or other resources) and also minimizes the negative effects of risks. Risk is defined as the possibility that an event will occur that adversely affects the achievement of an objective. Uncertainty, therefore, is a key aspect of risk. Systems like the Committee of Sponsoring Organizations of the Treadway Commission Enterprise Risk Management (COSO ERM), can assist managers in mitigating risk factors. Each company may have different internal control components, which leads to different outcomes. For example, the framework for ERM components includes Internal Environment, Objective Setting, Event Identification, Risk Assessment, Risk Response, Control Activities, Information and Communication, and Monitoring. For the most part, these methods consist of the following elements, performed, more or less, in the following order. The International Organization for Standardization (ISO) identifies the following principles of risk management: Risk management should: According to the standard ISO 31000 "Risk management – Principles and guidelines on implementation," the process of risk management consists of several steps as follows: This involves: After establishing the context, the next step in the process of managing risk is to identify potential risks. Risks are about events that, when triggered, cause problems or benefits. Hence, risk identification can start with the source of our problems and those of our competitors (benefit), or with the problem consequenses. Examples of risk sources are: stakeholders of a project, employees of a company or the weather over an airport. When either source or problem is known, the events that a source may trigger or the events that can lead to a problem can be investigated. For example: stakeholders withdrawing during a project may endanger funding of the project; confidential information may be stolen by employees even within a closed network; lightning striking an aircraft during takeoff may make all people on board immediate casualties. The chosen method of identifying risks may depend on culture, industry practice and compliance. The identification methods are formed by templates or the development of templates for identifying source, problem or event. Common risk identification methods are: Once risks have been identified, they must then be assessed as to their potential severity of impact (generally a negative impact, such as damage or loss) and to the probability of occurrence. These quantities can be either simple to measure, in the case of the value of a lost building, or impossible to know for sure in the case of an unlikely event, the probability of occurrence of which is unknown. Therefore, in the assessment process it is critical to make the best educated decisions in order to properly prioritize the implementation of the risk management plan. Even a short-term positive improvement can have long-term negative impacts. Take the "turnpike" example. A highway is widened to allow more traffic. More traffic capacity leads to greater development in the areas surrounding the improved traffic capacity. Over time, traffic thereby increases to fill available capacity. Turnpikes thereby need to be expanded in a seemingly endless cycles. There are many other engineering examples where expanded capacity (to do any function) is soon filled by increased demand. Since expansion comes at a cost, the resulting growth could become unsustainable without forecasting and management. The fundamental difficulty in risk assessment is determining the rate of occurrence since statistical information is not available on all kinds of past incidents and is particularly scanty in the case of catastrophic events, simply because of their infrequency. Furthermore, evaluating the severity of the consequences (impact) is often quite difficult for intangible assets. Asset valuation is another question that needs to be addressed. Thus, best educated opinions and available statistics are the primary sources of information. Nevertheless, risk assessment should produce such information for senior executives of the organization that the primary risks are easy to understand and that the risk management decisions may be prioritized within overall company goals. Thus, there have been several theories and attempts to quantify risks. Numerous different risk formulae exist, but perhaps the most widely accepted formula for risk quantification is: "Rate (or probability) of occurrence multiplied by the impact of the event equals risk magnitude." Risk mitigation measures are usually formulated according to one or more of the following major risk options, which are: Later research has shown that the financial benefits of risk management are less dependent on the formula used but are more dependent on the frequency and how risk assessment is performed. In business it is imperative to be able to present the findings of risk assessments in financial, market, or schedule terms. Robert Courtney Jr. (IBM, 1970) proposed a formula for presenting risks in financial terms. The Courtney formula was accepted as the official risk analysis method for the US governmental agencies. The formula proposes calculation of ALE (annualized loss expectancy) and compares the expected loss value to the security control implementation costs (cost-benefit analysis). Once risks have been identified and assessed, all techniques to manage the risk fall into one or more of these four major categories: Ideal use of these risk control strategies may not be possible. Some of them may involve trade-offs that are not acceptable to the organization or person making the risk management decisions. Another source, from the US Department of Defense (see link), Defense Acquisition University, calls these categories ACAT, for Avoid, Control, Accept, or Transfer. This use of the ACAT acronym is reminiscent of another ACAT (for Acquisition Category) used in US Defense industry procurements, in which Risk Management figures prominently in decision making and planning. This includes not performing an activity that could present risk. Refusing to purchase a property or business to avoid legal liability is one such example. Avoiding airplane flights for fear of hijacking. Avoidance may seem like the answer to all risks, but avoiding risks also means losing out on the potential gain that accepting (retaining) the risk may have allowed. Not entering a business to avoid the risk of loss also avoids the possibility of earning profits. Increasing risk regulation in hospitals has led to avoidance of treating higher risk conditions, in favor of patients presenting with lower risk. Risk reduction or "optimization" involves reducing the severity of the loss or the likelihood of the loss from occurring. For example, sprinklers are designed to put out a fire to reduce the risk of loss by fire. This method may cause a greater loss by water damage and therefore may not be suitable. Halon fire suppression systems may mitigate that risk, but the cost may be prohibitive as a strategy. Acknowledging that risks can be positive or negative, optimizing risks means finding a balance between negative risk and the benefit of the operation or activity; and between risk reduction and effort applied. By effectively applying Health, Safety and Environment (HSE) management standards, organizations can achieve tolerable levels of residual risk. Modern software development methodologies reduce risk by developing and delivering software incrementally. Early methodologies suffered from the fact that they only delivered software in the final phase of development; any problems encountered in earlier phases meant costly rework and often jeopardized the whole project. By developing in iterations, software projects can limit effort wasted to a single iteration. Outsourcing could be an example of risk sharing strategy if the outsourcer can demonstrate higher capability at managing or reducing risks. For example, a company may outsource only its software development, the manufacturing of hard goods, or customer support needs to another company, while handling the business management itself. This way, the company can concentrate more on business development without having to worry as much about the manufacturing process, managing the development team, or finding a physical location for a center. Briefly defined as "sharing with another party the burden of loss or the benefit of gain, from a risk, and the measures to reduce a risk." The term of 'risk transfer' is often used in place of risk sharing in the mistaken belief that you can transfer a risk to a third party through insurance or outsourcing. In practice if the insurance company or contractor go bankrupt or end up in court, the original risk is likely to still revert to the first party. As such, in the terminology of practitioners and scholars alike, the purchase of an insurance contract is often described as a "transfer of risk." However, technically speaking, the buyer of the contract generally retains legal responsibility for the losses "transferred", meaning that insurance may be described more accurately as a post-event compensatory mechanism. For example, a personal injuries insurance policy does not transfer the risk of a car accident to the insurance company. The risk still lies with the policy holder namely the person who has been in the accident. The insurance policy simply provides that if an accident (the event) occurs involving the policy holder then some compensation may be payable to the policy holder that is commensurate with the suffering/damage. Methods of managing risk fall into multiple categories. Risk retention pools are technically retaining the risk for the group, but spreading it over the whole group involves transfer among individual members of the group. This is different from traditional insurance, in that no premium is exchanged between members of the group up front, but instead losses are assessed to all members of the group. Risk retention involves accepting the loss, or benefit of gain, from a risk when the incident occurs. True self-insurance falls in this category. Risk retention is a viable strategy for small risks where the cost of insuring against the risk would be greater over time than the total losses sustained. All risks that are not avoided or transferred are retained by default. This includes risks that are so large or catastrophic that either they cannot be insured against or the premiums would be infeasible. War is an example since most property and risks are not insured against war, so the loss attributed to war is retained by the insured. Also any amounts of potential loss (risk) over the amount insured is retained risk. This may also be acceptable if the chance of a very large loss is small or if the cost to insure for greater coverage amounts is so great that it would hinder the goals of the organization too much. Select appropriate controls or countermeasures to mitigate each risk. Risk mitigation needs to be approved by the appropriate level of management. For instance, a risk concerning the image of the organization should have top management decision behind it whereas IT management would have the authority to decide on computer virus risks. The risk management plan should propose applicable and effective security controls for managing the risks. For example, an observed high risk of computer viruses could be mitigated by acquiring and implementing antivirus software. A good risk management plan should contain a schedule for control implementation and responsible persons for those actions. According to ISO/IEC 27001, the stage immediately after completion of the risk assessment phase consists of preparing a Risk Treatment Plan, which should document the decisions about how each of the identified risks should be handled. Mitigation of risks often means selection of security controls, which should be documented in a Statement of Applicability, which identifies which particular control objectives and controls from the standard have been selected, and why. Implementation follows all of the planned methods for mitigating the effect of the risks. Purchase insurance policies for the risks that it has been decided to transferred to an insurer, avoid all risks that can be avoided without sacrificing the entity's goals, reduce others, and retain the rest. Initial risk management plans will never be perfect. Practice, experience, and actual loss results will necessitate changes in the plan and contribute information to allow possible different decisions to be made in dealing with the risks being faced. Risk analysis results and management plans should be updated periodically. There are two primary reasons for this: Prioritizing the "risk management processes" too highly could keep an organization from ever completing a project or even getting started. This is especially true if other work is suspended until the risk management process is considered complete. It is also important to keep in mind the distinction between risk and uncertainty. Risk can be measured by impacts × probability. If risks are improperly assessed and prioritized, time can be wasted in dealing with risk of losses that are not likely to occur. Spending too much time assessing and managing unlikely risks is to be avoided. Unlikely events do occur but if the risk is unlikely enough to occur it may be better to simply retain the risk and deal with the result if the loss does in fact occur. Qualitative risk assessment is subjective and lacks consistency. The primary justification for a formal risk assessment process is legal and bureaucratic. As applied to corporate finance, "risk management" is the technique for measuring, monitoring and controlling the financial or operational risk on a firm's balance sheet, a traditional measure is the value at risk (VaR), but there also other measures like profit at risk (PaR) or margin at risk. The Basel II framework breaks risks into market risk (price risk), credit risk and operational risk and also specifies methods for calculating capital requirements for each of these components. In Information Technology, Risk management includes "Incident Handling", an action plan for dealing with intrusions, cyber-theft, denial of service, fire, floods, and other security-related events. According to the SANS Institute, it is a six step process: Preparation, Identification, Containment, Eradication, Recovery, and Lessons Learned. In enterprise risk management, a risk is defined as a possible event or circumstance that can have negative influences on the enterprise in question. Its impact can be on the very existence, the resources (human and capital), the products and services, or the customers of the enterprise, as well as external impacts on society, markets, or the environment. In a financial institution, enterprise risk management is normally thought of as the combination of credit risk, interest rate risk or asset liability management, liquidity risk, market risk, and operational risk. In the more general case, every probable risk can have a pre-formulated plan to deal with its possible consequences (to ensure "contingency" if the risk becomes a "liability"). From the information above and the average cost per employee over time, or cost accrual ratio, a project manager can estimate: Risk in a project or process can be due either to Special Cause Variation or Common Cause Variation and requires appropriate treatment. That is to re-iterate the concern about extremal cases not being equivalent in the list immediately above. ESRM is a security program management approach that links security activities to an enterprise's mission and business goals through risk management methods. The security leader's role in ESRM is to manage risks of harm to enterprise assets in partnership with the business leaders whose assets are exposed to those risks. ESRM involves educating business leaders on the realistic impacts of identified risks, presenting potential strategies to mitigate those impacts, then enacting the option chosen by the business in line with accepted levels of business risk tolerance For medical devices, risk management is a process for identifying, evaluating and mitigating risks associated with harm to people and damage to property or the environment. Risk management is an integral part of medical device design and development, production processes and evaluation of field experience, and is applicable to all types of medical devices. The evidence of its application is required by most regulatory bodies such as the US FDA. The management of risks for medical devices is described by the International Organization for Standardization (ISO) in ISO 14971:2007, Medical Devices—The application of risk management to medical devices, a product safety standard. The standard provides a process framework and associated requirements for management responsibilities, risk analysis and evaluation, risk controls and lifecycle risk management. The European version of the risk management standard was updated in 2009 and again in 2012 to refer to the Medical Devices Directive (MDD) and Active Implantable Medical Device Directive (AIMDD) revision in 2007, as well as the In Vitro Medical Device Directive (IVDD). The requirements of EN 14971:2012 are nearly identical to ISO 14971:2007. The differences include three "(informative)" Z Annexes that refer to the new MDD, AIMDD, and IVDD. These annexes indicate content deviations that include the requirement for risks to be reduced "as far as possible", and the requirement that risks be mitigated by design and not by labeling on the medical device (i.e., labeling can no longer be used to mitigate risk). Typical risk analysis and evaluation techniques adopted by the medical device industry include hazard analysis, fault tree analysis (FTA), failure mode and effects analysis (FMEA), hazard and operability study (HAZOP), and risk traceability analysis for ensuring risk controls are implemented and effective (i.e. tracking risks identified to product requirements, design specifications, verification and validation results etc.). FTA analysis requires diagramming software. FMEA analysis can be done using a spreadsheet program. There are also integrated medical device risk management solutions. Through a draft guidance, the FDA has introduced another method named "Safety Assurance Case" for medical device safety assurance analysis. The safety assurance case is structured argument reasoning about systems appropriate for scientists and engineers, supported by a body of evidence, that provides a compelling, comprehensible and valid case that a system is safe for a given application in a given environment. With the guidance, a safety assurance case is expected for safety critical devices (e.g. infusion devices) as part of the pre-market clearance submission, e.g. 510(k). In 2013, the FDA introduced another draft guidance expecting medical device manufacturers to submit cybersecurity risk analysis information. Project risk management must be considered at the different phases of acquisition. In the beginning of a project, the advancement of technical developments, or threats presented by a competitor's projects, may cause a risk or threat assessment and subsequent evaluation of alternatives (see Analysis of Alternatives). Once a decision is made, and the project begun, more familiar project management applications can be used: Megaprojects (sometimes also called "major programs") are large-scale investment projects, typically costing more than $1 billion per project. Megaprojects include major bridges, tunnels, highways, railways, airports, seaports, power plants, dams, wastewater projects, coastal flood protection schemes, oil and natural gas extraction projects, public buildings, information technology systems, aerospace projects, and defense systems. Megaprojects have been shown to be particularly risky in terms of finance, safety, and social and environmental impacts. Risk management is therefore particularly pertinent for megaprojects and special methods and special education have been developed for such risk management. It is important to assess risk in regard to natural disasters like floods, earthquakes, and so on. Outcomes of natural disaster risk assessment are valuable when considering future repair costs, business interruption losses and other downtime, effects on the environment, insurance costs, and the proposed costs of reducing the risk. The Sendai Framework for Disaster Risk Reduction is a 2015 international accord that has set goals and targets for disaster risk reduction in response to natural disasters. There are regular International Disaster and Risk Conferences in Davos to deal with integral risk management. The management of risks to persons and property in wilderness and remote natural areas has developed with increases in outdoor recreation participation and decreased social tolerance for loss. Organizations providing commercial wilderness experiences can now align with national and international consensus standards for training and equipment such as ANSI/NASBLA 101-2017 (boating), UIAA 152 (ice climbing tools), and European Norm 13089:2015 + A1:2015 (mountaineering equipment). The Association for Experiential Education offers accreditation for wilderness adventure programs. The Wilderness Risk Management Conference provides access to best practices, and specialist organizations provide wilderness risk management consulting and training. In his book, "Outdoor Leadership and Education", climber, outdoor educator, and author, Ari Schneider, notes that outdoor recreation is inherently risky, and there is no way to completely eliminate risk. However, he explains how that can be a good thing for outdoor education programs. According to Schneider, optimal adventure is achieved when real risk is managed and perceived risk is maintained in order to keep actual danger low and a sense of adventure high. One popular models for risk assessment is the Risk Assessment and Safety Management (RASM) Model developed by Rick Curtis, author of The Backpacker's Field Manual. The formula for the RASM Model is: Risk = Probability of Accident × Severity of Consequences. The RASM Model weighs negative risk—the potential for loss, against positive risk—the potential for growth. IT risk is a risk related to information technology. This is a relatively new term due to an increasing awareness that information security is simply one facet of a multitude of risks that are relevant to IT and the real world processes it supports. "Cybersecurity is tied closely to the advancement of technology. It lags only long enough for incentives like black markets to evolve and new exploits to be discovered. There is no end in sight for the advancement of technology, so we can expect the same from cybersecurity." ISACA's "Risk IT" framework ties IT risk to enterprise risk management. Duty of Care Risk Analysis (DoCRA) evaluates risks and their safeguards and considers the interests of all parties potentially affected by those risks. CIS RAM provides a method to design and evaluate the implementation of the CIS Controls™. For the offshore oil and gas industry, operational risk management is regulated by the safety case regime in many countries. Hazard identification and risk assessment tools and techniques are described in the international standard ISO 17776:2000, and organisations such as the IADC (International Association of Drilling Contractors) publish guidelines for Health, Safety and Environment (HSE) Case development which are based on the ISO standard. Further, diagrammatic representations of hazardous events are often expected by governmental regulators as part of risk management in safety case submissions; these are known as bow-tie diagrams (see Network theory in risk assessment). The technique is also used by organisations and regulators in mining, aviation, health, defence, industrial and finance. The principles and tools for quality risk management are increasingly being applied to different aspects of pharmaceutical quality systems. These aspects include development, manufacturing, distribution, inspection, and submission/review processes throughout the lifecycle of drug substances, drug products, biological and biotechnological products (including the use of raw materials, solvents, excipients, packaging and labeling materials in drug products, biological and biotechnological products). Risk management is also applied to the assessment of microbiological contamination in relation to pharmaceutical products and cleanroom manufacturing environments. Risk communication is a complex cross-disciplinary academic field related to core values of the targeted audiences. Problems for risk communicators involve how to reach the intended audience, how to make the risk comprehensible and relatable to other risks, how to pay appropriate respect to the audience's values related to the risk, how to predict the audience's response to the communication, etc. A main goal of risk communication is to improve collective and individual decision making. Risk communication is somewhat related to crisis communication. Some experts coincide that risk is not only enrooted in the communication process but also it cannot be dissociated from the use of language. Though each culture develops its own fears and risks, these construes apply only by the hosting culture. Risk communication and community engagement (RCCE) is a method that draws heavily on volunteers, frontline personnel and on people without prior training in this area.
https://en.wikipedia.org/wiki?curid=26404
Conversion therapy Conversion therapy is the pseudoscientific practice of trying to change an individual's sexual orientation from homosexual or bisexual to heterosexual using psychological or spiritual interventions. There is no reliable evidence that sexual orientation can be changed and medical institutions warn that conversion therapy practices are ineffective and potentially harmful. Medical, scientific, and government organizations in the United States and United Kingdom have expressed concern over the validity, efficacy and ethics of conversion therapy. Various jurisdictions around the world have passed laws against conversion therapy. The American Psychiatric Association (APA) opposes psychiatric treatment "based upon the assumption that homosexuality "per se" is a mental disorder or based upon the "a priori" assumption that a patient should change his/her sexual homosexual orientation" and describes attempts to change a person's sexual orientation by practitioners as unethical. It also states that the advancement of conversion therapy may cause social harm by disseminating unscientific views about sexual orientation. In 2001, United States Surgeon General David Satcher issued a report stating that "there is no valid scientific evidence that sexual orientation can be changed". Techniques used in conversion therapy in the United States and Western Europe have included ice-pick lobotomies; chemical castration with hormonal treatment; aversive treatments, such as "the application of electric shock to the hands and/or genitals"; "nausea-inducing drugs ... administered simultaneously with the presentation of homoerotic stimuli"; and masturbatory reconditioning. More recent clinical techniques used in the United States have been limited to counseling, visualization, social skills training, psychoanalytic therapy, and spiritual interventions such as "prayer and group support and pressure", though there are some reports of aversive treatments through unlicensed practice as late as the early 2000s. The term reparative therapy has been used as a synonym for conversion therapy in general, but it has been argued that strictly speaking it refers to a specific kind of therapy associated with the psychologists Elizabeth Moberly and Joseph Nicolosi. The National Association for Research & Therapy of Homosexuality (NARTH) was the main organization advocating for conversion therapy. Fundamentalist Christian groups, and some other organizations, have used religious justification for the therapy. However, the International Rehabilitation Council for Torture Victims describes conversion therapy as a form of torture. The history of conversion therapy can be divided broadly into three periods: an early Freudian period; a period of mainstream approval of conversion therapy, when the mental health establishment became the "primary superintendent" of sexuality; and a post-Stonewall period where the mainstream medical profession disavowed conversion therapy. During the earliest parts of psychoanalytic history, analysts granted that homosexuality was non-pathological in certain cases, and the ethical question of whether it ought to be changed was discussed. By the 1920s analysts assumed that homosexuality was pathological and that attempts to treat it were appropriate, although psychoanalytic opinion about changing homosexuality was largely pessimistic. Those forms of homosexuality that were considered perversions were usually held to be incurable. Analysts' tolerant statements about homosexuality arose from recognition of the difficulty of achieving change. Beginning in the 1930s and continuing for roughly twenty years, major changes occurred in how analysts viewed homosexuality, which involved a shift in the rhetoric of analysts, some of whom felt free to ridicule and abuse their gay patients. Sigmund Freud was a physician and the founder of psychoanalysis. Freud stated that homosexuality could sometimes be removed through hypnotic suggestion, and was influenced by Eugen Steinach, a Viennese endocrinologist who transplanted testicles from straight men into gay men in attempts to change their sexual orientation, stating that his research had "thrown a strong light on the organic determinants of homo-eroticism". Freud cautioned that Steinach's operations would not necessarily make possible a therapy that could be generally applied, arguing that such transplant procedures would be effective in changing homosexuality in men only in cases in which it was strongly associated with physical characteristics typical of women, and that probably no similar therapy could be applied to lesbianism. Steinach's method was doomed to failure because the immune system rejects transplanted glands, and was eventually exposed as ineffective and often harmful. Freud's main discussion of female homosexuality was the 1920 paper "The Psychogenesis of a Case of Homosexuality in a Woman", which described his analysis of a young woman who had entered therapy because her parents were concerned that she was a lesbian. Her father wanted this condition changed. In Freud's view, the prognosis was unfavourable because of the circumstances under which she entered therapy, and because homosexuality was not an illness or neurotic conflict. Freud wrote that changing homosexuality was difficult and possible only under unusually favourable conditions, observing that "in general to undertake to convert a fully developed homosexual into a heterosexual does not offer much more prospect of success than the reverse". Success meant making heterosexual feeling possible, not eliminating homosexual feelings. Gay people could seldom be convinced that heterosexual sex would provide them with the same pleasure they derived from homosexual sex. Patients often wanted to become heterosexual for reasons Freud considered superficial, including fear of social disapproval, an insufficient motive for change. Some might have no real desire to become heterosexual, seeking treatment only to convince themselves that they had done everything possible to change, leaving them free to return to homosexuality after the failure they expected. In 1935, a mother asked Freud to treat her son. Freud replied in a letter that later became famous: Sándor Ferenczi was an influential psychoanalyst. Ferenczi hoped to cure some kinds of homosexuality completely, but was content in practice with reducing what he considered gay men's hostility to women, along with the urgency of their homosexual desires, and with helping them to become attracted to and potent with women. In his view, a gay man who was confused about his sexual identity and felt himself to be "a woman with the wish to be loved by a man" was not a promising candidate for cure. Ferenczi believed that complete cures of homosexuality might become possible in the future when psychoanalytic technique had been improved. Daughter of Sigmund Freud, Anna Freud became an influential psychoanalytic theorist in the UK. Anna Freud reported the successful treatment of homosexuals as neurotics in a series of unpublished lectures. In 1949 she published "Some Clinical Remarks Concerning the Treatment of Cases of Male Homosexuality" in the "International Journal of Psychoanalysis". In her view, it was important to pay attention to the interaction of passive and active homosexual fantasies and strivings, the original interplay of which prevented adequate identification with the father. The patient should be told that his choice of a passive partner allows him to enjoy a passive or receptive mode, while his choice of an active partner allows him to recapture his lost masculinity. She claimed that these interpretations would reactivate repressed castration anxieties, and childhood narcissistic grandiosity and its complementary fear of dissolving into nothing during heterosexual intercourse would come with the renewal of heterosexual potency. Anna Freud in 1951 published "Clinical Observations on the Treatment of Male Homosexuality" in "The Psychoanalytic Quarterly" and "Homosexuality" in the "American Psychoanalytic Association (APsaA) Bulletin". In these articles, she insisted on the attainment of full object-love of the opposite sex as a requirement for cure of homosexuality. In 1951 she gave a lecture about treatment of homosexuality which was criticised by Edmund Bergler, who emphasised the oral fears of patients and minimized the importance of the phallic castration fears she had discussed. Anna Freud recommended in 1956 to a journalist who was preparing an article about psychoanalysis for "The Observer" of London that she not quote Freud's letter to the American mother, on the grounds that "nowadays we can cure many more homosexuals than was thought possible in the beginning. The other reason is that readers may take this as a confirmation that all analysis can do is to convince patients that their defects or 'immoralities' do not matter and that they should be happy with them. That would be unfortunate." Melanie Klein was a pupil of Ferenczi. Her seminal book "The Psycho-Analysis of Children", based on lectures given to the British Psychoanalytical Society in the 1920s, was published in 1932. Klein claimed that entry into the Oedipus Complex is based on mastery of primitive anxiety from the oral and anal stages. If these tasks are not performed properly, developments in the Oedipal stage will be unstable. Complete analysis of patients with such unstable developments would require uncovering these early concerns. The analysis of homosexuality required dealing with paranoid trends based on the oral stage. "The Psycho-Analysis of Children" ends with the analysis of Mr. B., a gay man. Klein claimed that he illustrated pathologies that enter into all forms of homosexuality: a gay man idealizes "the good penis" of his partner to allay the fear of attack he feels due to having projected his paranoid hatred onto the imagined "bad penis" of his mother as an infant. She stated that Mr. B.'s homosexual behaviour diminished after he overcame his need to adore the "good penis" of an idealized man. This was made possible by his recovering his belief in the good mother and his ability to sexually gratify her with his good penis and plentiful semen. In March 2018, a majority of 435 against 109 representatives in the European parliament passed a resolution condemning conversion therapy and urging European Union member states to ban the practice. On May 2020, Albania became the third european country (after Malta (2016) and Germany (2020)) to ban conversion therapy or any pseudo-therapeutic attempts to change a person's sexual orientation or gender identity. On May 7, 2020, German parliament Bundestag banned nationwide conversion therapy for minors until 18 years and forbids advertising of conversion therapy. It also forbids conversion therapy for adults, if they are decided by force, fraud or pressure. On December 6, 2016, Malta became the first country in the European Union to prohibit the use of conversion therapy. Psychoanalysis started to receive recognition in the United States in 1909, when Sigmund Freud delivered a series of lectures at Clark University in Massachusetts at the invitation of G. Stanley Hall. In 1913, Abraham Brill wrote "The Conception of Homosexuality", which he published in the "Journal of the American Medical Association" and read before the American Medical Association's annual meeting. Brill criticised physical treatments for homosexuality such as bladder washing, rectal massage, and castration, along with hypnosis, but referred approvingly to Freud and Sadger's use of psychoanalysis, calling its results "very gratifying". Since Brill understood curing homosexuality as restoring heterosexual potency, he claimed that he had cured his patients in several cases, even though many remained homosexual. Wilhelm Stekel, an Austrian, published his views on treatment of homosexuality, which he considered a disease, in the American "Psychoanalytic Review" in 1930. Stekel believed that "success was fairly certain" in changing homosexuality through psychoanalysis provided that it was performed correctly and the patient wanted to be treated. In 1932, "The Psychoanalytic Quarterly" published a translation of Helene Deutsch's paper "On Female Homosexuality". Deutsch reported her analysis of a lesbian, who did not become heterosexual as a result of treatment, but who managed to achieve a "positive libidinal relationship" with another woman. Deutsch indicated that she would have considered heterosexuality a better outcome. Edmund Bergler was the most important psychoanalytic theorist of homosexuality in the 1950s. He was vociferous in his opposition to Alfred Kinsey. Kinsey's work, and its reception, led Bergler to develop his own theories for treatment, which were essentially to "blame the victim", in the evaluation of Jennifer Terry, associate professor of Woman's Studies. Bergler claimed that if gay people wanted to change, and the right therapeutic approach was taken, then they could be cured in 90% of cases. Bergler used confrontational therapy in which gay people were punished in order to make them aware of their masochism. Bergler openly violated professional ethics to achieve this, breaking patient confidentiality in discussing the cases of patients with other patients, bullying them, calling them liars and telling them they were worthless. He insisted that gay people could be cured. Bergler confronted Kinsey because Kinsey thwarted the possibility of cure by presenting homosexuality as an acceptable way of life, which was the basis of the gay rights activism of the time. Bergler popularised his views in the United States in the 1950s using magazine articles and books aimed at non-specialists. In 1951, the mother who wrote to Freud asking him to treat her son sent Freud's response to the "American Journal of Psychiatry", in which it was published. The 1952 first edition of the American Psychiatric Association's Diagnostic and Statistical Manual of Mental Disorders (DSM-I) classified homosexuality as a mental disorder. During the three decades between Freud's death in 1939 and the Stonewall riots in 1969, conversion therapy received approval from most of the psychiatric establishment in the United States. In 1962, Irving Bieber "et al." published "", in which they concluded that "although this change may be more easily accomplished by some than by others, in our judgment a heterosexual shift is a possibility for all homosexuals who are strongly motivated to change". There was a riot in 1969 at the Stonewall Bar in New York after a police raid. The Stonewall riot acquired symbolic significance for the gay rights movement and came to be seen as the opening of a new phase in the struggle for gay liberation. Following these events, conversion therapy came under increasing attack. Activism against conversion therapy increasingly focused on the DSM's designation of homosexuality as a psychopathology. In 1973, after years of criticism from gay activists and bitter dispute among psychiatrists, the American Psychiatric Association removed homosexuality as a mental disorder from the "Diagnostic and Statistical Manual of Mental Disorders". Supporters of the change used evidence from researchers such as Kinsey and Evelyn Hooker. Psychiatrist Robert Spitzer, a member of the APA's Committee on Nomenclature, played an important role in the events that led to this decision. Critics argued that it was a result of pressure from gay activists, and demanded a referendum among voting members of the Association. The referendum was held in 1974 and the APA's decision was upheld by a 58% majority. The APA removed ego-dystonic homosexuality from the DSM-III-R in 1987 and opposes the diagnosis of either homosexuality or ego-dystonic homosexuality as any type of disorder. Joseph Nicolosi had a significant role in the development of conversion therapy as early as the 1990s, publishing his first book, "Reparative Therapy of Male Homosexuality", in 1991. In 1992, Nicolosi, with Charles Socarides and Benjamin Kaufman, founded the National Association for Research & Therapy of Homosexuality (NARTH), an organization that opposed the mainstream medical view of homosexuality and aimed to "make effective psychological therapy available to all homosexual men and women who seek change". NARTH has operated under the name "Alliance for Therapeutic Choice and Scientific Integrity" (ATCSI) since 2014. In 1998, Christian right groups including the Family Research Council and the American Family Association spent $600,000 on advertising promoting conversion therapy. John Paulk and his then wife Anne featured in full-page newspaper spreads. United States Surgeon General David Satcher in 2001 issued a report stating that "there is no valid scientific evidence that sexual orientation can be changed". The same year, a study by Robert Spitzer concluded that some highly motivated individuals whose orientation is predominantly homosexual can become predominantly heterosexual with some form of reparative therapy. Spitzer based his findings on structured interviews with 200 self-selected individuals (143 males, 57 females). He told "The Washington Post" that the study "shows some people can change from gay to straight, and we ought to acknowledge that". Spitzer's study caused controversy and attracted media attention. Spitzer recanted his study in 2012, and apologized to the gay community for making unproven claims of the efficacy of reparative therapy, calling it his only professional regret. The American Psychoanalytic Association spoke against NARTH in 2004, stating "that organization does not adhere to our policy of nondiscrimination and ... their activities are demeaning to our members who are gay and lesbian". The same year, a survey of members of the American Psychological Association rated reparative therapy as "certainly discredited", though the authors warn that the results should be interpreted carefully as an initial step, not as a final deliberation. The American Psychological Association in 2007 convened a task force to evaluate its policies regarding reparative therapy. In 2008, the organizers of an APA panel on the relationship between religion and homosexuality canceled the event after gay activists objected that "conversion therapists and their supporters on the religious right use these appearances as a public relations event to try and legitimize what they do". In 2009, American Psychological Association stated that it "encourages mental health professionals to avoid misrepresenting the efficacy of sexual orientation change efforts by promoting or promising change in sexual orientation when providing assistance to individuals distressed by their own or others' sexual orientation and concludes that the benefits reported by participants in sexual orientation change efforts can be gained through approaches that do not attempt to change sexual orientation". The ethics guidelines of major mental health organizations in the United States vary from cautionary statements to recommendations that ethical practitioners refrain from practicing conversion therapy (American Psychiatric Association) or from referring patients to those who do (American Counseling Association). In a letter dated February 23, 2011 to the Speaker of the U.S. House of Representatives, the Attorney General of the United States stated "while sexual orientation carries no visible badge, a growing scientific consensus accepts that sexual orientation is a characteristic that is immutable". Gay rights groups and other groups concerned with mental health fear that reparative therapy can increase the chances of depression and suicide. President Barack Obama expressed opposition to the practice in 2015. Before the American Psychological Association's 1973 decision to remove homosexuality from the DSM, practitioners of conversion therapy employed aversive conditioning techniques, involving electric shock and nausea-inducing drugs during presentation of same-sex erotic images. Cessation of the aversive stimuli was typically accompanied by the presentation of opposite-sex erotic images, with the objective of strengthening heterosexual feelings. In "Aversion therapy for sexual deviation: a critical review", published in 1966, M. P. Feldman claimed a 58% cure rate, but Douglas Haldeman is skeptical that such stressful methods permit feelings of sexual responsiveness, and notes that Feldman defined success as suppression of homosexuality and increased capacity for heterosexual behavior. Another method used was the covert sensitization method, which involves instructing patients to imagine vomiting or receiving electric shocks, writing that only single case studies have been conducted, and that their results cannot be generalized. Haldeman writes that behavioral conditioning studies tend to decrease homosexual feelings, but do not increase heterosexual feelings, citing Rangaswami's "Difficulties in arousing and increasing heterosexual responsiveness in a homosexual: A case report", published in 1982, as typical in this respect. Haldeman concludes that such methods can be called torture, besides being ineffective. He writes that "Individuals undergoing such treatments do not emerge heterosexually inclined; rather they become shamed, conflicted, and fearful about their homosexual feelings." Some sources describe ex-gay ministries as a form of conversion therapy, while others state that ex-gay organizations and conversion therapy are distinct methods of attempting to convert gay people to heterosexuality. Ex-gay ministries have also been called transformational ministries. Some state that they do not conduct clinical treatment of any kind. Exodus International once believed reparative therapy could be a beneficial tool. The umbrella organization in the United States ceased activities in June 2013, and the three member board issued a statement which repudiated its aims and apologized for the harm their pursuit has caused to LGBT people. Haldeman writes that psychoanalytic treatment of homosexuality is exemplified by the work of Irving Bieber "et al." in "Homosexuality: A Psychoanalytic Study of Male Homosexuals". They advocated long-term therapy aimed at resolving the unconscious childhood conflicts that they considered responsible for homosexuality. Haldeman notes that Bieber's methodology has been criticized because it relied upon a clinical sample, the description of the outcomes was based upon subjective therapist impression, and follow-up data were poorly presented. Bieber reported a 27% success rate from long-term therapy, but only 18% of the patients in whom Bieber considered the treatment successful had been exclusively homosexual to begin with, while 50% had been bisexual. In Haldeman's view, this makes even Bieber's unimpressive claims of success misleading. Haldeman discusses other psychoanalytic studies of attempts to change homosexuality. Curran and Parr's "Homosexuality: An analysis of 100 male cases", published in 1957, reported no significant increase in heterosexual behavior. Mayerson and Lief's "Psychotherapy of homosexuals: A follow-up study of nineteen cases", published in 1965, reported that half of its 19 subjects were exclusively heterosexual in behavior four and a half years after treatment, but its outcomes were based on patient self-report and had no external validation. In Haldeman's view, those participants in the study who reported change were bisexual at the outset, and its authors wrongly interpreted capacity for heterosexual sex as change of sexual orientation. The term "reparative therapy" has been used as a synonym for conversion therapy generally, but according to Jack Drescher it properly refers to a specific kind of therapy associated with the psychologists Elizabeth Moberly and Joseph Nicolosi. The term "reparative" refers to Nicolosi's postulate that same-sex attraction is a person's unconscious attempt to "self-repair" feelings of inferiority. Most mental health professionals and the American Psychological Association consider reparative therapy discredited, but it is still practiced by some. In 2014 the Republican Party of Texas endorsed "counseling, which offers reparative therapy and treatment" in their party platform. Exodus International regarded reparative therapy as a useful tool to eliminate "unwanted same-sex attraction" but ceased activities in June 2013 and issued a statement repudiating its aims and apologizing for the harm the organization had caused to LGBT people. Psychoanalysts critical of Nicolosi's theories have offered gay-affirmative approaches as an alternative to reparative therapy. Haldeman has described William Masters' and Virginia Johnson's work on sexual orientation change as a form of conversion therapy. In "Homosexuality in Perspective", published in 1979, Masters and Johnson viewed homosexuality as the result of blocks that prevented the learning that facilitated heterosexual responsiveness, and described a study of 54 gay men who were dissatisfied with their sexual orientation. The original study did not describe the treatment methodology used, but this was published five years later. John C. Gonsiorek criticized their study on several grounds in 1981, pointing out that while Masters and Johnson stated that their patients were screened for major psychopathology or severe neurosis, they did not explain how this screening was performed, or how the motivation of the patients to change was assessed. Nineteen of their subjects were described as uncooperative during therapy and refused to participate in a follow-up assessment, but all of them were assumed without justification to have successfully changed. Haldeman writes that Masters and Johnson's study was founded on heterosexist bias, and that it would be tremendously difficult to replicate. In his view, the distinction Masters and Johnson made between "conversion" (helping gay men with no previous heterosexual experience to learn heterosexual sex) and "reversion" (directing men with some previous heterosexual experience back to heterosexuality) was not well founded. Many of the subjects Masters and Johnson labelled homosexual may not have been homosexual, since, of their participants, only 17% identified themselves as exclusively homosexual, while 83% were in the predominantly heterosexual to bisexual range. Haldeman observed that since 30% of the sample was lost to the follow-up, it is possible that the outcome sample did not include any people attracted mainly or exclusively to the same sex. Haldeman concludes that it is likely that, rather than converting or reverting gay people to heterosexuality, Masters and Johnson only strengthened heterosexual responsiveness in people who were already bisexual. In the 1940s and 1950s, U.S. neurologist Walter Freeman popularized the ice-pick lobotomy as a treatment for homosexuality. He personally performed as many as 3,439 lobotomy surgeries in 23 states, of which 2,500 used his ice-pick procedure, despite the fact that he had no formal surgical training. Up to 40% of Freeman's patients were gay individuals subjected to a lobotomy in order to change their homosexual orientation, leaving most of these individuals severely disabled for the rest of their lives. While promoted at the time as a treatment for various psychoses, the effectiveness of lobotomy in changing sexual orientation was already the subject of critical research in 1948 when a single case was investigated by Joseph Friedlander and Ralph Banay. A video depicting the "ice-pick lobotomy" of a homosexual man was featured in the documentary film, "". In May 2001, Robert Spitzer presented "Can Some Gay Men and Lesbians Change Their Sexual Orientation? 200 Participants Reporting a Change from Homosexual to Heterosexual Orientation", a study of attempts to change homosexual orientation through ex-gay ministries and conversion therapy, at the American Psychiatric Association's convention in New Orleans. The study was partly a response to the APA's 2000 statement cautioning against clinical attempts at changing homosexuality, and was aimed at determining whether such attempts were ever successful rather than how likely it was that change would occur for any given individual. Spitzer wrote that some earlier studies provided evidence for the effectiveness of therapy in changing sexual orientation, but that all of them suffered from methodological problems. In 2012, Spitzer renounced and retracted this study, stating "I was quite wrong in the conclusions that I made from this study. The study does not provide evidence, really, that gays can change. And that's quite an admission on my part." He also apologized to the gay community for making unproven claims of the efficacy of reparative therapy, calling it his only professional regret. Spitzer has requested that all "ex-gay" therapy organizations such as NARTH, PFOX, American College of Pediatricians, and Focus on the Family stop citing his study as evidence for conversion therapy. Spitzer reported that after intervention, 66% of the men and 44% of the women had achieved "Good Heterosexual Functioning", which he defined as requiring five criteria (being in a loving heterosexual relationship during the last year, overall satisfaction in emotional relationship with a partner, having heterosexual sex with the partner at least a few times a month, achieving physical satisfaction through heterosexual sex, and not thinking about having homosexual sex more than 15% of the time while having heterosexual sex). He found that the most common reasons for seeking change were lack of emotional satisfaction from gay life, conflict between same-sex feelings and behavior and religious beliefs, and desire to marry or remain married. This paper was widely reported in the international media and taken up by politicians in the United States, Germany, and Finland, and by conversion therapists. In 2003, Spitzer published the paper in the "Archives of Sexual Behavior". Spitzer's study has been criticized on numerous ethical and methodological grounds, and "press releases from both NGLTF and HRC sought to undermine Spitzer's credibility by connecting him politically to right-wing groups that had backed the ex-gay movement". Gay activists argued that the study would be used by conservatives to undermine gay rights. Spitzer acknowledged that the study sample consisted of people who sought treatment primarily because of their religious beliefs (93% of the sample), served in various church-related functions, and who publicly spoke in favor of changing homosexual orientation (78%), and thus were strongly motivated to overreport success. Critics felt he dismissed this source of bias, without even attempting to measure deception or self-deception (a standard practice in self-reporting psychological tests like MMPI-2). That participants had to rely upon their memories of what their feelings were before treatment may have distorted the findings. It was impossible to determine whether any change that occurred was due to the treatment because it was not clear what it involved and there was no control group. Spitzer's own data showed that claims of change were reflected mostly in changes in self-labelling and behavior, less in attractions, and least in the homoerotic content during the masturbatory fantasies; this particular finding was consistent with other studies in this area. Participants may have been bisexual before treatment. Follow-up studies were not conducted. Spitzer stressed the limitations of his study. Spitzer said that the number of gay people who could successfully become heterosexual was likely to be "pretty low", and conceded that his subjects were "unusually religious". Ariel Shidlo and Michael Schroeder found in "Changing Sexual Orientation: A Consumer's Report", a peer-reviewed study of 202 respondents published in 2002, that 88% of participants failed to achieve a sustained change in their sexual behavior and 3% reported changing their orientation to heterosexual. The remainder reported either losing all sexual drive or attempting to remain celibate, with no change in attraction. Some of the participants who failed felt a sense of shame and had gone through conversion therapy programs for many years. Others who failed believed that therapy was worthwhile and valuable. Many respondents felt harmed by the attempt to change, and reported depression, suicidal ideation and attempts, hypervigilance of gender-deviant mannerisms, social isolation, fear of being a child abuser and poor self-esteem. Of the 8 respondents (out of a sample of 202) who reported a change in sexual orientation, 7 worked as ex-gay counselors or group leaders. Although no national ban exists, several US states and individual counties ban therapy attempting to change sexual orientation as shown in the map below. Many health organizations around the world have denounced and criticized sexual orientation change efforts. National health organizations in the United States have announced that there has been no scientific demonstration of conversion therapy's efficacy in the last forty years. They find that conversion therapy is ineffective, risky and can be harmful. Anecdotal claims of cures are counterbalanced by assertions of harm, and the American Psychiatric Association, for example, cautions ethical practitioners under the Hippocratic oath to do no harm and to refrain from attempts at conversion therapy. Mainstream medical bodies state that conversion therapy can be harmful because it may exploit guilt and anxiety, thereby damaging self-esteem and leading to depression and even suicide. There is also concern in the mental health community that the advancement of conversion therapy can cause social harm by disseminating inaccurate views about sexual orientation and the ability of gay and bisexual people to lead happy, healthy lives. Major health organizations critical of conversion therapy include: The American Psychological Association undertook a study of the peer-reviewed literature in the area of sexual orientation change efforts (SOCE) and found myriad issues with the procedures used in conducting the research. The taskforce did find that some participants experienced a lessening of same sex attraction and arousal, but that these instances were "rare" and "uncommon". The taskforce concluded that, "given the limited amount of methodically sound research, claims that recent SOCE is effective are not supported". Two issues with SOCE claims are that conversion therapists falsely assume that homosexuality is a mental disorder and that their research focuses almost exclusively on gay men and rarely includes lesbians. The American Psychological Association's code of conduct states that "Psychologists respect the dignity and worth of all people, and the rights of individuals to privacy, confidentiality, and self-determination", but also that "Psychologists are aware that special safeguards may be necessary to protect the rights and welfare of persons or communities whose vulnerabilities impair autonomous decision making." The American Counseling Association says that "it is of primary importance to respect a client's autonomy to request a referral for a service not offered by a counselor". They said that no one should be forced to attempt to change their sexual orientation against their will, including children being forced by their parents. Supporters of SOCE focus on patient self-determination when discussing whether therapy should be available. Mark Yarhouse, of Pat Robertson's Regent University, wrote that "psychologists have an ethical responsibility to allow individuals to pursue treatment aimed at curbing experiences of same-sex attraction or modifying same-sex behaviors, not only because it affirms the client's rights to dignity, autonomy, and agency, as persons presumed capable of freely choosing among treatment modalities and behavior, but also because it demonstrates regard for diversity". Yarhouse and Throckmorton, of the private Christian school Grove City College, argue that the procedure should be available out of respect for a patient's values system and because they find evidence that it can be effective. Haldeman similarly argues for a client's right to access to therapy if requested from a fully informed position: "For some, religious identity is so important that it is more realistic to consider changing sexual orientation than abandoning one's religion of origin ... and if there are those who seek to resolve the conflict between sexual orientation and spirituality with conversion therapy, they must not be discouraged." In response to Yarhouse's paper, Jack Drescher argued that "any putative ethical obligation to refer a patient for reparative therapy is outweighed by a stronger ethical obligation to keep patients away from mental health practitioners who engage in questionable clinical practices". Chuck Bright wrote that refusing to endorse a procedure that "has been deemed unethical and potentially harmful by most medical and nearly every professional psychotherapy regulating body cannot be justifiably identified as prohibiting client self-determination". Some commentators, recommending a hard stand against the practice, have found therapy inconsistent with a psychologist's ethical duties because "it is more ethical to let a client continue to struggle honestly with her or his identity than to collude, even peripherally, with a practice that is discriminatory, oppressive, and ultimately ineffective in its own stated ends". They argue that clients who request it do so out of social pressure and internalized homophobia, pointing to evidence that rates of depression, anxiety, alcohol and drug abuse and suicidal feelings are roughly doubled in those who undergo therapy. Haldeman argues that, due to concern for people whose "spiritual or religious concerns" may assume priority over their sexual orientation, mental health organizations do not ban conversion therapy outright. In 1998, the American Psychiatric Association issued a statement opposing any treatment which is based upon the assumption that homosexuality is a mental disorder or that a person should change their orientation, but did not have a formal position on other treatments that attempt to change a person's sexual orientation. In 2000, they augmented that statement by saying that as a general principle, a therapist should not determine the goal of treatment, but recommends that ethical practitioners refrain from attempts to change clients' sexual orientation until more research is available. The American Counseling Association has stated that they do not condone any training to educate and prepare a counselor to practice conversion therapy. Counselors who do offer training in conversion therapy must inform students that the techniques are unproven. They suggest counselors do not refer clients to a conversion therapist or to proceed cautiously once they know the counselor fully informs clients of the unproven nature of the treatment and the potential risks. However, "it is of primary importance to respect a client's autonomy to request a referral for a service not offered by a counselor". A counselor performing conversion therapy must provide complete information about the treatment, offer referrals to gay-affirmative counselors, discuss the right of clients, understand the client's request within a cultural context, and only practice within their level of expertise. NARTH stated in 2012 that refusing to offer therapy aimed at change to a client who requests it, and telling them that their only option is to claim a gay identity, could also be considered ethically unacceptable. In 2012 the British Psychological Society issued a position statement opposing any treatments that are based on an assumption that non-heterosexual orientations are pathological. A 2013 article by the Committee on Adolescence of the American Academy of Pediatrics stated "Referral for 'conversion' or 'reparative therapy' is never indicated; therapy is not effective and may be harmful to LGBTQ individuals by increasing internalized stigma, distress, and depression." In 2014, the American Association of Christian Counselors amended its code of ethics to eliminate the promotion of conversion therapy for homosexuals and encouraged them to be celibate instead. An article in the American Medical Association's Journal of Ethics argues that if a pediatrician learns that parents of a 12-year-old patient seek conversion therapy, the pediatrician can advise against "the ineffective and potentially harmful intervention" while being culturally sensitive of their religious objections to homosexuality. The authors argue that the doctor's medical ethics means they should place the interests of the patient above the cultural sensitivities of the parents, and confidentially counsel the patient about resources for LGBT youth facing bullying, and advise the parents about resources for parents of LGBT children. In 2014, major therapy professional bodies in the United Kingdom issued a joint consensus statement opposing conversion therapy. Professional bodies supporting the statement included the UK Council for Psychotherapy, the British Psychoanalytic Council, the Royal College of Psychiatrists, the British Association for Counselling and Psychotherapy, the British Psychological Society and the National Counselling Society. In 2015, with support of the UK Government's Department of Health, a wide range of UK organisations signed a memorandum of understanding (MoU) setting out an agreed framework for activities by parties concerned to help address the issues raised by the practice of conversion therapy in the UK. In addition to many of the professional bodies that previously issued the consensus statement, signatories included the UK Association of Christian Counsellors, the Royal College of General Practitioners, NHS England and NHS Scotland. The signatory organisations recognised a shared commitment to protecting the public from the risks of conversion therapy. They committed to raise awareness among healthcare professionals and psychological therapists of ethical issues involved in conversion therapy and to provide training to enable therapists to support clients in distress in an appropriate way. The World Health Organization's ICD-10, which along with the DSM-IV is widely used internationally, states that "sexual orientation by itself is not to be regarded as a disorder". It lists ego-dystonic sexual orientation as a disorder instead, which it defines as occurring where "the gender identity or sexual preference (heterosexual, homosexual, bisexual, or prepubertal) is not in doubt, but the individual wishes it were different because of associated psychological and behavioural disorders, and may seek treatment in order to change it". In 2012, the Pan American Health Organization (the North and South American branch of the World Health Organization) released a statement cautioning against services that purport to "cure" people with non-heterosexual sexual orientations as they lack medical justification and represent a serious threat to the health and well-being of affected people, and noted that the global scientific and professional consensus is that homosexuality is a normal and natural variation of human sexuality and cannot be regarded as a pathological condition. The Pan American Health Organization further called on governments, academic institutions, professional associations and the media to expose these practices and to promote respect for diversity. The World Health Organization affiliate further noted that gay minors have sometimes been forced to attend these "therapies" involuntarily, being deprived of their liberty and sometimes kept in isolation for several months, and that these findings were reported by several United Nations bodies. Additionally, the Pan American Health Organization recommended that such practices be denounced and subject to sanctions and penalties under national legislation, as they constitute a violation of the ethical principles of health care and violate human rights that are protected by international and regional agreements. The development of theoretical models of sexual orientation in countries outside the United States that have established mental health professions often follows the history within the U.S. (although often at a slower pace), shifting from pathological to non-pathological conceptions of homosexuality. Major medical and psychological bodies in Australia uniformly prohibit conversion therapy practices, with published statements having come from peak bodies representing psychologists, psychiatrists, and medical practitioners. In a statement issued jointly with the College of Psychiatrists, Royal Australasian College of Physicians President Catherine Yelland summarised the view of the Australian medical community: "[g]ay conversion therapy is unethical, harmful and not supported by medical evidence." The approaches taken by peak medical bodies is exemplified by the 2015 Australian Psychological Society Position Statement, which declares (emphasis in original) that the: The Position Statement supports this position by reference to the Society's Code of Ethics, which were adopted in 2007 and mandated as the Code of Ethics for Australian psychologists in 2010 by the Psychology Board of Australia. Under the Code, psychologists are required to "avoid discriminating unfairly against people on the basis of age, religion, sexuality, ethnicity, gender, disability, or any other basis proscribed by law" and mandates that they The Position Statement explicitly states that this ethical "requirement not to discriminate and to respect clients' moral rights does not equate to a justification to treat homosexuality or bisexuality as a disorder requiring treatment," relying on the Code of Ethics' section on propriety: "psychologists only provide psychological services within the boundaries of their professional competence [which] includes but is not restricted to ... basing their service on established knowledge of the discipline and profession of psychology". Regarding the knowledge base relating to conversion therapy, the statement is unequivocal (emphasis in original): The Society's position concludes by noting that it "is, of course, appropriate for psychologists to provide clinical services to clients who experience distress in regards to their sexual orientation ... [but this practice] should seek to understand "the reasons for" distress and how it may be alleviated. Evidence-based strategies to alleviate distress do not include attempts at changing sexual orientation, but could include challenging negative stereotypes, seeking social support, and self-acceptance, among others." The Government of Victoria announced in 2016 that it would be legislating to ban all LGBTQI conversion therapy. The new law began operating in February 2017 and allows the Health Complaints Commissioner to act against any health professional engaged in practices that are "found to be making false claims and to be acting in a manner that puts people's physical, mental or psychological health, safety or welfare at risk"—and in a world first, this law applies to conversion therapy for adults as well as for minors. Western Australia and the Australian Capital Territory announced in September 2017 that they are investigating similar laws. Advocates for a ban on conversion therapy argued that reviews need to go beyond the practices of health professionals and into activities of religious groups and the unregulated (non-medical) counselling sector. A Fairfax Media investigation in 2018 reported that "across Australia, organisations who believe that LGBTI people can or should change are hard at work. Conversion practices are hidden in evangelical churches and ministries, taking the form of exorcisms, prayer groups or counselling disguised as pastoral care. They're also present in some religious schools or practised in the private offices of health professionals. They're pushed out through a thriving network of courses and mentors in the borderless world of cyberspace, cloaked in the terminology of 'self improvement' or 'spiritual healing.'" A study of Pentecostal-Charismatic Churches found that LGBTI parishioners were faced with four options: remain closeted, come out but commit to remaining celibate, undergo conversion therapy, or leave the church... the majority took the last option, though typically only after "agonising attempts to reconcile their faith and their sexuality." The study provides corroboration that conversion therapy remains practiced within religious communities. Following the Fairfax investigation, Victorian Premier Daniel Andrews called on Prime Minister Malcolm Turnbull to support outlawing conversion therapy as part of the national mental health strategy. Federal Health Minister Greg Hunt declared that the issue is one for the states as no Commonwealth funding goes to sexual orientation change efforts—though "gay conversion ideology has been quietly pushed in schools as part of the federal government's chaplaincy program." The report noted that the Victorian law applies only to people offering health services and so does not catch religious groups and charities "who say they are helping same-sex attracted people to live in accordance with their faith." Chris, a survivor of conversion therapy, joined Andrews in calling for the Federal Government to outlaw conversion therapy, declaring that "praying the gay away nearly killed me." He established a petition calling on Turnbull and Hunt to act to outlaw conversion therapy, declaring: "I prayed to God asking him to either heal me, or kill me. I was so depressed, I wanted to die." In April 2018, Shadow Health Minister Catherine King wrote a response to the petition: "I'm writing to let you know that Labor stands with you, Chris Csabs and the medical experts in opposing gay conversion therapy...two Turnbull Government ministers—the Acting Prime Minister and the Health Minister—have now failed to condemn the practice when given the chance." Shortly after Catherine King's response, Health Minister for Queensland Dr Steven Miles voiced his concerns over the practise and stated that the Federal Health Minister should be working with the states to enact change. In May 2018, the Victorian Health Minister Jill Hennessy called for an inquiry into gay conversion therapies. In an unprecedented move, the state government indicated it would not only investigate health professionals but will focus on religious and faith-based ministries propagating Gay Conversion ideologies. The following day, Health Minister for the Australian Capital Territory Meegan Fitzharris followed Catherine King's lead by also responding to the petition, stating that, "The ACT government will ban gay conversion therapy. It is abhorrent and completely inconsistent with the inclusive values of Canberrans." On June 25, 2015, a New Jersey jury found the Jewish conversion therapy organization JONAH guilty of consumer fraud in the case Ferguson v JONAH for promising to be able to change its clients' sexual urges and determined its commercial practices to be unconscionable. In a 1997 U.S. case, the Ninth Circuit addressed conversion therapy in the context of an asylum application. A Russian citizen "had been apprehended by the Russian militia, registered at a clinic as a 'suspected lesbian', and forced to undergo treatment for lesbianism, such as 'sedative drugs' and hypnosis. ... The Ninth Circuit held that the conversion treatments to which Pitcherskaia had been subjected constituted mental and physical torture." The court rejected the argument that the treatments to which Pitcherskaia had been subjected did not constitute persecution because they had been intended to help her, not harm her, and stated "human rights laws cannot be sidestepped by simply couching actions that torture mentally or physically in benevolent terms such as 'curing' or 'treating' the victims". In 1993, the Superior Court of San Francisco's Family Court placed 15-year-old lesbian Lyn Duff under the guardianship of a foster couple after her mother committed her to Rivendell Psychiatric Center in West Jordan, Utah, where she allegedly endured physical abuse under the guise of conversion therapy. Lyn Duff's petition to leave her mother was granted without court opinion.
https://en.wikipedia.org/wiki?curid=26410
Ring homomorphism In ring theory, a branch of abstract algebra, a ring homomorphism is a structure-preserving function between two rings. More explicitly, if "R" and "S" are rings, then a ring homomorphism is a function such that "f" is Additive inverses and the additive identity are part of the structure too, but it is not necessary to require explicitly that they too are respected, because these conditions are consequences of the three conditions above. On the other hand, neglecting to include the condition "f"(1"R") = 1"S" would cause several of the properties below to fail. If in addition "f" is a bijection, then its inverse "f"−1 is also a ring homomorphism. In this case, "f" is called a ring isomorphism, and the rings "R" and "S" are called isomorphic. From the standpoint of ring theory, isomorphic rings cannot be distinguished. If "R" and "S" are rngs (also known as "pseudo-rings", or "non-unital rings"), then the natural notion is that of a rng homomorphism, defined as above except without the third condition "f"(1"R") = 1"S". It is possible to have a rng homomorphism between (unital) rings that is not a ring homomorphism. The composition of two ring homomorphisms is a ring homomorphism. It follows that the class of all rings forms a category with ring homomorphisms as the morphisms (cf. the category of rings). In particular, one obtains the notions of ring endomorphism, ring isomorphism, and ring automorphism. Let be a ring homomorphism. Then, directly from these definitions, one can deduce: Moreover, Injective ring homomorphisms are identical to monomorphisms in the category of rings: If is a monomorphism that is not injective, then it sends some "r"1 and "r"2 to the same element of "S". Consider the two maps "g"1 and "g"2 from Z["x"] to "R" that map "x" to "r"1 and "r"2, respectively; and are identical, but since "f" is a monomorphism this is impossible. However, surjective ring homomorphisms are vastly different from epimorphisms in the category of rings. For example, the inclusion is a ring epimorphism, but not a surjection. However, they are exactly the same as the strong epimorphisms.
https://en.wikipedia.org/wiki?curid=26411
Real Madrid CF Real Madrid Club de Fútbol (, meaning "Royal Madrid Football Club"), commonly referred to as Real Madrid, is a Spanish professional football club based in Madrid. Founded on 6 March 1902 as Madrid Football Club, the club has traditionally worn a white home kit since inception. The word "real" is Spanish for "royal" and was bestowed to the club by King Alfonso XIII in 1920 together with the royal crown in the emblem. The team has played its home matches in the 81,044-capacity Santiago Bernabéu Stadium in downtown Madrid since 1947. Unlike most European sporting entities, Real Madrid's members ("socios") have owned and operated the club throughout its history. The club was estimated to be worth €3.8 billion ($4.2 billion) in 2019, and it was the second highest-earning football club in the world, with an annual revenue of €757,3 million in 2019. The club is one of the most widely supported teams in the world. Real Madrid is one of three founding members of La Liga that have never been relegated from the top division since its inception in 1929, along with Athletic Bilbao and Barcelona. The club holds many long-standing rivalries, most notably "El Clásico" with Barcelona and "El Derbi" with Atlético Madrid. Real Madrid established itself as a major force in both Spanish and European football during the 1950s, winning five consecutive European Cups and reaching the final seven times. This success was replicated in the league, which the club won five times in the space of seven years. This team, which consisted of players such as Alfredo Di Stéfano, Ferenc Puskás, Francisco Gento, and Raymond Kopa, is considered by some in the sport to be the greatest team of all time. In domestic football, the club has won 65 trophies; a record 33 La Liga titles, 19 Copa del Rey, 11 Supercopa de España, a Copa Eva Duarte, and a Copa de la Liga. In European and worldwide competitions, the club has won a record 26 trophies; a record 13 European Cup/UEFA Champions League titles, two UEFA Cups and four UEFA Super Cups. In international football, they have achieved a record seven club world championships. Real Madrid was recognised as the FIFA Club of the 20th Century on 11 December 2000, and received the FIFA Centennial Order of Merit on 20 May 2004. The club was also awarded Best European Club of the 20th Century by the IFFHS on 11 May 2010. In June 2017, the team succeeded in becoming the first club to win back to back Champions Leagues, then made it three in a row in May 2018, extending their lead atop the UEFA club rankings. Real Madrid's origins go back to when football was introduced to Madrid by the academics and students of the "Institución Libre de Enseñanza", which included several Cambridge and Oxford University graduates. They founded "(Sociedad) Sky Football" in 1897, commonly known as "La Sociedad" (The Society) as it was the only one based in Madrid, playing on Sunday mornings at Moncloa. In 1900, conflict between members caused some of them to leave and create a new club, "Nueva Sociedad de Football" (New Society of Football), to distinguish themselves from "Sky Football". Among the dissenters were Julián Palacios, recognized as the first Real Madrid president, Juan Padrós and Carlos Padrós, the latter two being brothers and future presidents of Real Madrid. In 1901 this new club was renamed as Madrid Football Club. Later, following a restructuring in 1902, "Sky" was renamed as "New Foot-Ball Club". On 6 March 1902, after a new Board presided by Juan Padrós had been elected, Madrid Football Club was officially founded. Three years after its foundation, in 1905, "Madrid FC" won its first title after defeating Athletic Bilbao in the Spanish Cup final. The club became one of the founding sides of the Royal Spanish Football Federation on 4 January 1909, when club president Adolfo Meléndez signed the foundation agreement of the Spanish FA. After moving between grounds the team moved to the "Campo de O'Donnell" in 1912. In 1920, the club's name was changed to Real Madrid after King Alfonso XIII granted the title of Real (Royal) to the club. In 1929, the first Spanish football league was founded. Real Madrid led the first league season until the last match, a loss to Athletic Bilbao, meant they finished runners-up to Barcelona. Real Madrid won its first League title in the 1931–32 season and retained the title the following year, becoming the first team to win the championship twice. On 14 April 1931, the arrival of the Second Spanish Republic caused the club to lose the title Real and went back to being named Madrid Football Club. Football continued during the Second World War, and on 13 June 1943 Madrid beat Barcelona 11–1 in the second leg of a semi-final of the Copa del Generalísimo, the Copa del Rey having been renamed in honour of General Franco. It has been suggested Barcelona players were intimidated by police, including by the director of state security who "allegedly told the team that some of them were only playing because of the regime's generosity in permitting them to remain in the country." The Barcelona chairman, Enrique Piñeyro, was assaulted by Madrid fans. However, none of these allegations have been proven and FIFA and UEFA still consider the result as legitimate. According to Spanish journalist and writer, Juan Carlos Pasamontes, Barcelona player Josep Valle denied that the Spanish security forces came before the match. Instead, at the end of the first half, Barcelona coach Juan José Nogués and all of his players were angry with the hard-style of play Real Madrid was using and with the aggressiveness of the home crowd. When they refused to take the field, the Superior Chief of Police of Madrid appeared, identified himself, and ordered the team to take the field. Santiago Bernabéu became president of Real Madrid in 1945. Under his presidency, the club, its stadium Santiago Bernabéu and its training facilities Ciudad Deportiva were rebuilt after the Spanish Civil War damages. Additionally, during the 1950s former "Real Madrid Amateurs" player Miguel Malbo founded Real Madrid's youth academy, or ""cantera"," known today as La Fábrica. Beginning in 1953, he embarked upon a strategy of signing world-class players from abroad, the most prominent being Alfredo Di Stéfano. In 1955, acting upon the idea proposed by Gabriel Hanot, a French sports journalist and editor of "L'Équipe", Bernabéu, Bedrignan and Gusztáv Sebes created a tournament for the champions teams around Europe, under invitation, that would eventually become what today is known as the UEFA Champions League. It was under Bernabéu's guidance that Real Madrid established itself as a major force in both Spanish and European football. The club won the European Cup five times in a row between 1956 and 1960, which included the 7–3 Hampden Park final against Eintracht Frankfurt in 1960. After these five consecutive successes, Real was permanently awarded the original cup and earning the right to wear the UEFA badge of honour. The club won the European Cup for a sixth time in 1966 defeating Partizan Belgrade 2–1 in the final with a team composed entirely of same nationality players, a first in the competition. This team became known as the Yé-yé. The name "Yé-yé" came from the "Yeah, yeah, yeah" chorus in The Beatles' song "She Loves You" after four members of the team posed for "Marca" and impersonated the Beatles. The Yé-yé generation was also European Cup runner-up in 1962 and 1964. In the 1970s, Real Madrid won five league championships and three Spanish Cups. The club played its first UEFA Cup Winners' Cup final in 1971 and lost to English side Chelsea 2–1. On 2 July 1978, club president Santiago Bernabéu died while the World Cup was being played in Argentina. FIFA decreed three days of mourning to honour him during the tournament. The following year, the club organized the first edition of the Trofeo Santiago Bernabéu in memory of its former president. By the early 1980s, Real Madrid had lost its grasp on the Liga title until a new cohort of home-grown stars brought domestic success back to the club. Spanish sport journalist Julio César Iglesias gave to this generation the name "La Quinta del Buitre" ("Vulture's Cohort"), which was derived from the nickname given to one of its members, Emilio Butragueño. The other four members were Manuel Sanchís, Martín Vázquez, Míchel and Miguel Pardeza; all five footballers were graduates of Real Madrid's youth academy. With "La Quinta del Buitre" (reduced to four members when Pardeza left for Zaragoza in 1986) and notable players like goalkeeper Francisco Buyo, right-back Miguel Porlán "Chendo" and Mexican striker Hugo Sánchez, Real Madrid had one of the best teams in Spain and Europe during the second half of the 1980s, winning two UEFA Cups, five Spanish championships in a row, one Spanish cup and three Spanish Super Cups. In the early 1990s, "La Quinta del Buitre" split up after Martín Vázquez, Emilio Butragueño and Míchel left the club. In 1996, President Lorenzo Sanz appointed Fabio Capello as coach. Although his tenure lasted only one season, Real Madrid were proclaimed league champions, and players like Predrag Mijatović, Davor Šuker, Clarence Seedorf, Roberto Carlos and keeper Bodo Illgner, arrived at the club to strengthen a squad that already boasted the likes of Raúl, Fernando Hierro and Fernando Redondo. As a result, Real Madrid (with the addition of Fernando Morientes in 1997) finally ended its 32-year wait for its seventh European Cup: in 1998, under manager Jupp Heynckes, they defeated Juventus 1–0 in the final with a goal from Mijatović. In November 1999 Vicente del Bosque took over as coach. For the last season of the century, 1999–2000, the squad was still led by the older veterans such as Fernando Hierro, Fernando Redondo, Roberto Carlos and Raúl González. Real added the budding young talents of Fernando Morientes, Guti and Iker Casillas, supported by the arrival of Steve McManaman and Nicolas Anelka from the English Premier League, alongside local talents Míchel Salgado and Iván Helguera. In Del Bosque's first season in charge Real won the Champions League for the eighth time, following a 3–0 victory over Valencia in the final, with goals from Morientes, McManaman and Raúl. This victory marked the beginning of a successful period in Real Madrid's history. In July 2000, Florentino Pérez was elected club president. He vowed in his campaign to erase the club's €270 million debt and modernize the club's facilities. However, the primary electoral promise that propelled Pérez to victory was the signing of Luís Figo from arch-rivals Barcelona. The following year, the club had its training ground rezoned and used the money to begin assembling the "Galácticos" team by signing a global star every summer, which included Zinedine Zidane, Ronaldo, Luís Figo, David Beckham and Fabio Cannavaro. It is debatable whether the gamble paid off, as despite winning the UEFA Champions League and an Intercontinental Cup in 2002, followed by La Liga in 2003, the club failed to win a major trophy for the next three seasons. The few days after the capturing of the 2003 Liga title were surrounded with controversy. The first controversial decision came when Pérez sacked winning coach Vicente del Bosque. Over a dozen players left the club, including Madrid captain Fernando Hierro, while defensive midfielder Claude Makelele refused to take part in training in protest at being one of the lowest-paid players at the club and subsequently moved to Chelsea. "That's a lot [of players leaving] when the normal rule is: never change a winning team," stated Zidane. Real Madrid, with newly appointed coach Carlos Queiroz, started their domestic league slowly after a hard win over Real Betis. The 2005–06 season began with the promise of several new signings: Júlio Baptista (€24 million), Robinho (€30 million) and Sergio Ramos (€27 million). However, Real Madrid suffered from some poor results, including a 0–3 loss at the hands of Barcelona at the Santiago Bernabéu in November 2005. Madrid's coach Wanderley Luxemburgo was sacked the following month and his replacement was Juan Ramón López Caro. A brief return to form came to an abrupt halt after losing the first leg of the Copa del Rey quarterfinal, 6–1 to Real Zaragoza. Shortly after, Real Madrid were eliminated from the Champions League for a fourth successive year, this time at the hands of Arsenal. On 27 February 2006, Florentino Pérez resigned. Ramón Calderón was elected as club president on 2 July 2006 and subsequently appointed Fabio Capello as the new coach and Predrag Mijatović as the new sporting director. Real Madrid won the Liga title in 2007 for the first time in four years, but Capello was nonetheless sacked at the end of the campaign. The title was won on 17 June, where Real faced Mallorca at the Bernabéu while Barcelona and Sevilla, the other title challengers, faced Gimnàstic de Tarragona and Villarreal, respectively. At half-time, Real were 0–1 down, while Barcelona had surged ahead into a 0–3 lead in Tarragona. However, three goals in the last half-hour secured Madrid a 3–1 win and their first league title since 2003. On 1 June 2009, Florentino Pérez regained Real Madrid's presidency. Pérez continued with the "Galácticos" policy pursued in his first term, buying Kaká from Milan for a record-breaking (in pounds sterling) sum of £56 million, and then breaking the record again by purchasing Cristiano Ronaldo from Manchester United for £80 million. José Mourinho took over as manager in May 2010. In April 2011, a rare occurrence happened when, for the first time ever, four "Clásicos" were to be played in a span of just 18 days. The first fixture was for the Liga campaign on 17 April (which ended 1–1 with penalty goals for both sides), the Copa del Rey final (which ended 1–0 to Madrid) and the controversial two-legged Champions League semifinal on 27 April and 2 May (3–1 loss on aggregate) to Barcelona. In the 2011–12 La Liga season, Real Madrid won La Liga for a record 32nd time in the league's history, also finishing the season with numerous club-level records set, including 100 points reached in a single season, a total of 121 goals scored, a goal difference of +89 and 16 away wins, with 32 wins overall. In the same season, Cristiano Ronaldo become the fastest player to reach 100 goals scored in Spanish league history. In reaching 101 goals in 92 games, Ronaldo surpassed Real Madrid legend Ferenc Puskás, who scored 100 goals in 105 matches. Ronaldo set a new club mark for individual goals scored in one year (60), and became the first player ever to score against all 19 opposition teams in a single season. Real Madrid began the 2012–13 season winning the Supercopa de España, defeating Barcelona on away goals, but finished second in the league competition. A major transfer of the season was signing of Luka Modrić from Tottenham Hotspur for a fee in the region of £33 million. After a disappointing extra time loss to Atlético Madrid in the 2013 Copa del Rey Final, Pérez announced the departure of José Mourinho at the end of the season by "mutual agreement". On 25 June 2013, Carlo Ancelotti succeeded Mourinho to become the manager of Real Madrid on a three-year deal, with Zinedine Zidane named as one of his assistants. On 1 September 2013, the long-awaited transfer of Gareth Bale from Tottenham Hotspur was announced. The transfer of the Welshman was reportedly a new world record signing, with the transfer price approximated at €100 million. In Ancelotti's first season at the club, Real Madrid won the Copa del Rey, with Bale scoring the winner in the final against Barcelona. On 24 May, Real Madrid defeated city rivals Atlético Madrid in the 2014 Champions League Final, winning their first European title since 2002, and becoming the first team to win ten European Cups/Champions League titles, an achievement known as ""La Décima"". After winning the 2014 Champions League, Real Madrid signed goalkeeper Keylor Navas, midfielder Toni Kroos and attacking midfielder James Rodríguez. The club won the 2014 UEFA Super Cup against Sevilla, the club's 79th official trophy. During the last week of the 2014 summer transfer window, Real Madrid sold two players key to the previous season's successes: Xabi Alonso to Bayern Munich and Ángel Di María to Manchester United. This decision by the club was surrounded by controversy, with Cristiano Ronaldo stating, "If I was in charge, maybe I would have done things differently," while Carlo Ancelotti admitted, "We must start again from zero." After a slow start to the 2014–15 La Liga season, Real Madrid went on a record-breaking 22-match winning streak, which included wins against Barcelona and Liverpool, surpassing the previous Spanish record of 18 successive wins set by Frank Rijkaard's "Barça" in the 2005–06 season. The streak came to an end in their opening match of 2015 with a loss to Valencia, leaving the club two short of equalling the world record of 24 consecutive wins. The club failed to retain the Champions League (losing to Juventus in the semi-finals) and the Copa del Rey, and also failed to land the league title (finishing two points and a place behind champions Barcelona), shortcomings that all preceded Ancelotti's dismissal on 25 May 2015. On 3 June 2015, Rafael Benítez was confirmed as the new Real Madrid manager, signing a three-year contract. Real Madrid remained unbeaten in the league until a 3–2 loss at Sevilla on the matchday 11. This was followed by a 0–4 home loss in the first "Clásico" of the season against Barcelona. In the Copa del Rey Round of 32, Real fielded an ineligible player in Denis Cheryshev in a 1–3 first leg win away against Cádiz, resulting in the second leg being cancelled and Real being disqualified. Benítez was relieved of his duties on 4 January 2016 following allegations of unpopularity with supporters, displeasure with players and a failure to get good results against top teams. On 4 January 2016, Benítez's departure was announced along with the promotion of Zinedine Zidane to his first head coaching role. Under Zidane, Real ended up finishing in second place, just one point behind champions Barcelona, in the 2015–16 La Liga. On 28 May, Real Madrid won their 11th Champions League title, extending their record for most successes in the competition, with the achievement being termed ""La Undécima"". Real Madrid began their 2016–17 campaign, which was to be Zidane's first full season in charge of the club, with victory in the 2016 UEFA Super Cup. On 10 December 2016, Madrid won their 35th-straight match without a loss, which set a new club record. On 18 December 2016, Madrid defeated Japanese club Kashima Antlers 4–2 in the final of the 2016 FIFA Club World Cup. With a 3–3 draw at Sevilla on 12 January 2017, Madrid's unbeaten run extended to 40 matches, breaking Barcelona's Spanish record of 39 matches unbeaten in all competitions from the previous season. Their unbeaten streak ended after a 1–2 away loss against Sevilla in La Liga three days later. In May that year, Madrid won the 2016–17 La Liga for a record 33rd time, their first title in five years. On 3 June, the club's Champions League Final win against Juventus resulted in Real Madrid being the first team to successfully defend their title in the UEFA Champions League era, and the first to win consecutive titles in the competition since Milan in 1989 and 1990, when the tournament was known as the European Cup. Real Madrid's title was its 12th, extending its record, and its third in four years. The achievement is also known as ""La Duodécima"". The 2016–17 season was the greatest campaign in terms of trophies won in the history of Real Madrid. Real Madrid won the 2017 UEFA Super Cup 2–1 against Manchester United. Five days later, Real Madrid beat Barcelona at the Camp Nou in the first leg of the 2017 Supercopa de España, before winning the second leg 2–0, ending a 24 consecutive match scoring record of Barcelona in "El Clásico" matches, and with a 5–1 aggregate score. On 16 December 2017, Real beat Brazilian club Grêmio 1–0 in the Final of 2017 FIFA Club World Cup and became the first club to retain the trophy. Real Madrid also won their third successive UEFA Champions League in 2018, becoming the first club to win three straight UEFA Champions League titles since the tournament's inception, as well as the first team to win three straight titles in European competition since Bayern Munich in the 1970s. On 31 May, only five days after winning the final, Zidane announced his resignation as Real Madrid manager, citing the club's "need for change" as his rationale for departing. On 12 June, Real Madrid named Julen Lopetegui, the head coach of the Spanish national team, as their new manager. It was announced that he would officially become manager after the 2018 FIFA World Cup. However, the Spanish national team sacked Lopetegui a day prior to the tournament, stating that he negotiated terms with the club without informing them. The club then began aggressively re-shaping the squad in the summer of 2018, which included the sale of Cristiano Ronaldo to Juventus for a reported €100 million. After a string of poor performances and losses from the team, Lopetegui was dismissed and replaced by then Castilla coach, Santiago Solari. On 22 December 2018, Real Madrid beat Al Ain by a 4–1 margin in the final of 2018 FIFA Club World Cup. With their win, Real Madrid became the outright record winners of the Club World Cup with four titles. They are considered to have been the world champions for grand total of seven time because FIFA recognises the Intercontinental Cup as the predecessor of FIFA Club World Cup. They also extended the record for most consecutive titles with their third in a row. On 11 March 2019, Real Madrid reinstated Zidane as the head coach of the club. On 12 January 2020, Real Madrid beat Atlético Madrid in a penalty shootout to win their eleventh Supercopa de España title. The first crest had a simple design consisting of a decorative interlacing of the three initials of the club, "MCF" for Madrid Club de Fútbol, in dark blue on a white shirt. The first change in the crest occurred in 1908 when the letters adopted a more streamlined form and appeared inside a circle. The next change in the configuration of the crest did not occur until the presidency of Pedro Parages in 1920. At that time, King Alfonso XIII granted the club his royal patronage which came in the form of the title "Real Madrid," meaning "Royal." Thus, Alfonso's crown was added to the crest and the club styled itself "Real Madrid Club de Fútbol". With the dissolution of the monarchy in 1931, all the royal symbols (the crown on the crest and the title of Real) were eliminated. The crown was replaced by the dark mulberry band of the Region of Castile. In 1941, two years after the end of the Civil War, the crest's "Real Corona", or "Royal Crown", was restored while the mulberry stripe of Castile was retained as well. In addition, the whole crest was made full color, with gold being the most prominent, and the club was again called Real Madrid Club de Fútbol. The most recent modification to the crest occurred in 2001 when the club wanted to better situate itself for the 21st century and further standardize its crest. One of the modifications made was changing the mulberry stripe to a more bluish shade. Real Madrid has maintained the white shirt for its home kit throughout the history of the club. There was, however, one season that the shirt and shorts were not both white. It was an initiative undertaken by Escobal and Quesada in 1925; the two were traveling through England when they noticed the kit worn by London-based team Corinthian F.C., one of the most famous teams at the time known for its elegance and sportsmanship. It was decided that Real Madrid would wear black shorts in an attempt to replicate the English team, but the initiative lasted just one year. After being eliminated from the cup by Barcelona with a 1–5 defeat in Madrid and a 2–0 defeat in Catalonia, President Parages decided to return to an all-white kit, claiming that the other kit brought bad luck. By the early 1940s, the manager changed the kit again by adding buttons to the shirt and the club's crest on the left breast, which has remained ever since. On 23 November 1947, in a game against Atlético Madrid at the Metropolitano Stadium, Real Madrid became the first Spanish team to wear numbered shirts. English club Leeds United permanently switched their blue shirt for a white one in the 1960s, to emulate the dominant Real Madrid of the era. Real's traditional away colours are all blue or all purple. Since the advent of the replica kit market, the club has also released various other one colour designs, including red, green, orange and black. The club's kit is manufactured by Adidas, whose contract extends from 1998. Real Madrid's first shirt sponsor, Zanussi, agreed for the 1982–83, 1983–84 and 1984–85 seasons. Following that, the club was sponsored by Parmalat and Otaysa before a long-term deal was signed with Teka in 1992. In 2001, Real Madrid ended their contract with Teka and for one season and used the Realmadrid.com logo to promote the club's website. Then, in 2002, a deal was signed with Siemens Mobile and in 2006, the BenQ Siemens logo appeared on the club's shirt. Real Madrid's shirt sponsor from 2007 until 2013 was bwin.com following the economic problems of BenQ Siemens. Fly Emirates became their shirt sponsor in 2013, and in 2017 the club renewed their sponsorship with the airliner, signing a deal until 2022 worth €70 million per year. In 2015, Madrid signed a new 10-year contract with Adidas believed to be worth a total of £850 million (€1 billion), earning £59 million (€64 million) per season including contract penalty or contract termination clause anytime if Real Madrid failed to qualify for any European competitions even relegation. "Note: Early contract termination clauses are activated at any time depending on team's on-pitch performance." After moving between grounds, the team moved to the Campo de O'Donnell in 1912, which remained its home ground for 11 years. After this period, the club moved for one year to the Campo de Ciudad Lineal, a small ground with a capacity of 8,000 spectators. After that, Real Madrid moved its home matches to Estadio Chamartín, which was inaugurated on 17 May 1923 with a match against Newcastle United. In this stadium, which hosted 22,500 spectators, Real Madrid celebrated its first Spanish league title. After some successes, the 1943 elected president Santiago Bernabéu decided that the Estadio Chamartín was not big enough for the ambitions of the club, and thus a new stadium was built and was inaugurated on 14 December 1947. This was the Santiago Bernabéu Stadium as it is known today, although it did not acquire the present name until 1955. The first match at the Bernabéu was played between Real Madrid and the Portuguese club Belenenses and won by "Los Blancos", 3–1, the first goal being scored by Sabino Barinaga. The capacity has changed frequently, peaking at 120,000 after a 1953 expansion. Since then, there have been a number of reductions due to modernizations (the last standing areas were removed in 1998–99 in response to UEFA regulations which forbids standing at matches in UEFA competitions), countered to some extent by expansions. The latest capacity is 81,044 spectators. A plan to add a retractable roof has been announced. Real Madrid has the fourth-highest of the average attendances of European football clubs, behind only Borussia Dortmund, Barcelona and Manchester United. The Bernabéu has hosted the 1964 UEFA European Championship final, the 1982 FIFA World Cup final, the 1957, 1969 and 1980 European Cup finals and the 2010 UEFA Champions League Final. The stadium has its own Madrid Metro station along the 10 line called "Santiago Bernabéu". On 14 November 2007, the Bernabéu was upgraded to Elite Football Stadium status by UEFA. On 9 May 2006, the Alfredo Di Stéfano Stadium was inaugurated in the Real Madrid City, where Real Madrid usually trains. The inaugural match was played between Real Madrid and Stade de Reims, a rematch of the 1956 European Cup final. Real Madrid won the match 6–1, with goals from Sergio Ramos, Antonio Cassano (2), Roberto Soldado (2) and José Manuel Jurado. The venue is now part of the Ciudad Real Madrid, the club's training facility located outside Madrid, in Valdebebas. The stadium holds 5,000 people and is Real Madrid Castilla's home ground. It is named after former Real legend Alfredo Di Stéfano. Since 14 June 2020, the stadium is used by the first team to play their remaining 2019–20 La Liga home games due to remodelling works at the Santiago Bernabéu stadium. Raúl holds the record for most Real Madrid appearances, having played 741 first-team matches from 1994 to 2010. Iker Casillas comes second with 725 appearances, followed by Manuel Sanchis, Jr., having played 710 times. The record for a goalkeeper is held by Iker Casillas, with 725 appearances. With 170 caps (167 while at the club), Sergio Ramos is Real's most capped international player. Cristiano Ronaldo is Real Madrid's all-time top goalscorer, with 450 goals. Six other players have also scored over 200 goals for Real: Alfredo Di Stéfano (1953–64), Santillana (1971–88), Ferenc Puskás (1958–66), Hugo Sánchez (1985–92), Karim Benzema (2009–current) and the previous goalscoring record-holder Raúl (1994–2010). Cristiano Ronaldo also holds the record for the most league goals scored in one season (48 in 2014–15), alongside being Real's top goalscorer of all time in La Liga history with 311 goals. Di Stéfano's 49 goals in 58 matches was for decades the all-time highest tally in the European Cup, until it was surpassed by Raúl in 2005, and is now held by Cristiano Ronaldo with 105 goals. The fastest goal in the history of the club (12 seconds) was scored by the Brazilian Ronaldo on 3 December 2003 during a league match against Atlético Madrid. Officially, the highest home attendance figure for a Real Madrid match is 83,329, which was for a Copa del Rey match in 2006. The current official capacity of the Santiago Bernabéu is 81,044. The club's average attendance in the 2007–08 season was 76,234, the highest in European Leagues. Real has also set records in Spanish football, most notably the most domestic titles (33 as of 2016–17) and the most seasons won in a row (five, during 1960–65 and 1985–90). With 121 matches (from 17 February 1957 to 7 March 1965), the club holds the record for longest unbeaten run at home in La Liga. The club also holds the record for winning the European Cup/UEFA Champions League thirteen times and for the most semi-final appearances (28). As of December 2019, Cristiano Ronaldo is the all-time top scorer in the UEFA Champions League, with 128 (129 including qualifiers) goals in total, 105 while playing for Real Madrid. The team has the record number of consecutive participations in the European Cup (before it became the Champions League) with 15, from 1955–56 to 1969–70. Among the club's on-field records is a 22-game winning streak in all competitions during the 2014–15 season, a Spanish record and fourth worldwide. The same season the team tied the win-streak for games in the Champions League, with ten. In September 2017, the club equalled the record of the Brazilian club Santos, starring Pelé, by scoring in their 73rd consecutive game. In June 2009, the club broke its own record for the highest transfer fee ever paid in the history of football by agreeing to pay Manchester United €94 million (£80 million) for the services of Cristiano Ronaldo. The fee of €77.5 million (100 billion lire) for Zinedine Zidane's transfer from Juventus to Real Madrid in 2001 was the previous highest transfer fee ever paid. This record (in pounds sterling) had been broken previously in June 2009, for a few days, when Real Madrid agreed to buy Kaká from Milan for €67 million (£65 million). The transfer of Tottenham Hotspur's Gareth Bale in 2013 was reportedly a new world record signing, with the transfer fee reported at around €100 million. In January 2016, documents pertaining to Bale's transfer were leaked which confirmed a world record transfer fee of €100,759,418. Real Madrid equalled their record signing in 2019, when the club reportedly signed Eden Hazard from Chelsea for €100 million. The club's sale record came on 10 July 2018, when Juventus signed Cristiano Ronaldo for €100 million. During most home matches the majority of the seats in the stadium are occupied by season-ticket holders, of which the figure is capped at 65,000. To become a season ticket holder one must first be a "socio", or club member. In addition to members, the club has more than 1,800 "peñas" (official, club-affiliated supporters' groups) in Spain and around the world. Real Madrid has the second highest average all-time attendance in Spanish football and regularly attracts over 74,000 fans to the Bernabéu. One of the best supported teams globally, Real Madrid was the first sports team (and first brand) to reach 100 million fans on Facebook in April 2017. Real Madrid's hardcore supporters are the so-called "Ultras Sur" supporters, or simply Ultras. They are known for their extreme right-wing politics, akin to Barcelona's hardcore supporters group Boixos Nois. The Ultras Surs have developed an alliance with other right wing groups, most notably Lazio "Irriducibili" fans, and have also developed an alliance with left-wing groups. On several occasions, they have racially abused opposing players and have been investigated by UEFA for doing so. Florentino Pérez took it upon himself to ban the Ultras from the Bernabéu and assign their seats to the general public. This decision was controversial with some of the Bernabéu faithful, however, as the lively atmosphere of games would suffer as a result. The Ultras have since held protests outside the Bernabéu and have demanded to be reinstated and allowed to enter the grounds. Questioned over Pope Francis' adherence to 2014 FIFA Club World Cup Final opponents San Lorenzo, Madrid captain Sergio Ramos stated, "In the semi-finals we noticed the love from supporters in Marrakesh and it seemed like we were playing at home. That sums up the greatness of this team. Madrid is God's team and the team of the world". There is often a fierce rivalry between the two strongest teams in a national league, and this is particularly the case in La Liga, where the game between Real Madrid and Barcelona is known as "The Classic" ("El Clásico"). From the start of national competitions, the clubs were seen as representatives of two rival regions in Spain, Catalonia and Castile, as well as of the two cities. The rivalry reflects what many regard as the political and cultural tensions felt between Catalans and the Castilians, seen by one author as a re-enactment of the Spanish Civil War. Over the years, the record from Real Madrid and Barcelona is 95 victories for Madrid, 96 victories for Barcelona, and 51 draws. During the dictatorships of Primo de Rivera and especially of Francisco Franco (1939–1975), all regional cultures were suppressed. All of the languages spoken in Spanish territory, except Spanish (Castilian) itself, were officially banned. Symbolising the Catalan people's desire for freedom, Barcelona became "More than a club" (""Més que un club"") for the Catalans. According to Manuel Vázquez Montalbán, the best way for the Catalans to demonstrate their identity was by joining Barcelona. It was less risky than joining a clandestine anti-Franco movement, and allowed them to express their dissidence. On the other hand, Real Madrid was widely seen as the embodiment of the sovereign oppressive centralism and the fascist regime at management level and beyond– Santiago Bernabéu, the former club president for whom Real Madrid's stadium is named, fought on the Nationalist side during the Spanish Civil War. During the war, however, members of both clubs, such as Josep Sunyol and Rafael Sánchez Guerra, suffered at the hands of Francoists. During the 1950s, the rivalry was exacerbated further when there was a controversy surrounding the transfer of Alfredo Di Stéfano, who eventually played for Real Madrid and was key to their subsequent success. The 1960s saw the rivalry reach the European stage when they met twice in a controversial knock-out round of the European Cup, with Madrid receiving unfavourable treatment from the referee. In 2002, the European encounter between the clubs was dubbed the "Match of The Century" by Spanish media, and Madrid's win was watched by more than 500 million people. The fixture has featured memorable goal celebrations from both teams, often involving mocking the opposition. A notable example of this occurred in October 1999 when Raúl silenced the hostile crowd of 100,000 Barcelona fans when he scored and then celebrated his goal by putting a finger to his lips as though to tell them to be quiet. The club's nearest neighbour is Atlético Madrid, a rivalry being shared between fans of both football teams. Although Atlético was founded by three Basque students in 1903, it was joined in 1904 by dissident members of "Madrid FC". Tensions escalated further after Atlético were merged with the football team of the Spanish airforce (and thus renamed Atlético Aviación), and in the 1940s, Atlético was perceived as the preferred team of Franco's regime before he revelled in Real's European success in the 1950s. Furthermore, Real supporters initially came from the middle and upper classes while the Atlético supporters were drawn from the working class. Today, however, these distinctions are largely blurred. They met for the first time on 21 February 1929 in matchday three of the first League Championship at the former Chamartín. It was the first official derby of the new tournament, and Real won 2–1. The rivalry first gained international attention in 1959 during the European Cup when the two clubs met in the semi-final. Real won the first leg 2–1 at the Bernabéu while Atlético won 1–0 at the "Metropolitano". The tie went to a replay, which Real won 2–1. Atlético, however, gained some revenge when, led by former Real Madrid coach José Villalonga, it defeated its city rivals in two successive "Copa del Generalísimo" finals in 1960 and 1961. Between 1961 and 1989, when Real dominated La Liga, only Atlético offered it any serious challenge, winning Liga titles in 1966, 1970, 1973 and 1977. In 1965, Atlético became the first team to beat Real at the Bernabéu in eight years. Real Madrid's record against Atlético in more recent times is very favorable. A high point coming in the 2002–03 season, when Real clinched the La Liga title after a 0–4 victory at Atlético at the Vicente Calderón Stadium. Atlético's first win over its city rivals since 1999 came with the Copa del Rey win in May 2013. In 2013–14, Real and Atlético were finalists of UEFA Champions League, the first final which hosted two clubs from same city. Real Madrid triumphed with 4–1 in extra time. On 7 February 2015, Real suffered their first defeat in 14 years at the Vicente Calderón, a 4–0 loss. On 28 May 2016, Real and Atlético met again for the Champions League title in Milan, which resulted in a win for Real after a penalty shootout. A further minor rivalry exists between Real Madrid and Athletic Bilbao. This is known as El Viejo Clásico ("the old classic"), so named as the two clubs were dominant in the first half of the 20th century, meeting in nine Copa del Rey finals including the first in 1903. Until 10 December 2011, this fixture was the most played in the history of Spanish football, when it was surpassed by "El Clásico". Athletic Bilbao, who operate a policy of only using local players, have long since ceased to be a competitive rival to clubs such as Real Madrid who scour the globe for the best talent; the "Lions" have collected no major trophies since 1984 and won only two of the 26 matches between the teams from 2005–06 to 2016–17. However, the matches remain keenly fought due to their historical and cultural significance, with some parallels to the political aspect of the Barcelona/Catalonia rivalry as Athletic are the largest club in the Basque region. Real Madrid and Germany's Bayern Munich are two of the most successful clubs in the UEFA Champions League/European Cup competition, Real winning thirteen times and Bayern winning five times. Although they have never met in a final, Real Madrid versus Bayern is the match that has historically been played most often in the Champions League/European Cup with 26 matches (12 wins for Madrid, 11 wins for Bayern, with 3 draws), Real Madrid supporters often refer to Bayern as the ""Bestia negra"" ("Black Beast"). During the 2010s, the two teams met in the 2011–12 Champions League semi-finals, which ended 3–3 on aggregate (Bayern won 3–1 on penalties after extra time, but lost the final at their own stadium), and then at the same stage in the 2013–14 edition with Real Madrid winning 5–0 on aggregate on their way to winning the competition. They were also drawn together in the 2016–17 quarter-finals; Real Madrid won 6–3 on aggregate and subsequently lifted the trophy. The following year, they met in the semi-finals, with Real Madrid again progressing 4–3. Until the 2018–19 season when they were twice defeated in Madrid by three-goal margins, Real's biggest loss at home in the Champions League had been at the hands of Bayern on 29 February 2000, 2–4. Arjen Robben, Xabi Alonso, Toni Kroos, James Rodríguez and Álvaro Odriozola are among the players to appear for both clubs in the early 21st century. Another match that is often played in the European Cup/Champions League is Real Madrid vs Juventus, the most decorated Italian club. They have played each other in 21 matches and have an almost perfectly balanced record (9 wins for Juventus, 10 wins for Real Madrid, 2 draws), as well as nearly the same goal difference (Madrid ahead 26 to 25). Their first meeting was in the 1961–62 European Cup, which Real Madrid won 3–1 in a replay held in Paris. At the quarter-final stage in 1995–96, Juventus prevailed 2–1 and went on to lift the trophy. In the 1998 UEFA Champions League Final between the teams in Amsterdam, Real Madrid won 1–0. They met again in the 2002–03 UEFA Champions League semi-finals, when both clubs were in their respective 'golden eras'; Juventus won 4–3 on aggregate. By that time, star midfielder Zinedine Zidane, who played for the "Bianconeri" in the 1998 final, had moved from Turin to Madrid in a world record €77 million deal. In the 2014–15 UEFA Champions League semi-finals, former Real Madrid player Álvaro Morata scored one goal in each leg to take Juventus to the final, winning 3–2 on aggregate. They faced off again in the 2017 UEFA Champions League Final in Cardiff, which Real Madrid won 4–1. Portuguese player Cristiano Ronaldo scored two goals in the match, and was named man of the match. The latest Champions League meeting was in the 2017–18 quarter-finals, which Real Madrid won 4–3 on aggregate; the tie ended in dramatic and controversial fashion, with a debatable penalty awarded to Real Madrid in the last minute of the second leg after Juventus built a 3–0 lead at the Bernabeu to pull level in the tie following a defeat at their Juventus Stadium by the same scoreline. Cristiano Ronaldo scored three goals over the two matches including the decisive penalty and a spectacular overhead kick, and having won the Champions League with Madrid for a fourth time, he transferred to Juventus a few months later for a €100 million transfer fee. In previous decades, Luis Del Sol, Michael Laudrup, Robert Jarni, Fabio Cannavaro and Emerson also featured for both clubs. It was under Florentino Pérez's first presidency (2000–2006) that Real Madrid started its ambition of becoming the world's richest professional football club. The club ceded part of its training grounds to the city of Madrid in 2001, and sold the rest to four corporations: Repsol YPF, Mutua Automovilística de Madrid, Sacyr Vallehermoso and OHL. The sale eradicated the club's debts, paving the way for it to buy the world's most expensive players, such as Zinedine Zidane, Luís Figo, Ronaldo and David Beckham. The city had previously rezoned the training grounds for development, a move which in turn increased their value, and then bought the site. The European Commission started an investigation into whether the city overpaid for the property, to be considered a form of state subsidy. The sale of the training ground for office buildings cleared Real Madrid's debts of €270 million and enabled the club to embark upon an unprecedented spending spree which brought big-name players to the club. In addition, profit from the sale was spent on a state-of-the-art training complex on the city's outskirts. Although Pérez's policy resulted in increased financial success from the exploitation of the club's high marketing potential around the world, especially in Asia, it came under increasing criticism for being too focused on marketing the Real Madrid brand and not enough on the performances of the team. By September 2007, Real Madrid was considered the most valuable football brand in Europe by BBDO. In 2008, it was ranked the second-most valuable club in football, with a value of €951 million (£640 million / $1.285 billion), only beaten by Manchester United, which was valued at €1.333 billion (£900 million). In 2010, Real Madrid had the highest turnover in football worldwide. In September 2009, Real Madrid's management announced plans to open its own dedicated theme park by 2013. A study at Harvard University concluded that Real Madrid "is one of the 20 most important brand names and the only one in which its executives, the players, are well-known. We have some spectacular figures in regard to worldwide support of the club. There are an estimated 287 million people worldwide who follow Real Madrid." In 2010, "Forbes" evaluated Real Madrid's worth to be around €992 million (US$1.323 billion), ranking them second after Manchester United, based on figures from the 2008–09 season. According to Deloitte, Real Madrid had a recorded revenue of €401 million in the same period, ranking first. Along with Barcelona, Athletic Bilbao and Osasuna, Real Madrid is organised as a registered association. This means that Real Madrid is owned by its supporters who elect the club president. The club president cannot invest his own money into the club and the club can only spend what it earns, which is mainly derived through merchandise sales, television rights and ticket sales. Unlike a limited company, it is not possible to purchase shares in the club, but only membership. The members of Real Madrid, called "socios", form an assembly of delegates which is the highest governing body of the club. As of 2010, the club has 60,000 "socios". At the end of the 2009–10 season, the club's board of directors stated that Real Madrid had a net debt of €244.6 million, €82.1 million lower than the previous fiscal year. Real Madrid announced that it had a net debt of €170 million after the 2010–11 season. From 2007 to 2011, the club made a net profit of €190 million. During the 2009–10 season, Real Madrid made €150 million through ticket sales, which was the highest in top-flight football. The club has the highest number of shirt sales a season, around 1.5 million. For the 2010–11 season its wage bill totalled €169 million, which was second-highest in Europe behind Barcelona. However, its wage bill to turnover ratio was the best in Europe at 43 percent, ahead of Manchester United and Arsenal at 46 percent and 50 percent, respectively. In 2013, "Forbes" listed the club as the world's most valuable sports team, worth $3.3 billion. It was valued at €3.47 billion ($4.1 billion) in 2018, and in the 2016–17 season it was the second highest-earning football club in the world, with an annual revenue of €674.6 million. The second highest paid sports team in the world – after Barcelona – in November 2018 the average first-team pay at Real Madrid was £8.1m ($10.6m) per year. Real Madrid was the featured club in the second installment of the "Goal!" football movie trilogy, "" (2007). The film follows former Newcastle United star Santiago Muñez as he is first scouted, and then signed by Real Madrid for the 2005–06 season. The film's creators wanted to put emphasis on the changes in Muñez's life after his move to Madrid. Production was done with the full support of UEFA, allowing the film crew to use many real life players in cameo roles. Real Madrid squad members featured in the film included Iker Casillas, Zinedine Zidane, David Beckham, Ronaldo, Roberto Carlos, Raúl, Sergio Ramos, Robinho, Michael Owen, Míchel Salgado, Júlio Baptista, Steve McManaman and Iván Helguera. Non-Real Madrid players to make cameo appearances included Ronaldinho, Thierry Henry, Lionel Messi, Samuel Eto'o, Andrés Iniesta, Pablo Aimar, Freddie Ljungberg, Cesc Fàbregas and Santiago Cañizares. In the film, both Florentino Pérez and Alfredo Di Stéfano presented the fictional player Muñez to the club after his signing. "Real, The Movie" is a 2005 part feature, part documentary film that showcases the worldwide passion for Real Madrid. Produced by the club and directed by Borja Manso, it follows five sub-stories of fans from around the world and their love for the club. Along with the fictional portion of the film, it also contains real footage of the squad, during training at Ciudad Real Madrid, matches, and interviews. Although the film mentions all of the squad, it mainly focuses on "galácticos" such as David Beckham, Zinedine Zidane, Raúl, Luís Figo, Ronaldo, Iker Casillas, and Roberto Carlos, among others. The film was originally produced in Spanish, but has been dubbed for their worldwide fanbase. The book "White Storm: 100 years of Real Madrid" by Phil Ball was the first English-language history of Real Madrid. Published in 2002, it talks about the most successful moments of the club during its first centenary, having been translated into various languages. In late 2011, Real Madrid released a digital music album, entitled "Legends", and a remix of the club's anthem, "Himno del Real Madrid," was released as the first single from the album. Real Madrid TV is an encrypted digital television channel, operated by Real Madrid and specialising in the club. The channel is available in Spanish and English. It is located at Ciudad Real Madrid in Valdebebas (Madrid), Real Madrid's training centre. "Hala Madrid" is a magazine published quarterly for the Real Madrid club members and the "Madridistas Fan Club" card holders. The phrase Hala Madrid, meaning "Forward Madrid" or "Go Madrid", is also the title of the club's official anthem, which is often sung by the Madridistas (the club's fans). The magazine includes reports on the club's matches in the previous month, as well as information about the reserve and youth teams. Features often include interviews with players, both past and present, and the club's historic matches. Real Madrid has appeared in many football-based video games, namely in the FIFA and Pro Evolution Soccer series. A Real Madrid player has appeared on the cover of both titles a combined seven times. In 2007, Spanish game publisher Virgin Play signed a deal with the club to make officially licensed Real Madrid video games. The only one released under the deal (due to Virgin Play's liquidation in September 2009) would end up being "Real Madrid: The Game", which was developed by Atomic Planet Entertainment and was published under Virgin Play's publishing division V.2 Play in May 2009 for the PlayStation 2, PlayStation Portable, Windows, Wii and Nintendo DS exclusively in European territories Virgin Play released their products in. The game featured a career mode with a mixture of role-playing and simulation as well as arcade-styled Football gameplay. Spanish teams are limited to three players without EU citizenship. The squad list includes only the principal nationality of each player; several non-European players on the squad have dual citizenship with an EU country. Also, players from the ACP countries in Africa, the Caribbean, and the Pacific that are signatories to the Cotonou Agreement are not counted against non-EU quotas due to the Kolpak ruling.
https://en.wikipedia.org/wiki?curid=26413
Rube Foster Andrew "Rube" Foster (September 17, 1879 – December 9, 1930) was an American baseball player, manager, and executive in the Negro leagues. He was elected to the Baseball Hall of Fame in 1981. Foster, considered by historians to have been perhaps the best African-American pitcher of the first decade of the 1900s, also founded and managed the Chicago American Giants, one of the most successful black baseball teams of the pre-integration era. Most notably, he organized the Negro National League, the first long-lasting professional league for African-American ballplayers, which operated from 1920 to 1931. He is known as the "father of Black Baseball." Foster adopted his longtime nickname, "Rube", as his official middle name later in life. Foster was born in Calvert, Texas on September 17, 1879. His father, also named Andrew, was a reverend and elder of the local American Methodist Episcopal Church. Foster started his professional career with the Waco Yellow Jackets, an independent black team, in 1897. Over the next few years he gradually built up a reputation among white and black fans alike, until he was signed by Frank Leland's Chicago Union Giants, a team in the top ranks of black baseball, in 1902. After a slump, he was released, and signed with a white semipro team based in Otsego, Michigan—Bardeen's Otsego Independents. According to Phil Dixon's American Baseball Chronicles: Great Teams, The 1905 Philadelphia Giants, Volume III "In completing the summer of 1902 with Otsego's multi-ethnic team—the only multi-race team with which he would ever regularly perform—Foster is reported to have pitched twelve games. He finished with a documented record of eight wins and four losses along with eighty-two documented strikeouts. Ironically, strikeout totals for five games which he appeared were not recorded. If found the totals would likely show that Foster struck out more than one-hundred batters for Otsego. In the seven games where details exist, Foster average eleven strikeouts per outing." Toward the end of the season he joined the Cuban X-Giants of Philadelphia, perhaps the best team in black baseball. The 1903 season saw Foster establish himself as the X-Giants' pitching star. In a post-season series for the eastern black championship, the X-Giants defeated Sol White's Philadelphia Giants five games to two, with Foster himself winning four games. According to various accounts, including his own, Foster acquired the nickname "Rube" after defeating star Philadelphia Athletics left-hander Rube Waddell in a postseason exhibition game played sometime between 1902 and 1905. A newspaper story in the Trenton (NJ) "Times" from July 26, 1904, contains the earliest known example of Foster being referred to as "Rube", indicating that the supposed meeting with Waddell must have taken place earlier than that. Recent research has uncovered a game played on August 2, 1903, in which Foster met and defeated Waddell while the latter was playing under an assumed name for a semi-pro team in New York City. Now a star, Foster jumped to the Philadelphia Giants for the 1904 season. Legend has it that John McGraw, manager of the New York Giants, hired Foster to teach the young Christy Mathewson the "fadeaway", or screwball, though historians have cast this story in doubt. During the 1904 season, Foster won 20 games against all competition (including two no-hitters) and lost six. In a rematch with Foster's old team, the Cuban X-Giants, he won two games and batted .400 in leading the Philadelphia Giants to the black championship. In 1905, Foster (by his own account several years later) compiled a fantastic record of 51–4, though recent research has confirmed only a 25–3 record. He led the Giants to another championship series victory, this time over the Brooklyn Royal Giants. The "Philadelphia Telegraph" wrote that "Foster has never been equalled in a pitcher's box." The following season, the Philadelphia Giants helped form the International League of Independent Professional Ball Players, composed of both all-black and all-white teams in the Philadelphia and Wilmington, Delaware, areas. In 1907, Foster's manager Sol White published his "Official Baseball Guide: History of Colored Baseball", with Foster contributing an article on "How to Pitch." However, before the season began, he and several other stars (including, most importantly, the outfielder Pete Hill) left the Philadelphia Giants for the Chicago Leland Giants, with Foster named playing manager. Under his leadership, the Lelands won 110 games (including 48 straight) and lost only ten, and took the Chicago City League pennant. The following season the Lelands tied a national championship series with the Philadelphia Giants, each team winning three games. Foster suffered a broken leg in July 1909, but rushed himself back into the lineup in time for an October exhibition series against the Chicago Cubs. Foster, pitching the second game, squandered a 5–2 lead in the ninth inning, then lost the game on a controversial play when a Cubs runner stole home while Foster was arguing with the umpire. The Lelands lost the series, three games to nothing. The Lelands also lost the unofficial western black championship to the St. Paul Colored Gophers. In 1910, Foster wrested legal control of the team from its founder, Frank Leland. He proceeded to put together the team he later considered his finest. He signed John Henry Lloyd away from the Philadelphia Giants; along with Hill, second baseman Grant Johnson, catcher Bruce Petway, and pitchers Frank Wickware and Pat Dougherty, Lloyd sparked the Lelands to a 123–6 record (with Foster himself contributing a 13–2 record on the mound). The following season, Foster established a partnership with John Schorling, the son-in-law of Chicago White Sox owner Charlie Comiskey. The White Sox had just moved into Comiskey Park, and Schorling arranged for Foster's team to use the vacated South Side Park, at 39th and Wentworth. Settling into their new home (now called Schorling's Park), the Lelands became the Chicago American Giants. For the next four seasons, the American Giants claimed the western black baseball championship, though they lost a 1913 series to the Lincoln Giants for the national championship. By 1915, Foster's first serious rival in the midwest had emerged: C. I. Taylor's Indianapolis ABCs, who claimed the western championship after defeating the American Giants four games to none in July. One of the victories was a forfeit called after a brawl between the two teams broke out. After the series, Foster and Taylor engaged in a public dispute about that game and the championship. In 1916, both teams again claimed the western title. The continued wrangling led to calls for a black baseball league to be formed, but Foster, Taylor, and the other major clubs in the midwest were unable to come to any agreement. By this time, Foster was pitching very little, compiling only a 2–2 record in 1915. His last recorded outing on the mound was in 1917; from this time he became purely a bench manager. As a manager and team owner, Foster was a disciplinarian. He asserted control over every aspect of the game, and set a high standard for personal conduct, appearance, and professionalism among his players. Given Schorling Park's huge dimensions, Foster developed a style of play that emphasized speed, bunting, place hitting, power pitching, and defense. He was also considered a great teacher, and many of his players themselves eventually became managers, including Pete Hill, Bruce Petway, Bingo DeMoss, Dave Malarcher, Sam Crawford, Poindexter Williams, and many others. In 1919, Foster helped Tenny Blount finance a new club in Detroit, the Stars. He also transferred several of his veteran players there, including Hill, who was to manage the new team, and Petway. He may have been preparing the way for the formation, the following year, of the Negro National League (NNL). In 1920, Foster, Taylor, and the owners of six other midwestern clubs met in the spring to form a professional baseball circuit for African-American teams. Foster, as president, controlled league operations, while remaining owner and manager of the American Giants. He was periodically accused of favoring his own team, especially in matters of scheduling (the Giants in the early years tended to have a disproportionate number of home games) and personnel: Foster seemed able to acquire whatever talent he needed from other clubs, such as Jimmie Lyons, the Detroit Stars' best player in 1920, who was transferred to the American Giants for 1921, or Foster's own younger brother, Bill, who joined the American Giants unwillingly when Rube forced the Memphis Red Sox to give him up in 1926. His critics believed he had organized the league primarily for purposes of booking games for the American Giants. With a stable schedule and reasonably solvent opponents, Foster was able to improve receipts at the gate. It is also true that when opposing clubs lost money, he was known to help them meet payroll, sometimes out of his own pocket. His American Giants won the new league's first three pennants, before being overtaken by the Kansas City Monarchs in 1923. In the same year the Hilldale Club and Bacharach Giants, the most important eastern clubs, pulled out of an agreement with the NNL and founded their own league, the Eastern Colored League (ECL). The ECL raided the older circuit for players, Foster's own ace pitcher Dave Brown among them. Eventually the two leagues reached an agreement to respect one another's contracts, and to play a world series. After two years of finishing behind the Monarchs, Foster "cleaned house" in spring, 1925, releasing several veterans (including Lyons and pitchers Dick Whitworth and Tom Williams). On May 26, Foster was nearly asphyxiated by a gas leak in Indianapolis. Though he recovered and returned to his team, his behavior grew erratic from then on. Foster had instituted a split-season format, and his American Giants finished third in both halves. The year 1926 saw him complete his team's reshaping, leaving only a handful of veterans from the championship squads of 1920 to 1922. The club finished third in the season's first half; but Foster would never finish the second. Over the years, "Foster grew increasingly paranoid. Took to carrying a revolver with him everywhere he went." Suffering from serious delusions, including one where he believed he was about to receive a call to pitch in the World Series, he was institutionalized midway through the 1926 season at an asylum in Kankakee, Illinois. The American Giants and the NNL lived on—in fact, led by Dave Malarcher, the Giants won the pennant and World Series in both 1926 and 1927—but the league clearly suffered in the absence of Foster's leadership. Foster died in 1930, never having recovered his sanity, and a year later, the league he had founded fell apart. Foster is interred in Lincoln Cemetery in Blue Island, Illinois. At his funeral, it was said that his coffin was closed, according to attendees, "at the usual hour a ballgame ends." In 1981, Foster was elected to the National Baseball Hall of Fame. He was the first representative of the Negro leagues elected as a pioneer or executive. On December 30, 2009, the U.S. Postal Service announced that it planned to issue a pair of postage stamps in June honoring Negro leagues Baseball. On July 17, 2010, the Postal Service issued a se-tenant pair of 44-cent, first-class, U.S. commemorative postage stamps, to honor the all-black professional baseball leagues that operated from 1920 to about 1960. One of the stamps depicts Foster, along with his name and the words "NEGRO LEAGUES BASEBALL". The stamps were formally issued at the Negro Leagues Baseball Museum, during the celebration of the museum's twentieth anniversary. The Negro Leagues Baseball Museum hosts the annual Andrew "Rube" Foster Lecture, in September.
https://en.wikipedia.org/wiki?curid=26415
Ring Lardner Ringgold Wilmer "Ring" Lardner (March 5, 1885 – September 25, 1933) was an American sports columnist and short story writer best known for his satirical writings on sports, marriage, and the theatre. His contemporaries Ernest Hemingway, Virginia Woolf, and F. Scott Fitzgerald all professed strong admiration for his writing. Born in Niles, Michigan, Ring Lardner was the son of wealthy parents, Henry and Lena Phillips Lardner. He was the youngest of nine children. Lardner's name came from a cousin of the same name. The cousin had been named by Lardner's uncle, Rear Admiral James L. Lardner, who had decided to name his son after a friend, Rear Admiral Cadwalader Ringgold, who was from a distinguished military family. Lardner never liked his given name and abbreviated it to Ring, naming one of his sons Ring Jr. In childhood he wore a brace for his deformed foot until he was eleven. He also had a passion for baseball, stage, and music. He later attended the Armour Institute in Chicago. Lardner married Ellis Abbott of Goshen, Indiana, in 1911. They had four sons, John, James, Ring Jr., and David. Lardner died on September 25, 1933, at the age of 48 in East Hampton, New York, of a heart attack due to complications from tuberculosis. Lardner started his writing career as a sports columnist, finding work with the newspaper "South Bend Times" in 1905. In 1907, he relocated to Chicago, where he got a job with the " Inter-Ocean". Within a year, he quit to work for the "Chicago Examiner", and then for the "Tribune". Two years later, Lardner was in St. Louis, writing the humorous baseball column "Pullman Pastimes" for Taylor Spink and the "Sporting News". Some of this work was the basis for his book "You Know Me, Al". Within three months, he was an employee of the "Boston American". In 1913, Lardner returned to the "Chicago Tribune", which became the home newspaper for his syndicated column "In the Wake of the News" (started by Hugh Keough, who had died in 1912). The column appeared in more than 100 newspapers, and is still published in the "Tribune". Lardner's Tribune and syndicated writing was not exclusively sports related: his dispatches from/near the World War One front were collected in the book "My Four Weeks in France", and his immersive coverage of the 1920 Democratic Convention resulted in Lardner receiving 0.5 votes on the 23rd ballot. In 1916, Lardner published his first successful book, "You Know Me Al", an epistolary novel written in the form of letters by "Jack Keefe", a bush-league baseball player, to a friend back home. The letters made much use of the fictional author's idiosyncratic vernacular. It had initially been published as six separate but interrelated short stories in "The Saturday Evening Post", causing some to classify the book as a collection of stories, others as a novel. Like most of Lardner's stories, "You Know Me Al" employs satire, in this case, to show the stupidity and avarice of a certain type of athlete. The journalist Andrew Ferguson wrote that "Ring Lardner thought of himself as primarily a sports columnist whose stuff wasn't destined to last, and he held to that absurd belief even after his first masterpiece, "You Know Me Al", was published in 1916 and earned the awed appreciation of Virginia Woolf, among other very serious, unfunny people." Ferguson termed the book one of the top five pieces of American humor writing. Sarah Bembrey has written about a singular event in Lardner's sportswriting experience: "In 1919 something happened that changed his way of reporting about sports and changed his love for baseball. This was the Black Sox scandal when the Chicago White Sox sold out the World Series to the Cincinnati Reds. Ring was exceptionally close to the White Sox and felt he was betrayed by the team. After the scandal, Ring always wrote about sports as if there were some kink to the outcome." Lardner's last fictional baseball writing was collected in the book "Lose with a Smile" (1933). Lardner later published such stories as "Haircut", "Some Like Them Cold", "The Golden Honeymoon", "Alibi Ike", and "A Day with Conrad Green". He also continued to write follow-up stories to "You Know Me Al", with the protagonist of that book, the headstrong but gullible Jack Keefe, experiencing various ups and downs in his major league career and in his personal life. Private Keefe's World War I training camp letters home to his friend Al were collected in the book "Treat 'Em Rough: Letters From Jack the Kaiser Killer". The sequel, "The Real Dope", followed Keefe overseas to the trenches in France. Lardner also had a lifelong fascination with the theatre, although his only Broadway three-act successes were the thrice-filmed "Elmer The Great", co-written with George M. Cohan, and "June Moon", a comedy authored with Broadway veteran George S. Kaufman. Lardner also wrote skits for the Ziegfeld Follies. and a series of brief nonsense plays that ridiculed the conventions of the theatre using zany humor and outrageous, impossible stage directions, such as "The curtain is lowered for seven days to denote the lapse of a week." He was a dedicated composer and lyricist: both his first – "Zanzibar" (1903) – and last – "June Moon" (1929) – published stage works included several Lardner tunes. He wrote at least one recorded song for Bert Williams, co-wrote one for Nora Bayes, and provided the lyrics for the song "That Old Quartet" (1913) by Nathaniel D. Mann. Other collaborators of note included Aubrey Stauffer, Jerome Kern on Very Good Eddie (1915), and Vincent Youmans – with whom he toiled on the Ziegfeld - Marilyn Miller - Astaires musical, "Smiles" (1930) . Lardner was a good friend of F. Scott Fitzgerald and other authors of the Jazz Age. His books were published by Maxwell Perkins, who also served as Fitzgerald's editor. To create his first book of short stories Lardner had to get copies from the magazines who bought the stories — he held his own short stories in low regard and did not save copies. Lardner was in some respects the model for the tragic character Abe North of Fitzgerald's last completed novel, "Tender Is the Night". Lardner influenced Ernest Hemingway, who sometimes wrote articles for his high school newspaper using the pseudonym Ring Lardner, Jr. The two met during December 1928, thanks to Max Perkins, but did not become friends. Lardner's gift for dialogue heavily influenced the writer John O'Hara, who said he learned from reading Lardner "that if you wrote down speech as it is spoken truly, you produce true characters, and the opposite is also true: if your characters don't talk like people they aren't good characters" and added, "it's the attribute most lacking in American writers and almost totally lacking in the British." J.D. Salinger referred to Lardner in two of his works," The Catcher in the Rye" and "Franny and Zooey". The protagonist says "My favorite author is my brother D.B. and my next favorite is Ring Lardner". Wayne C. Booth mentioned Lardner's famous short story "The Haircut" in his essay "Telling and Showing." In his movie "Eight Men Out" (1988) about the Black Sox scandal, writer-director John Sayles portrayed Lardner as one of the clear-eyed observers who was not taken in by the conspiracy. In one scene, Lardner strolls through the White Sox train, singing a parody of the song "I'm Forever Blowing Bubbles", changed to "I'm Forever Throwing Ballgames". In 2016, Lardner was inducted into the Chicago Literary Hall of Fame. Harry Turtledove describes his short story "Batboy" as a Ring Lardner pastiche. Neil Simon references Ring Lardner in his play Brighton Beach Memoirs. In John DeChancie's novel "Castle for Rent," Lord Incarnadine mentions having been friends with Ring Lardner. John Lardner was a newspaperman, sports columnist, and magazine writer. James Lardner, also a newspaperman, was killed in the Spanish Civil War fighting with the International Brigades. "Somebody Had to Do Something. A Memorial to James Phillips Lardner" - A 500 copy book with contributions by Ernest Hemingway, Ring Lardner, Jr., Jay Allen, Don Jesus Hernandez, El Campesino, Dolores Ibarruri, Vincent Sheean and Drawings by Castelao was published in 1939 by the James Lardner Memorial Fund. Ring Lardner, Jr. was a screenwriter who was blacklisted after the Second World War as one of the Hollywood Ten, screenwriters who were incarcerated for contempt of Congress after refusing to answer questions posed by the House Un-American Activities Committee (HUAC). He won two Academy Awards for his screenplays—one before his imprisonment and blacklisting (for Woman of the Year in 1942), and one after (for M*A*S*H in 1970). His book, "The Lardners, My Family Remembered" (), is a source of information on his father. David Lardner worked for "The New Yorker" as a general reporter and war correspondent before he was killed by a landmine near Aachen, Germany on October 19, 1944, less than one month after his arrival in Europe. Ring Lardner was a great-uncle to 1993 Pulitzer Prize winner George Lardner, Jr., a journalist at "The Washington Post" since 1963.
https://en.wikipedia.org/wiki?curid=26416
River Clyde The River Clyde (, , , or "Watter o Clyde") is a river that flows into the Firth of Clyde in Scotland. It is the eighth-longest river in the United Kingdom, and the second-longest in Scotland. Traveling through the major city of Glasgow, it was an important river for shipbuilding and trade in the British Empire. To the Romans, it was "Clota", and in the early medieval Cumbric language, it was known as "Clud" or "Clut", and was central to the Kingdom of Strathclyde ("Teyrnas Ystrad Clut"). The Clyde is formed by the confluence of two streams, the Daer Water (the headwaters of which are dammed to form the Daer Reservoir) and the Potrail Water. The Southern Upland Way crosses both streams before they meet at Watermeetings () to form the River Clyde proper. At this point, the Clyde is only from Tweed's Well, the source of the River Tweed, and is a similar distance from Annanhead Hill, the source of the River Annan. From there, it meanders northeastward before turning to the west, its flood plain used for many major roads in the area, until it reaches the town of Lanark. On the banks of the Clyde, the industrialists David Dale and Robert Owen built their mills and the model settlement of New Lanark. The mills harness the power of the Falls of Clyde, the most spectacular of which is Cora Linn. A hydroelectric power station still generates electricity here, although the mills are now a museum and World Heritage Site. The river then makes its way north-west, past (although not through) the towns of Wishaw on its eastern side and Larkhall on its western, with the surrounding environment becoming increasingly suburban. Between the towns of Motherwell and Hamilton, the course of the river has been altered to create an artificial loch within Strathclyde Park. Part of the original course can still be seen, and lies between the island and the east shore of the loch. The river then flows through Blantyre and Bothwell, where the ruined Bothwell Castle stands on a defensible promontory. Past Uddingston and into the southeast of Glasgow, the river begins to widen, meandering a course through Cambuslang, Rutherglen, and Dalmarnock. Flowing past Glasgow Green, the river is artificially straightened and widened through the centre, and although the new Clyde Arc now hinders access to the traditional Broomielaw dockland area, seagoing ships can still come upriver as far as Finnieston, where the PS "Waverley" docks. From there, it flows past the shipbuilding heartlands, through Govan, Partick, Whiteinch, Scotstoun, and Clydebank, all of which housed major shipyards, of which only two remain. The river flows out west of Glasgow, past Renfrew, and under the Erskine Bridge past Dumbarton on the north shore to the sandbank at Ardmore Point between Cardross and Helensburgh. Opposite, on the south shore, the river continues past the last Lower Clyde shipyard at Port Glasgow to Greenock, where it reaches the Tail of the Bank as the river merges into the Firth of Clyde. A significant issue of oxygen depletion in the water column has occurred at the mouth of the River Clyde. The strath of the Clyde was the focus for the G-BASE project from the British Geological Survey in the summer of 2010. The success of the Clyde at the beginning of the Industrial Revolution was driven by the location of Glasgow, being a port facing the Americas. Tobacco and cotton trade began the drive in the early 18th century. However, the shallow Clyde was not navigable for the largest ocean-going ships, so cargo had to be transferred at Greenock or Port Glasgow to smaller ships to sail upstream into Glasgow itself. In 1768, John Golborne advised the narrowing of the river and the increasing of the scour by the construction of rubble jetties and the dredging of sandbanks and shoals. A particular problem was the division of the river into two shallow channels by the Dumbuck shoal near Dumbarton. After James Watt's report on this in 1769, a jetty was constructed at Longhaugh Point to block off the southern channel. This being insufficient, a training wall called the Lang Dyke was built in 1773 on the Dumbuck shoal to stop water flowing over into the southern channel. In the late 18th and early 19th centuries, hundreds of jetties were built out from the banks between Dumbuck and the Broomielaw quay in Glasgow itself. In some cases, this resulted in an immediate deepening as the constrained water flow washed away the river bottom; in others, dredging was required. In the mid-19th century, engineers took on a much greater dredging of the Clyde, removing millions of cubic feet of silt to deepen and widen the channel. The major stumbling block in the project was a massive geological intrusion known as Elderslie Rock. As a result, the work was not completed until the 1880s. At this time, the Clyde also became an important source of inspiration for artists, such as John Atkinson Grimshaw and James Kay, willing to depict the new industrial era and the modern world. The completion of the dredging was well-timed; as steelworking grew in the city, the channel finally became navigable all the way up to Glasgow. Shipbuilding replaced trade as the major activity on the river, and shipbuilding companies were rapidly establishing themselves on the river. Soon, the Clyde gained a reputation for being the best location for shipbuilding in the British Empire, and grew to become the world's pre-eminent shipbuilding centre. "Clydebuilt" became an industry benchmark of quality, and the river's shipyards were given contracts for prestigious ocean-going liners, as well as warships, including the "Queen Mary" and the "Queen Elizabeth 2" in later years, all built in the town of Clydebank. From the founding of the Scott family's shipyard at Greenock in 1712 to the present day, over 25,000 ships have been built on the River Clyde and its Firth and on the tributary River Kelvin and River Cart together with boatyards at Maryhill and Kirkintilloch on the Forth & Clyde Canal and Blackhill on the Monkland Canal. In the same time, an estimated over 300 firms have engaged in shipbuilding on Clydeside, although probably a peak of 30 to 40 at any one time. The shipbuilding firms became household names on Clydeside, but also known around the world – Denny of Dumbarton, Scott of Greenock, Lithgows of Port Glasgow, Simon and Lobnitz of Renfrew, Alexander Stephen and Sons of Linthouse, Fairfield of Govan, Inglis of Pointhouse, Barclay Curle of Whiteinch, Connell and Yarrow of Scotstoun, to name but a few. Almost as famous were the engineering firms that supplied the machinery to drive these vessels, the boilers and pumps, and steering gear – Rankin & Blackmore, Hastie's and Kincaid's of Greenock, Rowan's of Finnieston, Weir's of Cathcart, Howden's of Tradeston, and Babcock & Wilcox of Renfrew. One 'Clyde' shipyard was not even located on one of these waterways: Alley & MacLellan's Sentinel Works in Jessie Street at Polmadie is around half a mile from the Clyde, yet it is reputed to have constructed over 500 vessels, many of which were assembled then 'knocked down' to kit form for despatch to some far distant and remote location – one such vessel being the SS "Chauncy Maples", still in service on Lake Malawi. Clyde shipbuilding reached its peak in the years just before World War I, and over 370 ships were thought to have been completed in 1913. The first recorded Clyde racing yacht, a 46-ton cutter, was built by Scotts of Greenock in 1803, and the great Scottish yacht designer William Fyfe did not start designing yachts until 1807. The first yacht club on the Clyde was the Northern Yacht Club, which appeared in 1824 and received its Royal Charter in 1831. The club was founded to organise and encourage the sport, and by 1825, Scottish and Irish clubs were racing against each other on the Clyde. However, yachting and yacht building did not really take off until the middle of the 19th century. The Clyde became famous worldwide for its significant contribution to yachting and yachtbuilding, and was the home of many notable designers: William Fife III, Alfred Mylne, G. L. Watson, E. McGruer, and David Boyd. It was also the home of many famous yacht yards. Robertson's Yard started repairing boats in a small workshop at Sandbank in 1876, and went on to become one of the foremost wooden boat builders on the Clyde. The 'golden years' of Robertson's yard were in the early 20th century, when they started building classic racing yachts. More than 55 boats were built by Robertson's in preparation for the First World War, and the yard remained busy even during the Great Depression in the 1930s, as many wealthy businessmen developed a passion for yacht racing on the Clyde. During World War II, the yard was devoted to Admiralty work, producing large, high-speed Fairmile Marine motor boats (motor torpedo boats and motor gun boats). After the war, the yard built the successful one-class Loch Longs and two David Boyd-designed challengers for the America's Cup: "Sceptre" (1958) and "Sovereign" (1964). Due to difficult business conditions in 1965, the yard was turned over to GRP production work (mainly Pipers and Etchells) until it closed in 1980. During its 104-year history, Robertson's Yard built 500 boats, many of which are still sailing today. Other notable boatyards on the Clyde, both situated on the Rosneath peninsula on the banks of the Gare Loch within half a mile of each other, were Silvers (1910-1970) and McGruers (1910-1973). Both yards built many well-known and classic yachts, some of which are still sailing today. McGruers built over 700 boats but both yards are no longer in business as boatbuilders. The Glasgow Humane Society is the oldest lifesaving organisation in the world, founded in 1790 and has been responsible for the safety and preservation of life on Glasgow's waterways. The downfall of the Clyde as a major industrial centre came during and after World War II. Clydebank in particular was targeted by the Luftwaffe and sustained heavy damage. The immediate postwar period had a severe reduction in warship orders, which was balanced by a prolonged boom in merchant shipbuilding. By the end of the 1950s, however, the rise of other shipbuilding nations, recapitalised and highly productive, made many European yards uncompetitive. Several Clydeside yards booked a series of loss-making contracts in the hope of weathering the storm. However, by the mid-1960s, shipbuilding on the Clyde was becoming increasingly uneconomical and potentially faced collapse. This culminated in the closure of Harland and Wolff's Linthouse yard and a bankruptcy crisis facing Fairfields of Govan. The government responded by creating the Upper Clyde Shipbuilders consortium. After the consortium's controversial collapse in 1971, the Labour government of James Callaghan later passed the Aircraft and Shipbuilding Industries Act, which nationalised most of the Clyde's shipyards and grouped them with other major British shipyards such as British Shipbuilders. Today, two major shipyards remain in operation on the Upper Clyde; they are owned by a naval defence contractor, BAE Systems Surface Ships, which specialises in the design and construction of technologically advanced warships for the Royal Navy and other navies around the world. The two yards are the former Yarrow yard at Scotstoun and Fairfields at Govan. Also, the King George V Dock is operated by the Clyde Port Authority. On the Lower Clyde, the Scottish Government owned Ferguson Shipbuilders at Port Glasgow is the last survivor of the many shipyards that once dominated Port Glasgow and Greenock – its mainstay being the construction of car ferries. The Clyde Waterfront Regeneration project is expected to attract investment of up to £5.6bn in the area from Glasgow Green to Dumbarton. Market gardens and garden centres have grown up on the fertile plains of the Clyde Valley. Tourism has also brought many back to the riverside, especially in Glasgow, where former docklands have given way to housing and amenities on the banks in the city, for instance, the Glasgow Harbour project, the Glasgow Science Centre, and the creation of the Scottish Exhibition and Conference Centre. With the migration of the commercial Port of Glasgow downstream to the deeper waters of the Firth of Clyde, the river has been extensively cleaned up, once having a very poor reputation for pollution and sewage, to make it suitable for recreational use. The Clyde Walkway is a foot and mountain bike path that follows the course of the Clyde between Glasgow and New Lanark. It was completed in 2005, and is now designated as one of Scotland's Great Trails by Scottish Natural Heritage. Organic chemical pollution contained in sediments of the Clyde estuary have been evaluated by British Geological Survey. Surface sediments from the Glasgow reaches of the Clyde, Cunningar to Milton, contained polyaromatic hydrocarbons (PAH) from 630 µg/kg to 23,711 µg/kg and polychlorinated biphenyl (PCB) in the range of 5 to 130.5 µg/kg, which classifies these sediments as non-toxic. However, a later study showed PCB concentrations as high as 5,797 µg/kg, which is above published threshold levels for the chlorinated compounds. Comparison of individual PAH compounds with different thermal stabilities shows that the source of PAH pollution in the Clyde changes downriver. PAH in the inner Clyde (Cunningar to Milton) are from combustion sources (vehicle exhaust, coal burning), whereas PAH in the outer Clyde are from petroleum spills The amount and type of sedimentary pollution contained in the Clyde is closely related to the area's industrial history. Assessment of the changing type of pollution through time (1750 to 2002) has been gained through the collection of seven sediment cores of one metre's depth which were dated using lead concentrations and changing lead isotope ratios. The sediments showed a long but declining history of coal usage and an increasing reliance on petroleum fuels from about the 1950s. Thereafter, declining hydrocarbon pollution was followed by the onset (1950s), peak (1965–1977), and decline (after 1980s) in total PCB concentrations. Although pollution from heavy industry and power generation is falling, evidence indicates that man-made pollution from new synthetic compounds in electrical products and textiles is on the increase. The amounts of 16 polybrominated diphenyl ether (PBDE) compounds used as flame retardants in televisions, computers, and furniture upholstery were measured in sediment cores collected from six sites between Princes Dock and Greenock. Comparison of the amounts of PBDE compounds revealed a decline in certain compounds in line with the European ban on production of mixtures containing environmentally harmful PBDE with eight and nine bromine atoms and a rise in the less harmful mixture composed of ten bromine atoms. The Clyde is central to the "Para Handy" novels of Neil Munro, and subsequent adaptations, and features in novels by Alasdair Gray, Matthew Fitt, and Robin Jenkins. It appears in the "Ossian" poetry of James Macpherson, as well as the works of John Wilson, William McGonagall, Edwin Morgan, Norman McCaig, Douglas Dunn and W.S. Graham. It features in artworks by, amongst others, William McTaggart, J.M.W. Turner, Robert Salmon, John Atkinson Grimshaw, Stanley Spencer, and George Wyllie. The Clyde appears prominently in the films "Young Adam", "Sweet Sixteen", "Just a Boys' Game", and "Down Where The Buffalo Go", and was the subject of the Academy Award-winning film documentary "Seawards the Great Ships". It is referenced in the traditional folk songs "Clyde's Water" and "Black Is the Color (of My True Love's Hair)", as well as "Song of the Clyde", popularised by Kenneth McKellar.
https://en.wikipedia.org/wiki?curid=26417
Rational root theorem In algebra, the rational root theorem (or rational root test, rational zero theorem, rational zero test or theorem) states a constraint on rational solutions of a polynomial equation with integer coefficients formula_2 and formula_3. Solutions of the equation are also called roots or zeroes of the polynomial on the left side. The theorem states that each rational solution , written in lowest terms so that and are relatively prime, satisfies: The rational root theorem is a special case (for a single linear factor) of Gauss's lemma on the factorization of polynomials. The integral root theorem is the special case of the rational root theorem when the leading coefficient is . The theorem is used to find all rational roots of a polynomial, if any. It gives a finite number of possible fractions which can be checked to see if they are roots. If a rational root is found, a linear polynomial can be factored out of the polynomial using polynomial long division, resulting in a polynomial of lower degree whose roots are also roots of the original polynomial. The general cubic equation with integer coefficients has three solutions in the complex plane. If the rational root test finds no rational solutions, then the only way to express the solutions algebraically uses cube roots. But if the test finds a rational solution , then factoring out leaves a quadratic polynomial whose two roots, found with the quadratic formula, are the remaining two roots of the cubic, avoiding cube roots. Let Suppose for some coprime : Now multiply both sides by . By shifting the constant term (the term containing ) to the right side, and factoring out on the left side, produces Thus, divides . But is coprime to and therefore to , so by Euclid's lemma must divide the remaining factor of the product. On the other hand, shifting the leading term to the right side and factoring out on the left side, gives Reasoning as before, it follows that divides . Should there be a nontrivial factor dividing all the coefficients of the polynomial, then one can divide by the greatest common divisor of the coefficients so as to obtain a primitive polynomial in the sense of Gauss's lemma; this does not alter the set of rational roots and only strengthens the divisibility conditions. That lemma says that if the polynomial factors in , then it also factors in as a product of primitive polynomials. Now any rational root corresponds to a factor of degree 1 in of the polynomial, and its primitive representative is then , assuming that and are coprime. But any multiple in of has leading term divisible by and constant term divisible by , which proves the statement. This argument shows that more generally, any irreducible factor of can be supposed to have integer coefficients, and leading and constant coefficients dividing the corresponding coefficients of . In the polynomial any rational root fully reduced would have to have a numerator that divides evenly into 1 and a denominator that divides evenly into 2. Hence the only possible rational roots are ±1/2 and ±1; since neither of these equates the polynomial to zero, it has no rational roots. In the polynomial the only possible rational roots would have a numerator that divides 6 and a denominator that divides 1, limiting the possibilities to ±1, ±2, ±3, and ±6. Of these, 1, 2, and –3 equate the polynomial to zero, and hence are its rational roots. (In fact these are its only roots since a cubic has only three roots; in general, a polynomial could have some rational and some irrational roots.) Every rational root of the polynomial must be among the numbers symbolically indicated by: These 8 root candidates can be tested by evaluating , for example using Horner's method. It turns out there is exactly one with . This process may be made more efficient: if , it can be used to shorten the list of remaining candidates. For example, does not work, as . Substituting yields a polynomial in  with constant term , while the coefficient of remains the same as the coefficient of . Applying the rational root theorem thus yields the possible roots formula_14, so that True roots must occur on both lists, so list of rational root candidates has shrunk to just and If rational roots are found, Horner's method will also yield a polynomial of degree whose roots, together with the rational roots, are exactly the roots of the original polynomial. If none of the candidates is a solution, there can be no rational solution.
https://en.wikipedia.org/wiki?curid=26426
Round Table The Round Table (, , ) is King Arthur's famed table in the Arthurian legend, around which he and his knights congregate. As its name suggests, it has no head, implying that everyone who sits there has equal status. The table was first described in 1155 by Wace, who relied on previous depictions of Arthur's fabulous retinue. The symbolism of the Round Table developed over time; by the close of the 12th century it had come to represent the chivalric order associated with Arthur's court, the Knights of the Round Table. Though the Round Table is not mentioned in the earliest accounts, tales of Arthur having a marvelous court made up of many prominent warriors is ancient. Geoffrey of Monmouth, in his "Historia Regum Britanniae" (composed c. 1136) says that, after establishing peace throughout Britain, Arthur "increased his personal entourage by inviting very distinguished men from far-distant kingdoms to join it." The code of chivalry so important in later medieval romance figures in as well, as Geoffrey says Arthur established "such a code of courtliness in his household that he inspired peoples living far away to imitate him." Arthur's court was well known to Welsh storytellers; in the romance "Culhwch and Olwen", the protagonist Culhwch invokes the names of 225 individuals affiliated with Arthur. The fame of Arthur's entourage became so prominent in Welsh tradition that in the later additions to the Welsh Triads, the formula tying named individuals to "Arthur's Court" in the triad titles began to supersede the older "Island of Britain" formula. Though the code of chivalry crucial to later continental romances dealing with the Round Table is mostly absent from the Welsh material, some passages of "Culhwch and Olwen" seem to reference it. For instance, Arthur explains the ethos of his court, saying "[w]e are nobles as long as we are sought out: the greater the bounty we may give, the greater our nobility, fame and honour." Though no Round Table appears in the early Welsh texts, Arthur is associated with various items of household furniture. The earliest of these is Saint Carannog's mystical floating altar in that saint's 12th century "Vita". In the story Arthur has found the altar and tries unsuccessfully to use it as a table; he returns it to Carannog in exchange for the saint ridding the land of a meddlesome dragon. Elements of Arthur's household figure into local topographical folklore throughout Britain as early as the early 12th century, with various landmarks being named "Arthur's Seat", "Arthur's Oven", and "Arthur's Bed-chamber". A henge at Eamont Bridge near Penrith, Cumbria is known as "King Arthur's Round Table". The still-visible Roman amphitheatre at Caerleon has been associated with the Round Table, and it has been suggested as a possible source for the legend. Following archaeological discoveries at the Roman ruins in Chester, some writers suggested that the Chester Roman Amphitheatre was the true prototype of the Round Table; however, the English Heritage Commission, acting as consultants to a History Channel documentary in which the claim was made, stated that there was no archaeological basis to the story. The Round Table first appeared in Wace's "Roman de Brut", a Norman language adaptation of Geoffrey's "Historia" finished in 1155. Wace says Arthur created the Round Table to prevent quarrels among his barons, none of whom would accept a lower place than the others. Layamon added to the story when he adapted Wace's work into the Middle English "Brut" in the early 13th century, saying that the quarrel between Arthur's vassals led to violence at a Yuletide feast. In response, a Cornish carpenter built an enormous but easily transportable Round Table to prevent further dispute. Wace claims he was not the source of the Round Table; both he and Layamon credited it instead to the Bretons. Some scholars have doubted this claim, while others believe it may be true. There is some similarity between the chroniclers' description of the Round Table and a custom recorded in Celtic stories, in which warriors sit in a circle around the king or lead warrior, in some cases feuding over the order of precedence as in Layamon. There is a possibility that Wace, contrary to his own claims, derived Arthur's round table not from any Breton source, but rather from medieval biographies of Charlemagne—notably Einhard's "Vita Caroli" and Notker the Stammerer's "De Carolo Magno"—in which the king is said to have possessed a round table decorated with a map of Rome. The Round Table takes on new dimensions in the romances of the late 12th and early 13th century, where it becomes a symbol of the famed order of chivalry which flourishes under Arthur. In Robert de Boron's "Merlin", written around 1200, the magician Merlin creates the Round Table in imitation of the table of the Last Supper and of Joseph of Arimathea's Holy Grail table. This table, here made for Arthur's father Uther Pendragon rather than Arthur himself, has twelve seats and one empty place to mark the betrayal of Judas. This seat must remain empty until the coming of the knight who will achieve the Grail. The Didot "Perceval", a prose continuation of Robert's work, takes up the story, and the knight Percival sits in the seat and initiates the Grail quest. The prose cycles of the 13th century, the Lancelot-Grail (Vulgate) Cycle and the Post-Vulgate Cycle, further adapt the chivalric attributes of the Round Table. Here it is the perfect knight Galahad, rather than Percival, who assumes the empty seat, now called the Siege Perilous. Galahad's arrival marks the start of the Grail quest as well as the end of the Arthurian era. In these works the Round Table is kept by King Leodegrance of Cameliard after Uther's death; Arthur inherits it when he marries Leodegrance's daughter Guinevere. Other versions treat the Round Table differently, for instance Arthurian works from Italy like "La Tavola Ritonda" ("The Round Table") often distinguish between the knights of the "Old Table" of Uther's time and those of Arthur's "New Table". In the Post-Vulgate, the Table is eventually destroyed by King Mark during his invasion of Logres after the deaths of Arthur and almost all of the Knights, many of whom in fact had killed each other, especially in internal conflicts at the end of the cycle. During the Middle Ages, festivals called Round Tables were celebrated throughout Europe in imitation of Arthur's court. These events featured jousting, dancing, and feasting, and in some cases attending knights assumed the identities of Arthur's entourage. The artifact is known as the "Winchester Round Table", a large tabletop hanging in Winchester Castle bearing the names of various knights of Arthur's court, was probably created for a Round Table tournament. The current paintwork is late; it was done by order of Henry VIII of England for Holy Roman Emperor Charles V's 1522 state visit, and depicts Henry himself sitting in Arthur's seat above a Tudor rose. The table itself is considerably older; dendrochronology calculates the date of construction to 1250–1280—during the reign of Edward I—using timber from store felled over a period of years. Edward was an Arthurian enthusiast who attended at least five Round Tables and hosted one himself in 1299, which may have been the occasion for the creation of the Winchester Round Table. Martin Biddle, from an examination of Edward's financial accounts, links it instead with a tournament Edward held near Winchester on 20 April 1290, to mark the betrothal of one of his daughters.
https://en.wikipedia.org/wiki?curid=26427
Rosetta Stone The Rosetta Stone is a granodiorite stele discovered in 1799 which is inscribed with three versions of a decree issued in Memphis, Egypt in 196 BC during the Ptolemaic dynasty on behalf of King Ptolemy V Epiphanes. The top and middle texts are in Ancient Egyptian using hieroglyphic and Demotic scripts respectively, while the bottom is in Ancient Greek. The decree has only minor differences among the three versions, so the Rosetta Stone became key to deciphering Egyptian hieroglyphs, thereby opening a window into ancient Egyptian history. The stone was carved during the Hellenistic period and is believed to have originally been displayed within a temple, possibly at nearby Sais. It was probably moved in late antiquity or during the Mameluk period, and was eventually used as building material in the construction of Fort Julien near the town of Rashid (Rosetta) in the Nile Delta. It was discovered there in July 1799 by French soldier Pierre-François Bouchard during the Napoleonic campaign in Egypt. It was the first Ancient Egyptian bilingual text recovered in modern times, and it aroused widespread public interest with its potential to decipher this previously untranslated hieroglyphic script. Lithographic copies and plaster casts began circulating among European museums and scholars. The British defeated the French and took the stone to London under the Capitulation of Alexandria in 1801. It has been on public display at the British Museum almost continuously since 1802 and is the most visited object there. Study of the decree was already underway when the first full translation of the Greek text appeared in 1803. Jean-François Champollion announced the transliteration of the Egyptian scripts in Paris in 1822; it took longer still before scholars were able to read Ancient Egyptian inscriptions and literature confidently. Major advances in the decoding were recognition that the stone offered three versions of the same text (1799); that the demotic text used phonetic characters to spell foreign names (1802); that the hieroglyphic text did so as well, and had pervasive similarities to the demotic (1814); and that phonetic characters were also used to spell native Egyptian words (1822–1824). Three other fragmentary copies of the same decree were discovered later, and several similar Egyptian bilingual or trilingual inscriptions are now known, including three slightly earlier Ptolemaic decrees: the Decree of Alexandria in 243 BC, the Decree of Canopus in 238 BC, and the Memphis decree of Ptolemy  IV, c. 218 BC. The Rosetta Stone is no longer unique, but it was the essential key to the modern understanding of ancient Egyptian literature and civilisation. The term "Rosetta Stone" is now used to refer to the essential clue to a new field of knowledge. The Rosetta Stone is listed as "a stone of black granodiorite, bearing three inscriptions ... found at Rosetta" in a contemporary catalogue of the artefacts discovered by the French expedition and surrendered to British troops in 1801. At some period after its arrival in London, the inscriptions were coloured in white chalk to make them more legible, and the remaining surface was covered with a layer of carnauba wax designed to protect it from visitors' fingers. This gave a dark colour to the stone that led to its mistaken identification as black basalt. These additions were removed when the stone was cleaned in 1999, revealing the original dark grey tint of the rock, the sparkle of its crystalline structure, and a pink vein running across the top left corner. Comparisons with the Klemm collection of Egyptian rock samples showed a close resemblance to rock from a small granodiorite quarry at Gebel Tingar on the west bank of the Nile, west of Elephantine in the region of Aswan; the pink vein is typical of granodiorite from this region. The Rosetta Stone is high at its highest point, wide, and thick. It weighs approximately . It bears three inscriptions: the top register in Ancient Egyptian hieroglyphs, the second in the Egyptian demotic script, and the third in Ancient Greek. The front surface is polished and the inscriptions lightly incised on it; the sides of the stone are smoothed, but the back is only roughly worked, presumably because this would have not been visible when it was erected. The Rosetta Stone is a fragment of a larger stele. No additional fragments were found in later searches of the Rosetta site. Owing to its damaged state, none of the three texts is absolutely complete. The top register, composed of Egyptian hieroglyphs, suffered the most damage. Only the last 14 lines of the hieroglyphic text can be seen; all of them are broken on the right side, and 12 of them on the left. Below it, the middle register of demotic text has survived best; it has 32 lines, of which the first 14 are slightly damaged on the right side. The bottom register of Greek text contains 54 lines, of which the first 27 survive in full; the rest are increasingly fragmentary due to a diagonal break at the bottom right of the stone. The full length of the hieroglyphic text and the total size of the original stele, of which the Rosetta Stone is a fragment, can be estimated based on comparable stelae that have survived, including other copies of the same order. The slightly earlier decree of Canopus, erected in 238 BC during the reign of Ptolemy III, is and wide, and contains 36 lines of hieroglyphic text, 73 of demotic text, and 74 of Greek. The texts are of similar length. From such comparisons, it can be estimated that an additional 14 or 15 lines of hieroglyphic inscription are missing from the top register of the Rosetta Stone, amounting to another . In addition to the inscriptions, there would probably have been a scene depicting the king being presented to the gods, topped with a winged disc, as on the Canopus Stele. These parallels, and a hieroglyphic sign for "stela" on the stone itself, O26 (see Gardiner's sign list), suggest that it originally had a rounded top. The height of the original stele is estimated to have been about . The stele was erected after the coronation of King Ptolemy V and was inscribed with a decree that established the divine cult of the new ruler. The decree was issued by a congress of priests who gathered at Memphis. The date is given as "4 Xandikos" in the Macedonian calendar and "18 Mekhir" in the Egyptian calendar, which corresponds to . The year is stated as the ninth year of Ptolemy V's reign (equated with 197/196 BC), which is confirmed by naming four priests who officiated in that year: Aetos son of Aetos was priest of the divine cults of Alexander the Great and the five Ptolemies down to Ptolemy V himself; the other three priests named in turn in the inscription are those who led the worship of Berenice Euergetis (wife of Ptolemy III), Arsinoe Philadelphos (wife and sister of Ptolemy II), and Arsinoe Philopator, mother of Ptolemy V. However, a second date is also given in the Greek and hieroglyphic texts, corresponding to , the official anniversary of Ptolemy's coronation. The demotic text conflicts with this, listing consecutive days in March for the decree and the anniversary. It is uncertain why this discrepancy exists, but it is clear that the decree was issued in 196 BC and that it was designed to re-establish the rule of the Ptolemaic kings over Egypt. The decree was issued during a turbulent period in Egyptian history. Ptolemy V Epiphanes reigned from 204 to 181 BC, the son of Ptolemy IV Philopator and his wife and sister Arsinoe. He had become ruler at the age of five after the sudden death of both of his parents, who were murdered in a conspiracy that involved Ptolemy IV's mistress Agathoclea, according to contemporary sources. The conspirators effectively ruled Egypt as Ptolemy V's guardians until a revolt broke out two years later under general Tlepolemus, when Agathoclea and her family were lynched by a mob in Alexandria. Tlepolemus, in turn, was replaced as guardian in 201 BC by Aristomenes of Alyzia, who was chief minister at the time of the Memphis decree. Political forces beyond the borders of Egypt exacerbated the internal problems of the Ptolemaic kingdom. Antiochus III the Great and Philip V of Macedon had made a pact to divide Egypt's overseas possessions. Philip had seized several islands and cities in Caria and Thrace, while the Battle of Panium (198 BC) had resulted in the transfer of Coele-Syria, including Judaea, from the Ptolemies to the Seleucids. Meanwhile, in the south of Egypt, there was a long-standing revolt that had begun during the reign of Ptolemy IV, led by Horwennefer and by his successor Ankhwennefer. Both the war and the internal revolt were still ongoing when the young Ptolemy V was officially crowned at Memphis at the age of 12 (seven years after the start of his reign) and when, just over a year later, the Memphis decree was issued. Stelae of this kind, which were established on the initiative of the temples rather than that of the king, are unique to Ptolemaic Egypt. In the preceding Pharaonic period it would have been unheard of for anyone but the divine rulers themselves to make national decisions: by contrast, this way of honoring a king was a feature of Greek cities. Rather than making his eulogy himself, the king had himself glorified and deified by his subjects or representative groups of his subjects. The decree records that Ptolemy V gave a gift of silver and grain to the temples. It also records that there was particularly high flooding of the Nile in the eighth year of his reign, and he had the excess waters dammed for the benefit of the farmers. In return the priesthood pledged that the king's birthday and coronation days would be celebrated annually and that all the priests of Egypt would serve him alongside the other gods. The decree concludes with the instruction that a copy was to be placed in every temple, inscribed in the "language of the gods" (Egyptian hieroglyphs), the "language of documents" (Demotic), and the "language of the Greeks" as used by the Ptolemaic government. Securing the favour of the priesthood was essential for the Ptolemaic kings to retain effective rule over the populace. The High Priests of Memphis—where the king was crowned—were particularly important, as they were the highest religious authorities of the time and had influence throughout the kingdom. Given that the decree was issued at Memphis, the ancient capital of Egypt, rather than Alexandria, the centre of government of the ruling Ptolemies, it is evident that the young king was anxious to gain their active support. Thus, although the government of Egypt had been Greek-speaking ever since the conquests of Alexander the Great, the Memphis decree, like the three similar earlier decrees, included texts in Egyptian to show its connection to the general populace by way of the literate Egyptian priesthood. There can be no one definitive English translation of the decree, not only because modern understanding of the ancient languages continues to develop, but also because of the minor differences between the three original texts. Older translations by E. A. Wallis Budge (1904, 1913) and Edwyn R. Bevan (1927) are easily available but are now outdated, as can be seen by comparing them with the recent translation by R. S. Simpson, which is based on the demotic text and can be found online, or, best of all, with the modern translations of all three texts, with introduction and facsimile drawing, that were published by Quirke and Andrews in 1989. The stele was almost certainly not originally placed at Rashid (Rosetta) where it was found, but more likely came from a temple site farther inland, possibly the royal town of Sais. The temple from which it originally came was probably closed around AD 392 when Roman emperor Theodosius I ordered the closing of all non-Christian temples of worship. The original stele broke at some point, its largest piece becoming what we now know as the Rosetta Stone. Ancient Egyptian temples were later used as quarries for new construction, and the Rosetta Stone probably was re-used in this manner. Later it was incorporated in the foundations of a fortress constructed by the Mameluke Sultan Qaitbay (c. 1416/18–1496) to defend the Bolbitine branch of the Nile at Rashid. There it lay for at least another three centuries until its rediscovery. Three other inscriptions relevant to the same Memphis decree have been found since the discovery of the Rosetta Stone: the Nubayrah Stele, a stele found in Elephantine and Noub Taha, and an inscription found at the Temple of Philae (on the Philae obelisk). Unlike the Rosetta Stone, the hieroglyphic texts of these inscriptions were relatively intact. The Rosetta Stone had been deciphered long before they were found, but later Egyptologists have used them to refine the reconstruction of the hieroglyphs that must have been used in the lost portions of the hieroglyphic text on the Rosetta Stone. Napoleon's 1798 campaign in Egypt inspired a burst of Egyptomania in Europe, and especially France. A corps of 167 technical experts ("savants"), known as the "Commission des Sciences et des Arts", accompanied the French expeditionary army to Egypt. On 1799, French soldiers under the command of Colonel d'Hautpoul were strengthening the defences of Fort Julien, a couple of miles north-east of the Egyptian port city of Rosetta (modern-day Rashid). Lieutenant Pierre-François Bouchard spotted a slab with inscriptions on one side that the soldiers had uncovered. He and d'Hautpoul saw at once that it might be important and informed General Jacques-François Menou, who happened to be at Rosetta. The find was announced to Napoleon's newly founded scientific association in Cairo, the Institut d'Égypte, in a report by Commission member Michel Ange Lancret noting that it contained three inscriptions, the first in hieroglyphs and the third in Greek, and rightly suggesting that the three inscriptions were versions of the same text. Lancret's report, dated 1799, was read to a meeting of the Institute soon after . Bouchard, meanwhile, transported the stone to Cairo for examination by scholars. Napoleon himself inspected what had already begun to be called "la Pierre de Rosette", the Rosetta Stone, shortly before his return to France in August 1799. The discovery was reported in September in "Courrier de l'Égypte", the official newspaper of the French expedition. The anonymous reporter expressed a hope that the stone might one day be the key to deciphering hieroglyphs. In 1800 three of the Commission's technical experts devised ways to make copies of the texts on the stone. One of these experts was Jean-Joseph Marcel, a printer and gifted linguist, who is credited as the first to recognise that the middle text was written in the Egyptian demotic script, rarely used for stone inscriptions and seldom seen by scholars at that time, rather than Syriac as had originally been thought. It was artist and inventor Nicolas-Jacques Conté who found a way to use the stone itself as a printing block to reproduce the inscription. A slightly different method was adopted by Antoine Galland. The prints that resulted were taken to Paris by General Charles Dugua. Scholars in Europe were now able to see the inscriptions and attempt to read them. After Napoleon's departure, French troops held off British and Ottoman attacks for another 18 months. In March 1801, the British landed at Aboukir Bay. Menou was now in command of the French expedition. His troops, including the Commission, marched north towards the Mediterranean coast to meet the enemy, transporting the stone along with many other antiquities. He was defeated in battle, and the remnant of his army retreated to Alexandria where they were surrounded and besieged, the stone now inside the city. Menou surrendered on August 30. After the surrender, a dispute arose over the fate of the French archaeological and scientific discoveries in Egypt, including the artefacts, biological specimens, notes, plans, and drawings collected by the members of the commission. Menou refused to hand them over, claiming that they belonged to the Institute. British General John Hely-Hutchinson refused to end the siege until Menou gave in. Scholars Edward Daniel Clarke and William Richard Hamilton, newly arrived from England, agreed to examine the collections in Alexandria and claimed to have found many artefacts that the French had not revealed. In a letter home, Clarke said that "we found much more in their possession than was represented or imagined". Hutchinson claimed that all materials were property of the British Crown, but French scholar Étienne Geoffroy Saint-Hilaire told Clarke and Hamilton that the French would rather burn all their discoveries than turn them over, referring ominously to the destruction of the Library of Alexandria. Clarke and Hamilton pleaded the French scholars' case to Hutchinson, who finally agreed that items such as natural history specimens would be considered the scholars' private property. Menou quickly claimed the stone, too, as his private property. Hutchinson was equally aware of the stone's unique value and rejected Menou's claim. Eventually an agreement was reached, and the transfer of the objects was incorporated into the Capitulation of Alexandria signed by representatives of the British, French, and Ottoman forces. It is not clear exactly how the stone was transferred into British hands, as contemporary accounts differ. Colonel Tomkyns Hilgrove Turner, who was to escort it to England, claimed later that he had personally seized it from Menou and carried it away on a gun-carriage. In a much more detailed account, Edward Daniel Clarke stated that a French "officer and member of the Institute" had taken him, his student John Cripps, and Hamilton secretly into the back streets behind Menou's residence and revealed the stone hidden under protective carpets among Menou's baggage. According to Clarke, their informant feared that the stone might be stolen if French soldiers saw it. Hutchinson was informed at once and the stone was taken away—possibly by Turner and his gun-carriage. Turner brought the stone to England aboard the captured French frigate HMS "Egyptienne", landing in Portsmouth in February 1802. His orders were to present it and the other antiquities to King George III. The King, represented by War Secretary Lord Hobart, directed that it should be placed in the British Museum. According to Turner's narrative, he and Hobart agreed that the stone should be presented to scholars at the Society of Antiquaries of London, of which Turner was a member, before its final deposit in the museum. It was first seen and discussed there at a meeting on 1802. In 1802, the Society created four plaster casts of the inscriptions, which were given to the universities of Oxford, Cambridge and Edinburgh and to Trinity College Dublin. Soon afterwards, prints of the inscriptions were made and circulated to European scholars. Before the end of 1802, the stone was transferred to the British Museum, where it is located today. New inscriptions painted in white on the left and right edges of the slab stated that it was "Captured in Egypt by the British Army in 1801" and "Presented by King George III". The stone has been exhibited almost continuously in the British Museum since June 1802. During the middle of the 19th century, it was given the inventory number "EA 24", "EA" standing for "Egyptian Antiquities". It was part of a collection of ancient Egyptian monuments captured from the French expedition, including a sarcophagus of Nectanebo II (EA 10), the statue of a high priest of Amun (EA 81), and a large granite fist (EA 9). The objects were soon discovered to be too heavy for the floors of Montagu House (the original building of The British Museum), and they were transferred to a new extension that was added to the mansion. The Rosetta Stone was transferred to the sculpture gallery in 1834 shortly after Montagu House was demolished and replaced by the building that now houses the British Museum. According to the museum's records, the Rosetta Stone is its most-visited single object, a simple image of it was the museum's best selling postcard for several decades, and a wide variety of merchandise bearing the text from the Rosetta Stone (or replicating its distinctive shape) is sold in the museum shops. The Rosetta Stone was originally displayed at a slight angle from the horizontal, and rested within a metal cradle that was made for it, which involved shaving off very small portions of its sides to ensure that the cradle fitted securely. It originally had no protective covering, and it was found necessary by 1847 to place it in a protective frame, despite the presence of attendants to ensure that it was not touched by visitors. Since 2004 the conserved stone has been on display in a specially built case in the centre of the Egyptian Sculpture Gallery. A replica of the Rosetta Stone is now available in the King's Library of the British Museum, without a case and free to touch, as it would have appeared to early 19th-century visitors. The museum was concerned about heavy bombing in London towards the end of the First World War in 1917, and the Rosetta Stone was moved to safety, along with other portable objects of value. The stone spent the next two years below ground level in a station of the Postal Tube Railway at Mount Pleasant near Holborn. Other than during wartime, the Rosetta Stone has left the British Museum only once: for one month in October 1972, to be displayed alongside Champollion's "Lettre" at the Louvre in Paris on the 150th anniversary of the letter's publication. Even when the Rosetta Stone was undergoing conservation measures in 1999, the work was done in the gallery so that it could remain visible to the public. Prior to the discovery of the Rosetta Stone and its eventual decipherment, the ancient Egyptian language and script had not been understood since shortly before the fall of the Roman Empire. The usage of the hieroglyphic script had become increasingly specialised even in the later Pharaonic period; by the 4th century AD, few Egyptians were capable of reading them. Monumental use of hieroglyphs ceased as temple priesthoods died out and Egypt was converted to Christianity; the last known inscription is dated to , found at Philae and known as the Graffito of Esmet-Akhom. The last demotic text, also from Philae, was written in 452. Hieroglyphs retained their pictorial appearance, and classical authors emphasised this aspect, in sharp contrast to the Greek and Roman alphabets. In the 5th century, the priest Horapollo wrote "Hieroglyphica", an explanation of almost 200 glyphs. His work was believed to be authoritative, yet it was misleading in many ways, and this and other works were a lasting impediment to the understanding of Egyptian writing. Later attempts at decipherment were made by Arab historians in medieval Egypt during the 9th and 10th centuries. Dhul-Nun al-Misri and Ibn Wahshiyya were the first historians to study hieroglyphs, by comparing them to the contemporary Coptic language used by Coptic priests in their time. The study of hieroglyphs continued with fruitless attempts at decipherment by European scholars, notably Johannes Goropius Becanus in the 16th century, Athanasius Kircher in the 17th, and Georg Zoëga in the 18th. The discovery of the Rosetta Stone in 1799 provided critical missing information, gradually revealed by a succession of scholars, that eventually allowed Jean-François Champollion to solve the puzzle that Kircher had called the riddle of the Sphinx. The Greek text on the Rosetta Stone provided the starting point. Ancient Greek was widely known to scholars, but they were not familiar with details of its use in the Hellenistic period as a government language in Ptolemaic Egypt; large-scale discoveries of Greek papyri were a long way in the future. Thus, the earliest translations of the Greek text of the stone show the translators still struggling with the historical context and with administrative and religious jargon. Stephen Weston verbally presented an English translation of the Greek text at a Society of Antiquaries meeting in April 1802. Meanwhile, two of the lithographic copies made in Egypt had reached the Institut de France in Paris in 1801. There, librarian and antiquarian Gabriel de La Porte du Theil set to work on a translation of the Greek, but he was dispatched elsewhere on Napoleon's orders almost immediately, and he left his unfinished work in the hands of colleague Hubert-Pascal Ameilhon. Ameilhon produced the first published translations of the Greek text in 1803, in both Latin and French to ensure that they would circulate widely. At Cambridge, Richard Porson worked on the missing lower right corner of the Greek text. He produced a skilful suggested reconstruction, which was soon being circulated by the Society of Antiquaries alongside its prints of the inscription. At almost the same moment, Christian Gottlob Heyne in Göttingen was making a new Latin translation of the Greek text that was more reliable than Ameilhon's and was first published in 1803. It was reprinted by the Society of Antiquaries in a special issue of its journal "Archaeologia" in 1811, alongside Weston's previously unpublished English translation, Colonel Turner's narrative, and other documents. At the time of the stone's discovery, Swedish diplomat and scholar Johan David Åkerblad was working on a little-known script of which some examples had recently been found in Egypt, which came to be known as demotic. He called it "cursive Coptic" because he was convinced that it was used to record some form of the Coptic language (the direct descendant of Ancient Egyptian), although it had few similarities with the later Coptic script. French Orientalist Antoine-Isaac Silvestre de Sacy had been discussing this work with Åkerblad when he received one of the early lithographic prints of the Rosetta Stone in 1801 from Jean-Antoine Chaptal, French minister of the interior. He realised that the middle text was in this same script. He and Åkerblad set to work, both focusing on the middle text and assuming that the script was alphabetical. They attempted to identify the points where Greek names ought to occur within this unknown text, by comparing it with the Greek. In 1802, Silvestre de Sacy reported to Chaptal that he had successfully identified five names (""Alexandros"", ""Alexandreia"", ""Ptolemaios"", ""Arsinoe"", and Ptolemy's title ""Epiphanes""), while Åkerblad published an alphabet of 29 letters (more than half of which were correct) that he had identified from the Greek names in the demotic text. They could not, however, identify the remaining characters in the demotic text, which, as is now known, included ideographic and other symbols alongside the phonetic ones. Silvestre de Sacy eventually gave up work on the stone, but he was to make another contribution. In 1811, prompted by discussions with a Chinese student about Chinese script, Silvestre de Sacy considered a suggestion made by Georg Zoëga in 1797 that the foreign names in Egyptian hieroglyphic inscriptions might be written phonetically; he also recalled that as early as 1761, Jean-Jacques Barthélemy had suggested that the characters enclosed in cartouches in hieroglyphic inscriptions were proper names. Thus, when Thomas Young, foreign secretary of the Royal Society of London, wrote to him about the stone in 1814, Silvestre de Sacy suggested in reply that in attempting to read the hieroglyphic text, Young might look for cartouches that ought to contain Greek names and try to identify phonetic characters in them. Young did so, with two results that together paved the way for the final decipherment. In the hieroglyphic text, he discovered the phonetic characters ""p t o l m e s"" (in today's transliteration ""p t w l m y s"") that were used to write the Greek name ""Ptolemaios"". He also noticed that these characters resembled the equivalent ones in the demotic script, and went on to note as many as 80 similarities between the hieroglyphic and demotic texts on the stone, an important discovery because the two scripts were previously thought to be entirely different from one another. This led him to deduce correctly that the demotic script was only partly phonetic, also consisting of ideographic characters derived from hieroglyphs. Young's new insights were prominent in the long article "Egypt" that he contributed to the "Encyclopædia Britannica" in 1819. He could make no further progress, however. In 1814, Young first exchanged correspondence about the stone with Jean-François Champollion, a teacher at Grenoble who had produced a scholarly work on ancient Egypt. Champollion saw copies of the brief hieroglyphic and Greek inscriptions of the Philae obelisk in 1822, on which William John Bankes had tentatively noted the names "Ptolemaios" and "Kleopatra" in both languages. From this, Champollion identified the phonetic characters "k l e o p a t r a" (in today's transliteration "q l i҆ w p 3 d r 3.t"). On the basis of this and the foreign names on the Rosetta Stone, he quickly constructed an alphabet of phonetic hieroglyphic characters, completing his work on 14 September and announcing it publicly on 27 September in a lecture to the "Académie royale des Inscriptions et Belles-Lettres". On the same day he wrote the famous ""Lettre à M. Dacier"" to Bon-Joseph Dacier, secretary of the Académie, detailing his discovery. In the postscript Champollion notes that similar phonetic characters seemed to occur in both Greek and Egyptian names, a hypothesis confirmed in 1823, when he identified the names of pharaohs Ramesses and Thutmose written in cartouches at Abu Simbel. These far older hieroglyphic inscriptions had been copied by Bankes and sent to Champollion by Jean-Nicolas Huyot. From this point, the stories of the Rosetta Stone and the decipherment of Egyptian hieroglyphs diverge, as Champollion drew on many other texts to develop an Ancient Egyptian grammar and a hieroglyphic dictionary which were published after his death in 1832. Work on the stone now focused on fuller understanding of the texts and their contexts by comparing the three versions with one another. In 1824 Classical scholar Antoine-Jean Letronne promised to prepare a new literal translation of the Greek text for Champollion's use. Champollion in return promised an analysis of all the points at which the three texts seemed to differ. Following Champollion's sudden death in 1832, his draft of this analysis could not be found, and Letronne's work stalled. François Salvolini, Champollion's former student and assistant, died in 1838, and this analysis and other missing drafts were found among his papers. This discovery incidentally demonstrated that Salvolini's own publication on the stone, published in 1837, was plagiarism. Letronne was at last able to complete his commentary on the Greek text and his new French translation of it, which appeared in 1841. During the early 1850s, German Egyptologists Heinrich Brugsch and Max Uhlemann produced revised Latin translations based on the demotic and hieroglyphic texts. The first English translation followed in 1858, the work of three members of the Philomathean Society at the University of Pennsylvania. Whether one of the three texts was the standard version, from which the other two were originally translated, is a question that has remained controversial. Letronne attempted to show in 1841 that the Greek version, the product of the Egyptian government under the Macedonian Ptolemies, was the original. Among recent authors, John Ray has stated that "the hieroglyphs were the most important of the scripts on the stone: they were there for the gods to read, and the more learned of their priesthood". Philippe Derchain and Heinz Josef Thissen have argued that all three versions were composed simultaneously, while Stephen Quirke sees in the decree "an intricate coalescence of three vital textual traditions". Richard Parkinson points out that the hieroglyphic version strays from archaic formalism and occasionally lapses into language closer to that of the demotic register that the priests more commonly used in everyday life. The fact that the three versions cannot be matched word for word helps to explain why the decipherment has been more difficult than originally expected, especially for those original scholars who were expecting an exact bilingual key to Egyptian hieroglyphs. Even before the Salvolini affair, disputes over precedence and plagiarism punctuated the decipherment story. Thomas Young's work is acknowledged in Champollion's 1822 "Lettre à M. Dacier", but incompletely, according to early British critics: for example, James Browne, a sub-editor on the "Encyclopædia Britannica" (which had published Young's 1819 article), anonymously contributed a series of review articles to the "Edinburgh Review" in 1823, praising Young's work highly and alleging that the "unscrupulous" Champollion plagiarised it. These articles were translated into French by Julius Klaproth and published in book form in 1827. Young's own 1823 publication reasserted the contribution that he had made. The early deaths of Young (1829) and Champollion (1832) did not put an end to these disputes. In his work on the stone in 1904 E. A. Wallis Budge gave special emphasis to Young's contribution compared with Champollion's. In the early 1970s, French visitors complained that the portrait of Champollion was smaller than one of Young on an adjacent information panel; English visitors complained that the opposite was true. The portraits were in fact the same size. Calls for the Rosetta Stone to be returned to Egypt were made in July 2003 by Zahi Hawass, then Secretary-General of Egypt's Supreme Council of Antiquities. These calls, expressed in the Egyptian and international media, asked that the stele be repatriated to Egypt, commenting that it was the "icon of our Egyptian identity". He repeated the proposal two years later in Paris, listing the stone as one of several key items belonging to Egypt's cultural heritage, a list which also included: the iconic bust of Nefertiti in the Egyptian Museum of Berlin; a statue of the Great Pyramid architect Hemiunu in the Roemer-und-Pelizaeus-Museum in Hildesheim, Germany; the Dendera Temple Zodiac in the Louvre in Paris; and the bust of Ankhhaf in the Museum of Fine Arts in Boston. In 2005, the British Museum presented Egypt with a full-sized fibreglass colour-matched replica of the stele. This was initially displayed in the renovated Rashid National Museum, an Ottoman house in the town of Rashid (Rosetta), the closest city to the site where the stone was found. In November 2005, Hawass suggested a three-month loan of the Rosetta Stone, while reiterating the eventual goal of a permanent return. In December 2009, he proposed to drop his claim for the permanent return of the Rosetta Stone if the British Museum lent the stone to Egypt for three months for the opening of the Grand Egyptian Museum at Giza in 2013. As John Ray has observed, "the day may come when the stone has spent longer in the British Museum than it ever did in Rosetta." There is strong opposition among national museums to the repatriation of objects of international cultural significance such as the Rosetta Stone. In response to repeated Greek requests for return of the Elgin Marbles from the Parthenon and similar requests to other museums around the world, in 2002 over 30 of the world's leading museums—including the British Museum, the Louvre, the Pergamon Museum in Berlin and the Metropolitan Museum in New York City—issued a joint statement declaring that "objects acquired in earlier times must be viewed in the light of different sensitivities and values reflective of that earlier era" and that "museums serve not just the citizens of one nation but the people of every nation". Various ancient bilingual or even trilingual epigraphical documents have sometimes been described as "Rosetta stones", as they permitted the decipherment of ancient written scripts. For example, the bilingual Greek-Brahmi coins of the Greco-Bactrian king Agathocles have been described as "little Rosetta stones", allowing to secure the first steps towards the decipherment of the Brahmi script by Christian Lassen, thus unlocking ancient Indian epigraphy. The Behistun inscription has also been compared to the Rosetta stone, as it links the translations of three ancient Middle-Eastern languages: Old Persian, Elamite, and Babylonian. The term "Rosetta stone" has been also used idiomatically to represent a crucial key in the process of decryption of encoded information, especially when a small but representative sample is recognised as the clue to understanding a larger whole. According to the "Oxford English Dictionary", the first figurative use of the term appeared in the 1902 edition of the "Encyclopædia Britannica" relating to an entry on the chemical analysis of glucose. Another use of the phrase is found in H. G. Wells' 1933 novel "The Shape of Things to Come", where the protagonist finds a manuscript written in shorthand that provides a key to understanding additional scattered material that is sketched out in both longhand and on typewriter. Since then, the term has been widely used in other contexts. For example, Nobel laureate Theodor W. Hänsch in a 1979 "Scientific American" article on spectroscopy wrote that "the spectrum of the hydrogen atoms has proved to be the Rosetta Stone of modern physics: once this pattern of lines had been deciphered much else could also be understood". Fully understanding the key set of genes to the human leucocyte antigen has been described as "the Rosetta Stone of immunology". The flowering plant "Arabidopsis thaliana" has been called the "Rosetta Stone of flowering time". A Gamma ray burst (GRB) found in conjunction with a supernova has been called a Rosetta Stone for understanding the origin of GRBs. The technique of Doppler echocardiography has been called a Rosetta Stone for clinicians trying to understand the complex process by which the left ventricle of the human heart can be filled during various forms of diastolic dysfunction. The name has also become used in various forms of translation software. Rosetta Stone is a brand of language-learning software published by Rosetta Stone Inc., headquartered in Arlington County, US. "Rosetta" is the name of a "lightweight dynamic translator" that enables applications compiled for PowerPC processors to run on Apple Inc. systems using an x86 processor. It was later used for applications compiled for the Intel instruction set to be run on Macs built with Apple's ARM CPUs. "Rosetta" is an online language translation tool to help localisation of software, developed and maintained by Canonical as part of the Launchpad project. Similarly, Rosetta@home is a distributed computing project for predicting protein structures from amino acid sequences (or "translating" sequence into structure). The Rosetta Project brings language specialists and native speakers together to develop a meaningful survey and near-permanent archive of 1,500 languages, intended to last from AD 2000 to 12,000. The European Space Agency's "Rosetta" spacecraft was launched to study the comet 67P/Churyumov–Gerasimenko in the hope that determining its composition will reveal the origin of the Solar System.
https://en.wikipedia.org/wiki?curid=26428
Redshirt (stock character) A "redshirt" is a stock character in fiction who dies soon after being introduced. The term originates from the original "" (NBC, 1966–69) television series in which the red-shirted security personnel frequently die during episodes. Redshirt deaths are often used to dramatize the potential peril that the main characters face. In "Star Trek", red-uniformed security officers and engineers who accompany the main characters on landing parties often suffer quick deaths. The first instance of what now is an established trope can be seen in the episode "What Are Little Girls Made Of?" (1966). Of the 59 crew members killed in the series, 43 (73%) were wearing red shirts. The "" book "Legends of the Ferengi" says Starfleet security personnel "rarely survive beyond the second act break". An episode of "" titled "" (1998) also references red as a sort of bad luck omen, in which the plot centers around a group of cadets calling themselves "Red Squad", almost all of whom die in the episode. The cinematic reboot of the franchise features a character named Olson (portrayed by Greg Ellis) who dies early on during a mission; he wears a red uniform in homage to the trope from the original series. In other media, the term "redshirt" and images of characters wearing red shirts has come to represent disposable characters destined for suffering or death. The trope, and its particular usage in "Star Trek", has been parodied and deconstructed in other media. Parodies include "Galaxy Quest" (1999), a comedy about actors from a defunct science-fiction television series serving on a real starship, which includes an actor who is terrified that he's going to die because his only appearance in the show was as an unnamed character who was killed early in the episode. "Redshirts" is a novel by John Scalzi that satirizes the trope, as does the PC game "Redshirt".
https://en.wikipedia.org/wiki?curid=26429
Resolution-class submarine The "Resolution" class was a class of four nuclear ballistic missile submarines (SSBN) built for the Royal Navy as part of the UK Polaris programme. Each submarine was armed with up to 16 UGM-27 Polaris A-3 nuclear missiles. The class comprised , , and . They were built by Vickers Armstrong in Barrow-in-Furness and Cammell Laird in Birkenhead between 1964 and 1968. All four boats were based at HM Naval Base Clyde (HMS "Neptune"), west of Glasgow, Scotland. The "Resolution" class was the launch platform for the United Kingdom's strategic nuclear deterrent from the late 1960s until 1996, when it was replaced by the carrying the Trident II. During the 1950s and early 1960s, the United Kingdom's nuclear deterrent was based on the RAF's V-bombers. But in the early 1960s developments in radar and surface-to-air missiles made it clear that bombers were becoming vulnerable, and would be unlikely to penetrate Soviet airspace. Free-fall nuclear weapons would no longer be a credible deterrent. To address this problem, in May 1960 the British Prime Minister, Harold Macmillan arranged a deal with US President Eisenhower to equip the V bombers with the US-designed AGM-48 Skybolt. The Skybolt was a range ballistic missile that allowed the launching bombers to remain well away from Soviet defences and launch attacks that would be basically invulnerable. With this range, the V bombers would have to fly only a few hundred miles from their bases before being in range for an attack on Moscow. Under the agreement the UK's contribution to the programme was limited to developing suitable mounting points on the Avro Vulcan bomber, installing the required guidance systems that fed the missiles updated positioning information, and development of a British version of the US W47 warhead to arm it, the RE.179 . The incoming Kennedy administration expressed serious doubts of both Skybolt and the US deterrent force in general. Robert McNamara was highly critical of the US bomber fleet, which he saw as obsolete in an age of ICBMs. Skybolt was seen simply as a means of continuing the existence of a system he no longer considered credible, and given the rapidly improving capabilities of ICBM inertial guidance systems, a precision strike capability with free-fall bombs would no longer be needed. McNamara was equally concerned about the UK also having its own nuclear force, and worried that the US could be drawn into a war by the UK. He wanted to bring the UK into a dual-key arrangement. McNamara first broached the idea of cancelling Skybolt with the British in November 1962. When this was reported in the House of Commons, a storm of protest broke out. A meeting was arranged to settle the issue, and Macmillan stated in no uncertain terms that the UK would be retaining their independent deterrent capability, no matter what the cost. With development of their Polaris-derived warheads well along, a suitable launch platform would be developed, if need be. Faced with a clear failure in policy terms, Kennedy gave up on the idea of strong-arming Britain into accepting a dual-key arrangement. By the end of the series of meetings, the UK had gained the much more impressive Polaris system, and would start development of a new submarine to launch it. The SSBNs would then take over the nuclear deterrent role from the RAF's V bombers from 1968 onwards. Two pairs of the boats were ordered in May 1963 from Vickers Shipbuilding Ltd, Barrow in Furness and from Cammell Laird and Co. Ltd, Birkenhead. The option of buying a fifth unit, planned as , was cancelled in February 1965 as a cost-savings measure by the government of Prime Minister Harold Wilson. Traditional battleship names were used, signifying that they were the capital ships of their time. Vickers Armstrong in Barrow-in-Furness constructed "Resolution" and "Repulse" and Cammell Laird in Birkenhead constructed "Renown" and "Revenge". The construction was unusual in that the bow and stern were constructed separately before being assembled together with the American-designed missile compartment. The design was a modification of the fleet submarine, but greatly extended to incorporate the missile compartment between the fin and the nuclear reactor. The length was , breadth , height and the displacement submerged and surfaced. A Rolls-Royce pressurised water reactor (PWR1) and English Electric Company turbines gave them a speed of and they could dive to depths of . Sixteen Polaris A3 missiles were carried, in two rows of eight. For emergencies there was a diesel generator and six torpedo tubes located at the bow, firing the Tigerfish wire-guided homing torpedoes. The submarines put to sea with a crew of 143. According to former head of the Royal Corps of Naval Constructors R.J. Daniel, the "Resolution"-class SSBNs possessed five features that were envied by the United States Navy: the machinery loading hatch, automated hovering system, welded hull valves, standardised valves, and raft-mounted propulsion machinery. The first to be completed was "Resolution", laid down in February 1964 and launched in September 1966. After commissioning in 1967 she underwent a long period of sea trials, culminating in the test firing of a Polaris missile from the USAF Eastern Test Range off Cape Kennedy at 11:15 on 15 February 1968. "Resolution" commenced her first operational patrol on 15 June 1968, beginning 28 years of Polaris patrols. The class were part of the 10th Submarine Squadron, all based at Faslane Naval Base, Scotland. All four of the class underwent conversion during the 1980s so that they could be fitted with the Polaris A3TK missile which was fitted with the British-developed Chevaline MRV system. As the newer "Vanguard"-class submarines entered service, the "Resolution" class was eventually retired and all boats laid up at Rosyth dockyard with their used nuclear fuel removed. All four will eventually be disposed of via MOD's Submarine Dismantling Project (SDP). This project will begin in 2016 with as the first submarine to prove the technique. The selected method will first remove all Low-level radioactive waste from the vessel, followed by the more radioactive intermediate-level waste. All non-radioactive material in the remainder of the vessel will be recycled for re-use by conventional ship-breaking techniques. New methods of project management were used in the refits of the "Resolution" class, including:
https://en.wikipedia.org/wiki?curid=26432
Roger Angell Roger Sergeant Angell (born September 19, 1920) is an American essayist known for his writing on sports, especially baseball. He has been a regular contributor to "The New Yorker" and was its chief fiction editor for many years. He has written numerous works of fiction, non-fiction, and criticism, and for many years wrote an annual Christmas poem for "The New Yorker". He received a number of awards for his writing, including the George Polk Award for Commentary in 1980, the "Kenyon Review" Award for Literary Achievement in 2005 along with Umberto Eco, and the inaugural PEN/ESPN Lifetime Achievement Award for Literary Sports Writing in 2011. He was elected a Fellow of the American Academy of Arts and Sciences in 2007 and is a long-time ex-officio member of the council of the Authors Guild. Angell was inducted into the Baseball Reliquary's Shrine of the Eternals in 2010. He was named the 2014 recipient of the J. G. Taylor Spink Award by the Baseball Writers' Association of America on December 10, 2013. Angell is the son of Katharine Sergeant Angell White, "The New Yorker"’s first fiction editor, and the stepson of renowned essayist E. B. White, but was raised for the most part by his father, Ernest Angell, an attorney who became head of the American Civil Liberties Union. Angell is a 1938 graduate of the Pomfret School and attended Harvard University. He served in the United States Army Air Forces during World War II. Angell's earliest published works were pieces of short fiction and personal narratives. Several of these pieces were collected in "The Stone Arbor and Other Stories" (1960) and "A Day in the Life of Roger Angell" (1970). He first contributed to "The New Yorker" in March 1944 and continued to contribute into 2018. In 1948, Angell was employed at "Holiday Magazine", a travel magazine that featured literary writers. He first wrote professionally about baseball in 1962, when William Shawn, editor of "The New Yorker", had him travel to Florida to write about spring training. Angell has been called the "Poet Laureate of baseball" but dislikes the term. In a review of "Once More Around the Park" for the "Journal of Sport History", Richard C. Crepeau wrote that "Gone for Good", Angell's essay on the career of Steve Blass, "may be the best piece that anyone has ever written on baseball or any other sport". Angell contributed commentary to the Ken Burns series "Baseball", in 1994. Angell has three children: Callie, Alice, and John Henry. He had Alice and Callie with his first wife Evelyn, and John Henry with Carol. Callie Angell, who was an authority on the films of Andy Warhol, committed suicide on May 5, 2010, in Manhattan, where she worked as a curator at the Whitney Museum of American Art; she was 62. In a 2014 essay, Angell mentioned her death – "the oceanic force and mystery of that event" – and his struggle to comprehend that "a beautiful daughter of mine, my oldest child, had ended her life." Alice Angell lived in Portland, Maine and died from cancer on February 2, 2019, and John Henry Angell lives in Portland, Oregon. His second wife, Carol Rogge Angell, to whom he was married for 48 years, died on April 10, 2012, of metastatic breast cancer at the age of 73. In 2014, he married Margaret Moorman, a writer and teacher, as noted in the Ellsworth American newspaper. In 2019, University of Nebraska Press published "No Place I Would Rather Be: Roger Angell and a Life in Baseball Writing", a book about Angell's career written by Joe Bonomo.
https://en.wikipedia.org/wiki?curid=26435
Rent (musical) Rent (stylized as RENT) is a rock musical with music, lyrics, and book by Jonathan Larson, loosely based on Giacomo Puccini's 1896 opera "La Bohème". It tells the story of a group of impoverished young artists struggling to survive and create a life in Lower Manhattan's East Village in the thriving days of bohemian Alphabet City, under the shadow of HIV/AIDS. The musical was first seen in a workshop production at New York Theatre Workshop in 1993. This same off-Broadway theatre was also the musical's initial home following its official 1996 opening. The show's creator, Jonathan Larson, died suddenly of an aortic dissection, believed to have been caused by undiagnosed Marfan syndrome, the night before the off-Broadway premiere. The musical moved to Broadway's larger Nederlander Theatre on April 29, 1996. On Broadway, "Rent" gained critical acclaim and won several awards, including the Pulitzer Prize for Drama and the Tony Award for Best Musical. The Broadway production closed on September 7, 2008, after 12 years, making it one of the longest-running shows on Broadway. The production grossed over $280 million. The success of the show led to several national tours and numerous foreign productions. In 2005, it was adapted into a motion picture featuring most of the original cast members. In 1988, playwright Billy Aronson wanted to create "a musical based on Puccini's "La Bohème", in which the luscious splendor of Puccini's world would be replaced with the coarseness and noise of modern New York." In 1989, Jonathan Larson, a 29-year-old composer, began collaborating with Aronson on this project, and the two composed together "Santa Fe", "Splatter" (later re-worked into the song "Rent"), and "I Should Tell You". Larson suggested setting the play "amid poverty, homelessness, spunky gay life, drag queens and punk" in the East Village neighborhood of Manhattan, which happened to be down the street from his Greenwich Village apartment. He also came up with the show's ultimate title (a decision that Aronson was unhappy with, at least until Larson pointed out that "rent" also means torn apart). In 1991, he asked Aronson if he could use Aronson's original concept and make "Rent" his own. Larson had ambitious expectations for "Rent"; his ultimate dream was to write a rock opera "to bring musical theater to the MTV generation". Aronson and Larson made an agreement that if the show went to Broadway, Aronson would share in the proceeds and be given credit for "original concept & additional lyrics". Jonathan Larson focused on composing "Rent" in the early 1990s, waiting tables at the Moondance Diner to support himself. Over the course of years, Larson wrote hundreds of songs and made many drastic changes to the show, which in its final incarnation contained 42 songs. In the fall of 1992, Larson approached James Nicola, artistic director of New York Theatre Workshop, with a tape and copy of "Rent"s script. When "Rent" had its first staged reading at New York Theatre Workshop in March 1993, it became evident that, despite its very promising material and moving musical numbers, many structural problems needed to be addressed, including its cumbersome length and overly complex plot. As of 1994, the New York Theatre Workshop version of "Rent" featured songs that never made it into the final version, such as: This workshop version of "Rent" starred Anthony Rapp as Mark and Daphne Rubin-Vega as Mimi. Larson continued to work on "Rent", gradually reworking its flaws and staging more workshop productions. On January 24, 1996, after the musical's final dress rehearsal before its off-Broadway opening, Larson had his first (and only) newspaper interview with music critic Anthony Tommasini of "The New York Times", attracted by the coincidence that the show was debuting exactly 100 years after Puccini's opera. Larson would not live to see "Rent"s success; he died from an undiagnosed aortic aneurysm (believed to have resulted from Marfan syndrome) in the early morning of January 25, 1996. Friends and family gathered at the New York Theatre Workshop, and the first preview of "Rent" became a sing-through of the musical in Larson's memory. The show premiered as planned and quickly gained popularity fueled by enthusiastic reviews and the recent death of its composer. It proved extremely successful during its off-Broadway run, selling out all its shows at the 150-seat New York Theater Workshop. Due to such overwhelming popularity and a need for a larger theater, "Rent" moved to Broadway's then-under-renovation Nederlander Theatre on 41st Street on April 29, 1996. At the production's request, final touches on the theater's remodeling and renovation were put on hold before and during the run of "Rent" because the show's producers and creative team felt the unfinished look fit in well with the gritty setting of the show. Larson's inspiration for "Rent"s content came from several different sources. Many of the characters and plot elements are drawn directly from Giacomo Puccini's opera "La Bohème", the world premiere of which was in 1896, a century before "Rent"s premiere. "La Bohème" was also about the lives of poor young artists. Tuberculosis, the plague of Puccini's opera, is replaced by HIV/AIDS in "Rent"; 1800s Paris is replaced by New York's East Village in the late 1980s or early 1990s. The names and identities of "Rent"s characters also heavily reflect Puccini's original characters, though they are not all direct adaptations. For example, Joanne in "Rent" represents the character of Alcindoro in "Bohème", but is also partially based on Marcello. Also, Joanne is the only "Rent" character whose predecessor in "La Bohème" is a different sex. Other examples of parallels between Larson's and Puccini's work include Larson's song "Light My Candle", which draws melodic content directly from "Che gelida manina"; "Quando me'n vo'" ("Musetta's Waltz"), a melody taken directly from Puccini's opera; and "Goodbye Love", a long, painful piece that reflects a confrontation and parting between characters in both Puccini's and Larson's work. "Quando me'n vo'" is paralleled in the first verse of "Take Me or Leave Me", when Maureen describes the way people stare when she walks in the street. It is also directly referred to in the scene where the characters are celebrating their bohemian life. Mark says, "Roger will attempt to write a bittersweet, evocative song..." Roger plays a quick piece, and Mark adds, "...that "doesn't" remind us of 'Musetta's Waltz'." This part of "Musetta's Waltz" is also later used in "Your Eyes", a song Roger writes. "Rent" is also a somewhat autobiographical work, as Larson incorporated many elements of his life into his show. Larson lived in New York for many years as a starving artist with an uncertain future. He sacrificed a life of stability for his art, and shared many of the same hopes and fears as his characters. Like his characters he endured poor living conditions, and some of these conditions (e.g. illegal wood-burning stove, bathtub in the middle of his kitchen, broken buzzer [his guests had to call from the pay phone across the street and he would throw down the keys, as in "Rent"]) made their way into the play. Part of the motivation behind the storyline in which Maureen leaves Mark for a woman (Joanne) is based on the fact that Larson's own girlfriend left him for a woman. The Mark Cohen character is based on Larson's friends, cinematographer and producer Jonathan Burkhart and documentary filmmaker Eddie Rosenstein. Playwright Sarah Schulman alleged that "Rent" bore striking similarities to her novel "People in Trouble". The line, "I'm more of a man than you'll ever be... and more of a woman than you'll ever get!", attributed to Angel Dumott Schunard at her funeral, was previously used by the character Hollywood Montrose, who appeared in the films "Mannequin" (1987) and "" (1991). Like Angel, Hollywood performs a song and dance number and sometimes wears women's clothing. This line was originally in the film "Car Wash" (1976), delivered by Antonio Fargas as a flamboyant homosexual cross dresser. The earliest concepts of the characters differ largely from the finished products. Everyone except Mark had AIDS, including Maureen and Joanne; Maureen was a serious, angry character who played off Oedipus in her performance piece instead of Hey Diddle Diddle; Mark was, at one point, a painter instead of a filmmaker; Roger was named Ralph and wrote musical plays; Angel was a jazz philosopher, while Collins was a street performer; Angel and Collins were both originally described as Caucasian; and Benny had a somewhat enlarged role in the story, taking part in songs like "Real Estate", which was later cut. Many actual locations and events are included in, or are the inspiration for, elements of the musical. Life Café, where the "La Vie Bohème" numbers are set, was an actual restaurant (closed 2013) on 10th Street and Avenue B in the East Village of New York City. The riot at the end of the first act is based on the East Village riot in 1988 that arose as a result of the city-imposed curfew in Tompkins Square Park. "Will I?", a song which takes place during a Life Support meeting and expresses the pain and fear of living a life with AIDS, was inspired by a real event. Larson attended a meeting of Friends in Deed, an organization that helps people deal with illness and grief, much like Life Support. After that first time, Larson attended the meetings regularly. During one meeting, a man stood up and said that he was not afraid of dying. He did say, however, that there was one thing of which he was afraid: Would he lose his dignity? From this question stemmed the first line of this song. The people present at the Life Support meeting in the show, such as Gordon, Ali and Pam, carry the names of Larson's friends who died. In the Broadway show, the names of the characters in that particular scene (they introduce themselves) were changed nightly to honor the friends of the cast members who were living with or had died from AIDS. The scene and song "Life Support" were also based on Friends in Deed, as well as on Gordon, Pam, and Ali. Originally, the members of Life Support had a solid block of the "forget regret" refrain, and they talked about remembering love. When Jonathan's HIV positive friends heard this scene, they told him that having AIDS was not so easy to accept: it made you angry and resentful too, and the song did not match that. Jonathan then added a part where Gordon says that he has a problem with this "credo...my T-cells are low, I regret that news, okay?" Paul, the leader of the meeting, replies, "Okay...but, Gordon, how do you feel today?" Gordon admits that he is feeling the best that he has felt all year. Paul asks, "Then why choose fear?" Gordon says, "I'm a New Yorker. Fear's my life." Lynn Thomson was a dramaturg who was hired by New York Theatre Workshop to help rework "Rent". She claimed that between early May and the end of October 1995, she and Larson co-wrote a "new version" of the musical. She sued the Larson estate for $40 million and sought 16 percent of the show's royalties, claiming she had written a significant portion of the lyrics and the libretto of the "new version" of "Rent". During the trial, Thomson could not recall the lyrics to the songs that she allegedly wrote, nor the structures of the libretto she claimed to have created. The judge ruled against her and gave the Jonathan Larson Estate full credit and right to "Rent". A federal appellate court upheld the original ruling on appeal. In August 1998, the case was settled out of court. The terms of the settlement were not disclosed. On Christmas Eve in Manhattan's East Village, two roommates—Mark, a filmmaker, and Roger, a rock musician—struggle to stay warm and produce their art ("Tune Up #1"). Mark's mother leaves him a voicemail wishing him a merry Christmas and trying to comfort him since his ex-girlfriend Maureen dumped him ("Voice Mail #1"). Their friend Tom Collins, a gay anarchist professor of computer-age philosophy at New York University, calls and plans to surprise them at their apartment, but is mugged before entering. At the same time, Mark and Roger's former roommate and friend Benny, who has since become their harsh new landlord, has reneged on an earlier agreement and now demands last year's rent, before shutting down their electrical power ("Tune Up #2"). However, Mark and Roger rebel and resolve not to pay the rent they cannot pay and which they were promised would not be a problem ("Rent"). Meanwhile, Angel, a cross-dressing street drummer (presently out of drag), finds Collins wounded in an alley and tends to him ("You Okay Honey?") - the two are immediately attracted to each other, each learning that the other is HIV positive. It is revealed that Roger also has HIV, which he contracted from his last girlfriend, who committed suicide after learning of her diagnosis, causing Roger to fall into depression. Mark leaves the loft while Roger stays home ("Tune Up #3"), trying to compose on his guitar without success; he wishes desperately to write one last song to be remembered by before he dies ("One Song Glory"). An exotic dancer, junkie, and neighbor, Mimi, shows up at their apartment asking for help with lighting her candle, flirting with Roger in the process; however, he is clearly hesitant to return her affections ("Light My Candle"). Meanwhile, Joanne, a lawyer and Maureen's girlfriend, receives a voicemail from her parents ("Voice Mail #2"). At last, the missing Collins enters the apartment, presenting Angel, who is now in full drag and shares the money she made and the amusing story of how she killed a dog to earn it ("Today 4 U"). Mark comes home, and Benny arrives, speaking of Maureen's upcoming protest against his plans to evict the homeless from a lot where he is hoping to build a cyber arts studio. Benny offers that, if they convince Maureen to cancel the protest, then Mark and Roger can officially remain rent-free tenants. However, the two rebuff Benny's offer and he leaves ("You'll See"). Mark leaves the loft again to go help Maureen with the sound equipment for the protest, unexpectedly meeting Joanne at the stage. Initially hesitant with each other, the two eventually bond over their shared distrust of Maureen's "gaslighting" and promiscuous behaviours ("Tango: Maureen"). Mark then joins Collins and Angel to film their HIV support group meeting ("Life Support"), while Mimi attempts to seduce Roger alone in his apartment ("Out Tonight"). Roger is extremely upset by Mimi's intrusion, demanding she leave him alone and resisting any romantic feelings he may harbour for her ("Another Day"). After Mimi leaves, Roger reflects on his fear of dying an undignified death from AIDS, while the Life Support group echoes his thoughts ("Will I"). Collins, Mark, and Angel protect a homeless woman from police harassment, but she chastises them ("On the Street"). To lighten the mood, Collins talks about his dream of escaping New York City to open a restaurant in Santa Fe ("Santa Fe"). Soon, Mark leaves to check up on Roger; while alone, Collins and Angel confess their love for each other ("I'll Cover You"). Joanne hectically prepares for Maureen's show, trying to balance all of the people calling her at once ("We're Okay"). Before the performance, Roger apologizes to Mimi, inviting her to come to the protest and the dinner party his friends are having afterwards. At the same time, police, vendors, and homeless people prepare for the protest ("Christmas Bells"). Maureen begins her avant-garde, if not over the top, performance based on "Hey Diddle Diddle" ("Over the Moon"). At the post-show party at the Life Café, Benny arrives, criticizing the protest and the group's bohemian lifestyle. In response, Mark and all the café's bohemian patrons defiantly rise up to celebrate their way of living ("La Vie Bohème"). Mimi and Roger each discover that the other is HIV-positive and hesitantly decide to move forward with their relationship ("I Should Tell You"). Joanne explains that Mark and Roger's building has been padlocked and a riot has broken out; Roger and Mimi, unaware, share their first kiss. The celebration continues ("La Vie Bohème B"). The cast lines up to sing together before the plot of the second act begins, affirming that one should measure life "in love" ("Seasons of Love"). Afterwards, Mark and Roger gather to break back into their locked apartment with their friends ("Happy New Year"). A new voicemail reveals that Mark's footage of the riot has earned him a job offering at a tabloid news company called Buzzline ("Voice Mail #3"). The others finally break through the door just as Benny arrives, saying he wants to call a truce and revealing that Mimi––who used to be his girlfriend––convinced him to change his mind. Mimi denies rekindling her relationship with Benny, but Roger is upset, and although they apologize to each other, Mimi goes to her drug dealer for a fix ("Happy New Year B"). Around Valentine's Day, Mark tells the audience that Roger and Mimi have been living together, but they are tentative with each other. It is also told that Maureen and Joanne are preparing another protest; during rehearsal, Maureen criticizes Joanne's controlling behaviour, and Joanne criticizes Maureen's promiscuous mannerisms. They break up dramatically following an ultimatum ("Take Me or Leave Me"). Time progresses to spring ("Seasons of Love B"), but Roger and Mimi's relationship is strained by Mimi's escalating heroin usage and Roger's lasting jealousy and suspicion of Benny. Each alone, Roger and Mimi sing of love and loneliness, telling each other how they feel, as they watch Collins nurse Angel, whose health is declining due to AIDS ("Without You"). By the end of the summer, Mark continues to receive calls offering a corporate job at Buzzline ("Voice Mail #4"). A dance is performed representing all the couples' sex lives ("Contact"). At the climax of the number, the two former couples break up, and Angel suddenly dies. At the funeral, the friends briefly come together to share their memories, with Collins being the last to reminisce ("I'll Cover You [Reprise]"). Mark expresses his fear of being the only one left surviving when the rest of his friends die of AIDS, and he finally accepts the corporate job offer ("Halloween"). Roger reveals that he is leaving for Santa Fe, which sparks an argument about commitment between him and Mimi, and between Maureen and Joanne. Collins arrives and admonishes the entire group for fighting on the day of Angel's funeral, causing Maureen and Joanne to reconcile, but not Mimi and Roger. The group shares a sad moment, knowing that between deaths and leaving, their close-knit friendships will be breaking up. Everyone leaves except Mark and Roger, and so Mark tries to convince Roger to stay in New York. Roger, unable to handle Mimi's declining health, becomes angry with Mark and leaves. Mimi returns to say goodbye, overhears everything Roger says, and, terrified, agrees to go to rehab, which Benny pays for ("Goodbye Love"). Collins is forcibly removed from the church for being unable to pay for Angel's funeral; Benny shows compassion by paying and offering Mark and Collins drinks. Collins accepts, causing him and Collins to rekindle their old friendship, but Mark has to turn down the offer due to work commitments. Some time later, both Mark and Roger simultaneously reach an artistic epiphany, as Roger finds his song in Mimi and Mark finds his film in Angel's memory; Roger decides to return to New York in time for Christmas, while Mark quits his job to devote his efforts to working on his own film ("What You Own"). The characters' parents, concerned and confused about their respective situations, leave several worried messages on their phones ("Voice Mail #5"). On Christmas Eve, exactly one year having passed, Mark prepares to screen his now-completed film to his friends. Roger has written his song, but no one can find Mimi for him to play it to. Benny's wife, discovering Benny's relationship with Mimi, has pulled Benny out of the East Village. The power suddenly blows and Collins enters with handfuls of cash, revealing that he reprogrammed an ATM at a grocery store to provide money to anybody with the code 'ANGEL'. Maureen and Joanne abruptly enter carrying Mimi, who had been homeless and is now weak and close to death. She begins to fade, telling Roger that she loves him ("Finale"). Roger tells her to hold on as he plays her the song he wrote for her, revealing the depth of his feelings for her ("Your Eyes"). Mimi appears to die, but abruptly awakens, claiming to have been heading into a white light before a vision of Angel appeared, telling her to go back and stay with Roger. The remaining friends gather together in a final moment of shared happiness and resolve to enjoy whatever time they have left with each other, affirming that there is "no day but today" ("Finale B"). Act 1 Act 2 There are also many other non-named roles such as Cops, Bohemians, Vendors, Homeless People. "Rent" received several awards including a Pulitzer Prize and four Tony Awards. Critical reception of "Rent" was positive not only for its acting and musical components, but for its representation of HIV-positive individuals. Many critics praised the portrayal of characters such as Angel and Collins as being happy, with positive outlooks on life, rather than being resigned to death. While critics and theatre patrons had largely positive reviews of the show, it was criticized for its stereotypically negative portrayal of lesbian characters and the "glamourization" of the East Village in the late 1980s. Billy Aronson said, "For the record, although I was ambivalent about Jonathan’s ideas for "Rent" when we were working together on it, I have come to love the show. And as tragic as it is that he didn’t live to see his work become a huge success, I believe he knew it would be. In our last conversation I asked how the show was going and he said, with complete assurance, that it was incredible." The song "Seasons of Love" became a successful pop song and often is performed on its own. Because of its connection to New Years and looking back at times past, it is sometimes performed at graduations or school holiday programs. "Rent" gathered a following of fans who refer to themselves as "RENT-heads." The name originally referred to people who would camp out at the Nederlander Theater for hours in advance for the discounted $20 rush tickets to each show, though it generally refers to anyone who is obsessed with the show. These discounted tickets were for seats in the first two rows of the theater reserved for sale by lottery two hours prior to each show. Other Broadway shows have followed "Rent"s example and now also offer cheaper tickets in efforts to make Broadway theater accessible to people who would otherwise be unable to afford the ticket prices. The term originated in "Rent"s first months on Broadway. The show's producers offered 34 seats in the front two rows of the orchestra for $20 each, two hours before the performance. Fans and others interested in tickets would camp out for hours in front of the Nederlander Theater – which is on 41st Street, just outside Times Square – to buy these tickets. The television series "The Simpsons", "Family Guy", "Friends", "Will and Grace", "Scrubs", "Glee", "The Big Bang Theory", "Gilmore Girls", "Felicity", "Saturday Night Live", "The Office", "Franklin & Bash", "2 Broke Girls", "Girls", "Seinfeld", "The Neighbors", "Modern Family", "Smash", "Supernatural", "Superstore", and "Bob's Burgers" have included references to the show. The film "" includes a character who plays a lead role in "Lease", a Broadway musical parody of "Rent"; the finale song is "Everyone has AIDS!". In 2017, the song "Out Tonight" was covered by Ashleigh Murray, Asha Bromfield, Hayley Law and Camila Mendes in the episode "Chapter Eighteen: When a Stranger Calls" from the second season of the television series "Riverdale". In the film "Deadpool", Wade Wilson is seen wearing a "Rent" T-shirt. Stan Lee also referenced one of the songs ("Cover you") when he said as the DJ in the strip club "You can't buy love.." - "but you can rent it... " "Rent" has also been referenced in other musicals. Yitzhak in "Hedwig and the Angry Inch" wears a "Rent" T-shirt and speaks of his aspiration to play the role of Angel. The off-Broadway musical revue "Forbidden Broadway Strikes Back" includes parodies of "Rent" songs such as "Rant" ("Rent"), "Ouch! They're Tight" ("Out Tonight"), "Season of Hype" ("Seasons of Love"), "Too Gay 4 U (Too Het'ro 4 Me)" ("Today 4 U"), "Pretty Voices Singing" ("Christmas Bells") and "This Ain't Boheme" ("La Vie Bohème"). Lin-Manuel Miranda, the composer and writer of the Broadway show "Hamilton", has cited "Rent" as a main source of inspiration. He also referenced the show in a verse of the song "Wrote My Way Out" on "The Hamilton Mixtape" in the line "Running out of time like I'm Jonathan Larson's rent check". "Rent" had its first staged reading at New York Theatre Workshop in March 1993. A further two-week New York Theatre Workshop version was performed in 1994 starring Anthony Rapp as Mark and Daphne Rubin-Vega as Mimi, and more workshops followed. The show opened on 1996, again at New York Theatre Workshop, and quickly gained popularity off-Broadway, receiving enthusiastic reviews. "The New York Times" theater critic Ben Brantley called it an "exhilarating, landmark rock opera" with a "glittering, inventive score" that "shimmers with hope for the future of the American musical." Another reviewer wrote, ""Rent" speaks to Generation X the way that the musical "Hair" spoke to the baby boomers or those who grew up in the 1960s," while the "New York Times" similarly called it "a rock opera for our time, a "Hair" for the 90s." The show proved extremely successful off-Broadway, selling out all of its performances at the 150-seat theatre. Due to its overwhelming popularity and the need for a larger theater, "Rent" moved to Broadway's previously derelict Nederlander Theatre on 41st Street on April 29, 1996. On Broadway, the show achieved critical acclaim and word-of-mouth popularity. The production's ethnically diverse principal cast originally included Taye Diggs, Wilson Jermaine Heredia, Jesse L. Martin, Idina Menzel, Adam Pascal, Anthony Rapp, Daphne Rubin-Vega and Fredi Walker. The production's controversial topics and innovative pricing, including same day-of-performance $20 tickets, helped to increase the popularity of musical theater amongst the younger generation. The production was nominated for ten Tony Awards in 1996 and won four: Best Musical, Best Book, Best Original Score and Best Performance by a Featured Actor in a Musical (Heredia) On April 24, 2006, the original Broadway cast reunited for a one-night performance of the musical at the Nederlander Theatre. This performance raised over $2,000,000 for the Jonathan Larson Performing Arts Foundation, Friends In Deed and New York Theatre Workshop. Former cast members were invited, and many from prior tours and former Broadway casts appeared, performing an alternate version of "Seasons of Love" as the finale of the performance. "Rent" closed on September 7, 2008, after a 12-year run and 5,123 performances, making it the eleventh-longest-running Broadway show. The production grossed over $280 million. Original cast ensemble members Rodney Hicks and Gwen Stewart returned to the cast at the time of the Broadway closing. Hicks played Benny and Stewart played the role she created, the soloist in the song "Seasons of Love". In addition, actress Tracie Thoms joined the cast at the end of the run playing Joanne, the role she portrayed in the 2005 film version. The last Broadway performance was filmed and screened in movie theaters as "Rent: Filmed Live on Broadway" in September 2008. It was released on DVD and Blu-ray formats on February 3, 2009. Successful United States national tours, the "Angel Tour" and the "Benny Tour", launched in the 1990s. Later, the non-Equity tour started its run. There was also a Canadian tour (often referred to as the "Collins Tour"). The Angel tour began in November 1996 in Boston. Anthony Rapp joined the cast for the Chicago run, and Daphne Rubin-Vega joined for the Los Angeles run. The tour finished in San Francisco in September 1999. Other members of the Angel cast included Carrie Hamilton, Amy Spanger, Luther Creek, Kristoffer Cusick, and Tony Vincent. The Benny Tour began in July 1997 in San Diego, California, at the LaJolla Playhouse. Michael Grief, the original director of the Broadway show was also the artistic director of the LaJolla Playhouse and was instrumental in arranging for the Benny tour to begin in the smaller city of San Diego rather than Los Angeles, California. It originally featured Neil Patrick Harris in the role of Mark Cohen. The Benny tour generally played shorter stops and often-smaller markets than the Angel Tour did. Other cast members included Wilson Cruz and d'Monroe. Tours ran each season from 2005 to 2008. Cast members throughout the run included Aaron Tveit, Ava Gaudet, Declan Bennett, Rebecca Naomi Jones, Constantine Maroulis, Dan Rosenbaum, Heinz Winckler, Anwar Robinson, Christine Dwyer and Karen Olivo. In 2009, a national tour starring Adam Pascal and Anthony Rapp, reprising their original Broadway roles, launched in Cleveland, Ohio. Original Broadway Cast member Gwen Steward also appeared, alongside Michael McElroy as Collins, The tour ended on February 7, 2010, in Sacramento, California. A 20th-anniversary non-Equity touring production of "Rent" began in Dallas on September 20, 2016, and is scheduled to run through May 10, 2020. The show made its UK premiere on April 21, 1998, at the West End's Shaftesbury Theatre and officially opened on May 12, 1998. The original cast included Krysten Cummings as Mimi Marquez, Wilson Jermaine Heredia as Angel Schunard, Bonny Lockhart as Benny, Jesse L. Martin as Tom Collins, Adam Pascal as Roger Davis, Anthony Rapp as Mark Cohen, and Jessica Tezier as Maureen Johnson. The show closed on October 30, 1999, after one-and-a-half years. Limited revivals took place at the Prince of Wales Theatre from December 4, 2001, to January 6, 2002; December 6, 2002, to March 1, 2003 (featuring Adam Rickett as Mark and Caprice as Maureen). There was also a successful production for a limited run in Manchester in 2006 with an additional 'goodbye' performance in 2008 from the Manchester cast. On October 16, 2007, the heavily revised production titled "Rent Remixed" opened at the Duke of York's Theatre in London's West End. Directed by William Baker, it was set in the present day. The cast included Oliver Thornton (Mark), Luke Evans (Roger), Craig Stein (Benny), Leon Lopez (Collins), Francesca Jackson (Joanne), Jay Webb (Angel), Siobhán Donaghy (Mimi), and Denise Van Outen (Maureen). From December 24, 2007, the role of Maureen was played by Jessie Wallace. The production received generally unfavorable reviews. The Guardian gave it only one out of five stars, writing, "They call this 'Rent Remixed'. I'd dub it 'Rent Reduced', in that the late Jonathan Larson's reworking of La Bohème, while never a great musical, has been turned into a grisly, synthetic, pseudo pop concert with no particular roots or identity." The production closed on February 2, 2008. The production radically altered elements of the musical including defining the characters of Mimi, Angel and Mark as British. Songs were reordered (including Maureen's first appearance as the Act I finale). The rehaul of the score was masterminded by Steve Anderson and featured radically rearranged versions of Out Tonight, Today 4 U, Over the Moon and Happy New Year. A one-off Rent - The 20th Anniversary Concert was held at the Blackpool Opera house Monday November 11, 2013 A 20th anniversary tour opened at Theatr Clwyd in October 2016 before playing a two-month run at the St James Theatre, London. The cast included Layton Williams as Angel and Lucie Jones as Maureen. The production then continued to tour the UK. In 2018 an immersive production of RENT premiered at Frogmore Paper Mill in Apsley, Hemel Hempstead. The Cast included Aran Macrae (Roger), Connor Dyer (Mark) and Lizzie Emery (Mimi). The show opened on July 10, 2018, and ran until July 28. In 2020, the musical is set to be revived at Manchester's Hope Mill Theatre for a limited run from October 30 to December 19. The production was originally scheduled to begin in late July but was delayed due to the COVID-19 pandemic. The show was revived off-Broadway at Stage 1 of New World Stages with previews starting July 14, 2011, and a scheduled opening of August 11, 2011. This was the first New York Revival of the show since the original production closed less than three years earlier. The production was directed by "Rent"'s original director Michael Greif. Almost the entire show was different from the original, yet the reinvention did not please the critics, who complained that the new actors did not have a feel for the characters they were playing and that it made the show feel contrived. The off-Broadway production of "Rent" closed on September 9, 2012. In 1999, an Australian production featured Justin Smith as Mark, Rodger Corser as Roger and Christine Anu as Mimi. The tour began in Sydney and finished in Melbourne. A production in Perth, Western Australia was mounted in 2007 and featured Anthony Callea as Mark, Tim Campbell as Roger, Courtney Act as Angel and Nikki Webster as Maureen. The Dublin production had an extended run at the Olympia Theatre, Dublin in 2000. It starred Sean Pol McGreevy as Mark, Rachel Tucker as Maureen and Allyson Brown as Mimi under the direction of Phil Willmot. The Swedish production premiered on May 15, 2002 at The Göteborg Opera in Gothenburg, Sweden, playing until June 8, 2003. Sarah Dawn Finer played Joanne. "Rent" veteran Neil Patrick Harris directed a production at the Hollywood Bowl in Los Angeles, CA. The production played a three night engagement, August 6–8, 2010. The cast included Vanessa Hudgens as Mimi, Aaron Tveit as Roger, Skylar Astin as Mark, Wayne Brady as Collins, Telly Leung as Angel, Tracie Thoms as Joanne, Nicole Scherzinger as Maureen, Collins Pennie as Benny, and Gwen Stewart as Seasons of Love soloist (and additional roles). In 2017, the first tour for the German speaking countries was mounted by Berlin theatrical producer . The production travelled Germany, Austria and Switzerland and was directed by the British opera director Walter Sutcliffe. In 2007, an abridged edition of "Rent" was made available to five non-professional acting groups in the United States for production, primarily adapted by Jennifer and Peter Jones of Stuart, Florida's Starstruck Theater. Billed as , this version omits the song "Contact" and eliminates some of the coarse language and tones down some public displays of affection in the original. Shorewood High School in Shorewood, Wisconsin, became the first high school to perform an early version of the adaptation in May 2006. The high school was selected to present a workshop performance as part of Music Theatre International's work to adapt the musical for younger actors and potentially more conservative audiences. As of 2008, Music Theatre International began licensing ""Rent" School Edition" for performances by schools and non-professional amateur theaters in the United States and around the world. "Rent" has been performed in countries around the world, including Denmark, Estonia, Finland, Iceland, Norway, Sweden, Belgium, the Netherlands, Ireland, United Kingdom, France, Germany, Switzerland, Portugal, Spain, Italy, Hungary, Poland, Slovakia, Greece, Canada, the United States, Mexico, Panama, Bolivia, Brazil, Argentina, Russia, China, Hong Kong, South Korea, Taiwan, Japan, Philippines, Singapore, Thailand, South Africa, Australia, Guam, New Zealand, Israel, Puerto Rico, Austria, Peru, Trinidad and Tobago, Dominican Republic, Cuba, Czech Republic and Guatemala. The musical has been performed in twenty-five languages: Danish, Estonian, Finnish, Icelandic, Norwegian, Swedish, Dutch, English, French, German, Portuguese, Spanish, Italian, Hungarian, Polish, Slovak, Greek, Russian, Mandarin Chinese, Cantonese Chinese, Korean, Japanese, Hebrew, Czech, and Catalan. A cast recording of the original Broadway cast recording was released in 1996; it features all the music of the show on a double-disc "complete recording" collection along with a remixed version of the song "Seasons of Love" featuring Stevie Wonder. The later 2005 film version (see below) also resulted in a double-disc cast recording of the complete score used in the movie There are also many foreign cast recordings of international productions of the show. The final performance of the Broadway production of "Rent", which took place on September 7, 2008, was filmed live and, cut together with close-up footage from a day of filming in August of the same year, was released as "Rent: Filmed Live on Broadway" in cinemas with high definition digital projection systems in the U.S. and Canada between September 24 and 28, 2008. "Rent: Filmed Live" on Broadway was released on February 3, 2009, on DVD & Blu-ray formats. In 2005, "Rent" was adapted into a movie directed by Chris Columbus with a screenplay by Stephen Chbosky. With the exception of Daphne Rubin-Vega (who was pregnant at the time of filming) and Fredi Walker (who felt she was too old for her role), who played Mimi and Joanne respectively in the original Broadway cast, the original Broadway cast members reprised the principal roles. Released on November 23, 2005, the film remained in the box office top ten for three weeks, receiving mixed reviews. Several plot elements were changed slightly, and some songs were changed to spoken dialogue or cut entirely for the film. The soundtrack was produced by Rob Cavallo, engineered by Doug McKean and features renowned session musicians Jamie Muhoberac, Tim Pierce and Dorian Crozier. In May 2017, Fox announced plans to air a of "Rent" in late 2018. However, on September 25, 2017, Fox announced the official air date for Rent Live! would be Sunday, January 27, 2019. Marc Platt is set to serve as executive producer along with the estate of Jonathan Larson. Filmmaker and "Rent" alum Andy Señor, Jr. is currently producing a documentary, following his journey producing the musical in Cuba in late 2014. This production of "Rent" was the first Broadway musical to premiere in Cuba since diplomatic relations between the two countries became strained during the Cold War.
https://en.wikipedia.org/wiki?curid=26436
Restaurant A restaurant (), or an eatery, is a business that prepares and serves food and drinks to customers. Meals are generally served and eaten on the premises, but many restaurants also offer take-out and food delivery services. Restaurants vary greatly in appearance and offerings, including a wide variety of cuisines and service models ranging from inexpensive fast food restaurants and cafeterias, to mid-priced family restaurants, to high-priced luxury establishments. In Western countries, most mid-to high-range restaurants serve alcoholic beverages such as beer and wine. Some restaurants serve all the major meals, such as breakfast, lunch, and dinner (e.g., major fast food chains, diners, hotel restaurants, and airport restaurants). Other restaurants may serve only a single meal (for example, a pancake house may only serve breakfast) or they may serve two meals (e.g., lunch and dinner). The word derives from the French verb "restaurer" ("to restore", "to revive") and, being the present participle of the verb, it literally means "that which restores". The term "restaurant" was defined in 1507 as a "restorative beverage", and in correspondence in 1521 to mean "that which restores the strength, a fortifying food or remedy". The first use of the word to refer to a public venue where one can order food is believed to be in the 18th century. In 1765, a French chef by the name of A. Boulanger established a business selling soups and other "restaurants" ("restoratives"). Additionally, while not the first establishment where one could order food, or even soups, it is thought to be the first to offer a menu of available choices The "first real restaurant" is considered to have been "La Grande Taverne de Londres" in Paris, founded by Antoine Beauvilliers in either 1782 or 1786. According to Brillat-Savarin, this was "the first to combine the four essentials of an elegant room, smart waiters, a choice cellar, and superior cooking". In 1802 the term was applied to an establishment where restorative foods, such as bouillon, a meat broth, were served ("établissement de restaurateur"). Restaurants are classified or distinguished in many different ways. The primary factors are usually the food itself (e.g. vegetarian, seafood, steak); the cuisine (e.g. Italian, Chinese, Japanese, Indian, French, Mexican, Thai) or the style of offering (e.g. tapas bar, a sushi train, a tastet restaurant, a buffet restaurant or a yum cha restaurant). Beyond this, restaurants may differentiate themselves on factors including speed (see fast food), formality, location, cost, service, or novelty themes (such as automated restaurants). Some of these include fine dining, casual dining, contemporary casual, family style, fast casual, fast food, cafes, buffet, concession stands, food trucks, pop-up restaurants, and ghost restaurants. Restaurants range from inexpensive and informal lunching or dining places catering to people working nearby, with modest food served in simple settings at low prices, to expensive establishments serving refined food and fine wines in a formal setting. In the former case, customers usually wear casual clothing. In the latter case, depending on culture and local traditions, customers might wear semi-casual, semi-formal or formal wear. Typically, at mid- to high-priced restaurants, customers sit at tables, their orders are taken by a waiter, who brings the food when it is ready. After eating, the customers then pay the bill. In some restaurants, such as workplace cafeterias, there are no waiters; the customers use trays, on which they place cold items that they select from a refrigerated container and hot items which they request from cooks, and then they pay a cashier before they sit down. Another restaurant approach which uses few waiters is the buffet restaurant. Customers serve food onto their own plates and then pay at the end of the meal. Buffet restaurants typically still have waiters to serve drinks and alcoholic beverages. Fast food restaurants are also considered a restaurant. In addition, food trucks are another popular option for people who want quick food service. Tourists around the world can enjoy dining services on railway cars and cruise ships dining rooms, which are essentially traveling restaurants. Many railways dining services cater to the needs of travelers by providing railway refreshment rooms at railway stations. The fondness for such refreshment rooms, however, may be purely cultural as not one refreshing crumb of them has ever been witnessed in the United States. In the 2000s, a number of traveling restaurants, specifically designed for tourists, had been created. These can be found on trams, boats, buses, etc. Unfortunately, in 2020 little evidence remains of them in the United States. Many cruise ships provide a variety of dining experiences including a main restaurant, satellites restaurants, room service, specialty restaurants, cafes, bars, and buffets to name a few. Some restaurants on these cruise ships required reservations and specific dress codes. A restaurant's proprietor is called a "restaurateur", this derives from the French verb "restaurer", meaning "to restore". Professional cooks are called chefs, with there being various finer distinctions (e.g. sous-chef, chef de partie). Most restaurants (other than fast food restaurants and cafeterias) will have various waiting staff to serve food, beverages and alcoholic drinks, including busboys who remove used dishes and cutlery. In finer restaurants, this may include a host or hostess, a maître d'hôtel to welcome customers and to seat them, and a sommelier or wine waiter to help patrons select wines. A new route to becoming a restaurateur, rather than working one's way up through the stages, is to operate a food truck. Once a sufficient following has been obtained, a permanent restaurant site can be opened. This trend has become common in the UK and the US. A chef's table is a table located in the kitchen of a restaurant, reserved for VIPs and special guests. Patrons may be served a themed tasting menu prepared and served by the head chef. Restaurants can require a minimum party and charge a higher flat fee. Because of the demand on the kitchen's facilities, chef's tables are generally only available during off-peak times. In China, food catering establishments that may be described as restaurants have been known since the 11th century in Kaifeng, China's capital during the first half of the Song dynasty (960–1279). Probably growing out of the tea houses and taverns that catered to travellers, Kaifeng's restaurants blossomed into an industry catering to locals as well as people from other regions of China. There is a direct correlation between the growth of the restaurant businesses and institutions of theatrical stage drama, gambling and prostitution which served the burgeoning merchant middle class during the Song dynasty. Restaurants catered to different styles of cuisine, price brackets, and religious requirements. Even within a single restaurant choice were available, and people ordered the entree from written menus. An account from 1275 writes of Hangzhou, the capital city for the last half of the dynasty: The restaurants in Hangzhou also catered to many northern Chinese who had fled south from Kaifeng during the Jurchen invasion of the 1120s, while it is also known that many restaurants were run by families formerly from Kaifeng. In Ancient Greece and Ancient Rome, thermopolia (singular "thermopolium") were small restaurant-bars that offered food and drinks to customers. A typical thermopolium had L-shaped counters in which large storage vessels were sunk, which would contain either hot or cold food. Their popularity was linked to the lack of kitchens in many dwellings and the ease with which people could purchase prepared foods. Furthermore, eating out was considered an important aspect of socializing. In Pompeii, 158 thermopolia with service counters have been identified throughout the town. They were concentrated along the main axis of the town and the public spaces where they were frequented by the locals. France has had a rich history with the development of various forms of inns and eateries, eventually to form many of the now-ubiquitous elements of the modern restaurant. As far back as the thirteenth century, inns served a variety of food — bread, cheese, bacon, roasts, usually eaten at a common table. Parisians could buy what was essentially take-out food from "rôtisseurs", who prepared roasted meat dishes, and pastry-cooks, who could prepare meat pies and often more elaborate dishes. Municipal statutes stated that the official prices per item were to be posted at the entrance; this was the first official mention of menus. Taverns also served food, as did cabarets. A cabaret, however, unlike a tavern, served food at tables with tablecloths, provided drinks with the meal, and charged by the customers' choice of dish, rather than by the pot. Cabarets were reputed to serve better food than taverns and a few, such as the Petit Maure, became well-known. A few cabarets had musicians or singing, but most, until the late 19th century, were simply convivial eating places. The first café opened in Paris in 1672 at the Saint-Germain fair. By 1723 there were nearly four hundred cafés in Paris, but their menu was limited to simpler dishes or confectionaries, such as coffee, tea, chocolate, ice creams, pastries, and liqueurs. At the end of the 16th century, the guild of cook-caterers (later known as "traiteurs") was given its own legal status. The "traiteurs" dominated sophisticated food service, delivering or preparing meals for the wealthy at their residences. Taverns and cabarets were limited to serving little more than roast or grilled meats. Towards the end of the seventeenth century, both inns and then traiteurs began to offer "host's tables" ("tables d'hôte"), where one paid a set price to sit at a large table with other guests and eat a fixed menu meal. The earliest modern-format "restaurants" to use that name in Paris were the establishments which served bouillon, a broth made of meat and egg which was said to restore health and vigor. The first restaurant of this kind opened in 1765 or 1766 by Mathurin Roze de Chantoiseau on rue des Poulies, now part of the Rue de Louvre. The name of the owner is sometimes given as Boulanger. Unlike earlier eating places, it was elegantly decorated, and besides meat broth offered a menu of several other "restorative" dishes, including macaroni. Chantoiseau and other chefs took the title "traiteurs-restaurateurs". In June 1786 the Provost of Paris issued a decree giving the new kind of eating establishment official status, authorizing "restaurateurs" to receive clients and to offer them meals until eleven in the evening in winter and midnight in summer. Ambitious cooks from noble households began to open more elaborate eating places. The first luxury restaurant in Paris, the Taverne Anglaise, was opened at the Palais-Royal at the beginning of 1786, shortly before the French Revolution, by Antoine Beauvilliers, the former chef of the Count of Provence, It had mahogany tables, linen tablecloths, chandeliers, well-dressed and trained waiters, a long wine list and an extensive menu of elaborately prepared and presented dishes. The French Revolution caused a mass emigration of nobles, and many of their cook chose to open restaurants. One restaurant was started in 1791 by Méot, the former chef of the Duke of Orleans, which offered a wine list with twenty-two choices of red wine and twenty-seven of white wine. By the end of the century there were a collection of luxury restaurants at the Grand-Palais: Huré, the Couvert espagnol; Février; the Grotte flamande; Véry, Masse and the Café de Chartres (still open, now Le Grand Vefour) In the early 19th century traiteurs and restaurateurs, became known simply as "restaurateurs". The use of the term "restaurant" for the establishment itself only became common in the nineteenth century). The first restaurant guide, called "Almanach des Gourmandes", written by Grimod de La Reyniére, was published in 1804. During the French Restoration period, the most celebrated restaurant was the Rocher de Cancale, frequented by the characters of Balzac. In the middle of the century, Balzac's characters moved to the Cafe Anglais, which in 1867 also hosted the famous Three Emperors Dinner hosted by Napoleon III in honor of Tsar Alexander II, Kaiser Wilhelm I and Otto von Bismarck during the Exposition Universelle in 1867 Other restaurants that occupy a place in French history and literature include Maxim's and Fouquet's. The restaurant of Hotel Ritz Paris, opened in 1898, was made famous by its chef, Auguste Escoffier. The 19th century also saw the appearance of new kinds of more modest restaurants, including the bistrot. The brasserie featured beer and was made popular during the 1867 Paris Exposition. In Brazil, restaurants varieties mirrors the multitude of nationalities that arrived in the country: Japanese, Arab, German, Italian, Portuguese and many more. In Colombia, a "piqueteadero" is a type of casual or rustic eatery. Meals are often shared, and typical offerings include dishes such as chorizo, chicharrón, fried organs, fried yuca, maduro and corn on the cob. Customers order the foods they want and the prepared foods are served together on a platter to be shared. The word "piquete" can be used to refer to a common Colombian type of meal that includes meat, yuca and potatoes, which is a type of meal served at a piqueteaderos. The verb form of the word piquete, piquetear, means to participate in binging, liquor drinking, and leisure activities in popular areas or open spaces. In Peru, many indigenous, Spanish, and Chinese dishes are frequently found. Because of recent immigration from places such as China, and Japan, there are many Chinese and Japanese restaurants around the country, especially in the capital city of Lima. In the United States, it was not until the late 18th century that establishments that provided meals without also providing lodging began to appear in major metropolitan areas in the form of coffee and oyster houses. The actual term "restaurant" did not enter into the common parlance until the following century. Prior to being referred to as "restaurants" these eating establishments assumed regional names such as "eating house" in New York City, "restorator" in Boston, or "victualing house" in other areas. Restaurants were typically located in populous urban areas during the 19th century and grew both in number and sophistication in the mid-century due to a more affluent middle class and to suburbanization. The highest concentration of these restaurants were in the West, followed by industrial cities on the Eastern Seaboard. In the 1970s, there was one restaurant for every 7,500 persons. In 2016, there were 1,000,000 restaurants; one for every 310 people. The average person eats out five to six times weekly. 10% of the nation's workforce is composed of restaurant workers. According to a Gallup Poll in 2016, nearly 61% of Americans across the country eat out at a restaurant once a week or more, and this percent is only predicted to increase in future years. Before the COVID-19 pandemic, The National Restaurant Association estimated restaurant sales of $899 billion in 2020. The association now projects that the pandemic will decrease that to $675 billion, a decline of $274 billion over their previous estimate. Restaurant guides review restaurants, often ranking them or providing information to guide consumers (type of food, handicap accessibility, facilities, etc.). One of the most famous contemporary guides is the Michelin series of guides which accord from 1 to 3 stars to restaurants they perceive to be of high culinary merit. Restaurants with stars in the Michelin guide are formal, expensive establishments; in general the more stars awarded, the higher the prices. The main competitor to the Michelin guide in Europe is the guidebook series published by Gault Millau. Its ratings are on a scale of 1 to 20, with 20 being the highest. In the United States, the Forbes Travel Guide (previously the Mobil travel guides) and the AAA rate restaurants on a similar 1 to 5 star (Forbes) or diamond (AAA) scale. Three, four, and five star/diamond ratings are roughly equivalent to the Michelin one, two, and three star ratings while one and two star ratings typically indicate more casual places to eat. In 2005, Michelin released a New York City guide, its first for the United States. The popular Zagat Survey compiles individuals' comments about restaurants but does not pass an "official" critical assessment. FreshNYC recommends plausible New York City restaurants for busy New Yorkers and visitors alike. The "Good Food Guide," published by the Fairfax Newspaper Group in Australia, is the Australian guide listing the best places to eat. Chefs Hats are awarded for outstanding restaurants and range from one hat through three hats. The "Good Food Guide" also incorporates guides to bars, cafes and providers. "The Good Restaurant Guide" is another Australian restaurant guide that has reviews on the restaurants as experienced by the public and provides information on locations and contact details. Any member of the public can submit a review. Nearly all major American newspapers employ food critics and publish online dining guides for the cities they serve. Some news sources provide customary reviews of restaurants, while others may provide more of a general listings service. More recently Internet sites have started up that publish both food critic reviews and popular reviews by the general public. Many restaurants are small businesses, and franchise restaurants are common. There is often a relatively large immigrant representation, reflecting both the relatively low start-up costs of the industry (thus making restaurant ownership an option for immigrants with relatively few resources) and the cultural importance of food. Indian restaurant industry is highly fragmented with more than 1.5 million outlets of which only around 3000 of them are from the organized segment. Organized segment includes Quick Service Restaurants (QSRs), Casual Dining, Cafes, Fine Dining and Pubs, Bars, Clubs and Lounges. There are 86,915 commercial foodservice units in Canada, or 26.4 units per 10,000 Canadians. By segment, there are: Fully 63% of restaurants in Canada are independent brands. Chain restaurants account for the remaining 37%, and many of these are locally owned and operated franchises. The EU-27 has an estimated 1.6m businesses involved in 'accommodation & food services', more than 75% of which are small and medium enterprises. As of 2006, there are approximately 215,000 full-service restaurants in the United States, accounting for $298 billion in sales, and approximately 250,000 limited-service (fast food) restaurants, accounting for $260 billion. Starting in 2016, Americans spent more on restaurants than groceries. In October 2017, "The New York Times" reported there are 620,000 eating and drinking places in the United States, according to the Bureau of Labor Statistics. They also reported that the number of restaurants are growing almost twice as fast as the population. One study of new restaurants in Cleveland, Ohio found that 1 in 4 changed ownership or went out of business after one year, and 6 out of 10 did so after three years. (Not all changes in ownership are indicative of financial failure.) The three-year failure rate for franchises was nearly the same. Restaurants employed 912,100 cooks in 2013, earning an average $9.83 per hour. The waiting staff numbered 4,438,100 in 2012, earning an average $8.84 per hour. Jiaxi Lu of the "Washington Post" reports in 2014 that, "Americans are spending $683.4 billion a year dining out, and they are also demanding better food quality and greater variety from restaurants to make sure their money is well spent." Dining in restaurants has become increasingly popular, with the proportion of meals consumed outside the home in restaurants or institutions rising from 25% in 1950 to 46% in 1990. This is caused by factors such as the growing numbers of older people, who are often unable or unwilling to cook their meals at home and the growing number of single-parent households. It is also caused by the convenience that restaurants can afford people; the growth of restaurant popularity is also correlated with the growing length of the work day in the US, as well as the growing number of single parent households. Eating in restaurants has also become more popular with the growth of higher income households. At the same time, less expensive establishments such as fast food establishments can be quite inexpensive, making restaurant eating accessible to many. The restaurant industry in the United States is large and quickly growing, with 10 million workers. 1 in every 12 U.S. residents work in the business, and during the 2008 recession, the industry was an anomaly in that it continued to grow. Restaurants are known for having low wages, which they claim are due to thin profit margins of 4-5%. For comparison, however, Walmart has a 1% profit margin. As a result of these low wages, restaurant employees suffer from three times the poverty rate as other U.S. workers, and use food stamps twice as much. Restaurants also employ marginalized groups. They are the largest employer of people of color. Restaurants rank as the second largest employer of immigrants. These workers statistically are concentrated in the lowest paying positions in the restaurant industry. In the restaurant industry, 39% of workers earn minimum wage or lower. In many countries, restaurants are subject to inspections by health inspectors to maintain standards for public health, such as maintaining proper hygiene and cleanliness. As part of these inspections, cooking and handling practices of ground beef are taken into account to protect against the spread of E coli poisoning. The most common kind of violations of inspection reports are those concerning the storage of cold food at appropriate temperatures, proper sanitation of equipment, regular hand washing and proper disposal of harmful chemicals. Simple steps can be taken to improve sanitation in restaurants. As sickness is easily spread through touch, restaurants are encouraged to regularly wipe down tables, door knobs and menus. Depending on local customs, legislation and the establishment, restaurants may or may not serve alcoholic beverages. Restaurants are often prohibited from selling alcoholic beverages without a meal by alcohol sale laws; such sale is considered to be activity for bars, which are meant to have more severe restrictions. Some restaurants are licensed to serve alcohol ("fully licensed"), or permit customers to "bring your own" alcohol (BYO / BYOB). In some places restaurant licenses may restrict service to beer, or wine and beer.
https://en.wikipedia.org/wiki?curid=26437
Rolf Nevanlinna Rolf Herman Nevanlinna (né Neovius; 22 October 1895 – 28 May 1980) was a Finnish mathematician who made significant contributions to complex analysis. Nevanlinna was born Rolf Herman Neovius, becoming Nevanlinna in 1906 when his father changed the family name. The Neovius-Nevanlinna family contained many mathematicians: Edvard Engelbert Neovius (Rolf's grandfather) taught mathematics and topography at a military academy; Edvard Rudolf Neovius (Rolf's uncle) was a professor of mathematics at the University of Helsinki from 1883–1900; Lars Theodor Neovius-Nevanlinna (Rolf's uncle) was an author of mathematical textbooks; and Otto Wilhelm Neovius-Nevanlinna (Rolf's father) was a physicist, astronomer and mathematician. After Otto obtained his Ph.D. in physics from the University of Helsinki, he studied at the Pulkovo Observatory with the German astronomer Herman Romberg, whose daughter, Margarete Henriette Louise Romberg, he married in 1892. Otto and Margarete then settled in Joensuu, where Otto taught physics, and there their four children were born: Frithiof (born 1894; also a mathematician), Rolf (born 1895), Anna (born 1896) and Erik (born 1901). Nevanlinna began his formal education at the age of 7. Having already been taught to read and write by his parents, he went straight into the second grade but still found the work boring and soon refused to attend the school. He was then homeschooled before being sent to a grammar school in 1903 when the family moved to Helsinki, where his father took up a new post as a teacher at Helsinki High School. At the new school, Nevanlinna studied French and German in addition to the languages he already spoke: Finnish and Swedish. He also attended an orchestra school and had a love of music, which was encouraged by his mother: Nevanlinna then progressed onto the Helsinki High School, where his main interests were classics and mathematics. He was taught by a number of teachers during this time but the best of them all was his own father, who taught him physics and mathematics. He graduated in 1913 having performed very well, although he was not the top student of his year. He then went beyond the school syllabus in the summer of 1913 when he read Ernst Leonard Lindelöf's "Introduction to Higher Analysis"; from that time on, Nevanlinna had an enthusiastic interest in mathematical analysis. (Lindelöf was also a cousin of Nevanlinna's father, and so a part of the Neovius-Nevanlinna mathematical family.) Nevanlinna began his studies at the University of Helsinki in 1913, and received his Master of Philosophy in mathematics in 1917. Lindelöf taught at the university and Nevanlinna was further influenced by him. During his time at the University of Helsinki, World War I was underway and Nevanlinna wanted to join the 27th Jäger Battalion, but his parents convinced him to continue with his studies. He did however join the White Guard in the Finnish Civil War, but did not see active military action. In 1919, Nevanlinna presented his thesis, entitled "Über beschränkte Funktionen die in gegebenen Punkten vorgeschriebene Werte annehmen" ("On limited functions prescribed values at given points"), to Lindelöf, his doctoral advisor. The thesis, which was on complex analysis, was of high quality and Nevanlinna was awarded his Doctor of Philosophy on 2 June 1919. When Nevanlinna earned his doctorate in 1919, there were no university posts available so he became a school teacher. His brother, Frithiof, had received his doctorate in 1918 but likewise was unable to take up a post at a university, and instead began working as a mathematician for an insurance company. Frithiof recruited Rolf to the company, and Nevanlinna worked for the company and as a school teacher until he was appointed a Docent of Mathematics at the University of Helsinki in 1922. During this time, he had been contacted by Edmund Landau and requested to move to Germany to work at the University of Göttingen, but did not accept. After his appointment as Docent of Mathematics, he gave up his insurance job but did not resign his position as school teacher until he received a newly created full professorship at the university in 1926. Despite this heavy workload, it was between the years of 1922–25 that he developed what would become to be known as Nevanlinna theory. From 1947 Nevanlinna had a chair in the University of Zurich, which he held on a half-time basis after receiving in 1948 a permanent position as one of the 12 salaried Academicians in the newly created Academy of Finland. Rolf Nevanlinna's most important mathematical achievement is the "value distribution theory" of meromorphic functions. The roots of the theory go back to the result of Émile Picard in 1879, showing that a non-constant complex-valued function which is analytic in the entire complex plane assumes all complex values save at most one. In the early 1920s Rolf Nevanlinna, partly in collaboration with his brother Frithiof, extended the theory to cover meromorphic functions, i.e. functions analytic in the plane except for isolated points in which the Laurent series of the function has a finite number of terms with a negative power of the variable. Nevanlinna's value distribution theory or Nevanlinna theory is crystallised in its two "Main Theorems". Qualitatively, the first one states that if a value is assumed less frequently than average, then the function comes close to that value more often than average. The Second Main Theorem, more difficult than the first one, states roughly that there are relatively few values which the function assumes less often than average. Rolf Nevanlinna's article "Zur Theorie der meromorphen Funktionen" which contains the Main Theorems was published in 1925 in the journal Acta Mathematica. Hermann Weyl has called it "one of the few great mathematical events of the [twentieth] century." Nevanlinna gave a fuller account of the theory in the monographs "Le théoreme de Picard – Borel et la théorie des fonctions méromorphes" (1929) and "Eindeutige analytische Funktionen" (1936). Nevanlinna theory touches also on a class of functions called the Nevanlinna class, or functions of "bounded type". When the Winter War broke out (1939), Nevanlinna was invited to join the Finnish Army's Ballistics Office to assist in improving artillery firing tables. These tables had been based on a calculation technique developed by General Vilho Petter Nenonen, but Nevanlinna now came up with a new method which made them considerably faster to compile. In recognition of his work he was awarded the Order of the Cross of Liberty, Second Class, and throughout his life he held this honour in especial esteem. Among Rolf Nevanlinna's later interests in mathematics were the theory of Riemann surfaces (the monograph "Uniformisierung" in 1953) and functional analysis ("Absolute analysis" in 1959, written in collaboration with his brother Frithiof). Nevanlinna also published in Finnish a book on the foundations of geometry and a semipopular account of the Theory of Relativity. His Finnish textbook on the elements of complex analysis, "Funktioteoria" (1963), written together with Veikko Paatero, has appeared in German, English and Russian translations. Rolf Nevanlinna supervised at least 28 doctoral theses. His first and most famous doctoral student was Lars Ahlfors, one of the first two Fields Medal recipients. The research for which Ahlfors was awarded the prize (proving the Denjoy Conjecture, now known as the Denjoy–Carleman–Ahlfors theorem) was strongly based on Nevanlinna's work. Nevanlinna's work was recognised in the form of honorary degrees which he held from the universities of Heidelberg, the University of Bucharest, the University of Giessen, the Free University of Berlin, the University of Glasgow, the University of Uppsala, the University of Istanbul and the University of Jyväskylä. He was an honorary member of several learned societies, among them the London Mathematical Society and the Hungarian Academy of Sciences. — The 1679 Nevanlinna main belt asteroid is named after him. From 1954, Rolf Nevanlinna chaired the committee which set about the first computer project in Finland. Rolf Nevanlinna served as President of the International Mathematical Union (IMU) from 1959 to 1963 and as President of the International Congress of Mathematicians (ICM) in 1962. In 1964, Nevanlinna's connections with President Urho Kekkonen were instrumental in bringing about a total reorganization of the Academy of Finland. From 1965 to 1970 Nevanlinna was Chancellor of the University of Turku. Although Nevanlinna did not participate actively in politics, he was known to sympathise with the right-wing Patriotic People's Movement and, partly because of his half-German parentage, was also sympathetic towards Nazi Germany; with many mathematics professors fired in the 1930s due to the Nuremberg Laws, mathematicians sympathetic to the Nazi policies were sought as replacements, and Nevanlinna accepted a position as professor at the University of Göttingen in 1936 and 1937. His sympathy towards the Nazis led to his removal from his position as Rector of the University of Helsinki after Finland made peace with the Soviet Union in 1944. In the spring of 1941, Finland contributed a Volunteer Battalion to the Waffen-SS. At the time, the battalion was a symbolic bond between Germany and Finland as both fought against the Soviet Union but without a formal alliance between the two nations. In 1942, a committee was established for the Volunteer Battalion to take care of the battalion's somewhat strained relations with its German commanders, and Nevanlinna was chosen to be the chairman of the committee, as he was a person respected in Germany but loyal to Finland. He stated in his autobiography that he accepted this role due to a "sense of duty". Nevanlinna's political activities did not colour his relationships with his mathematical contacts; after World War II, the Soviet mathematical community was isolated from the Western mathematical community and the International Colloquium on Function Theory in Helsinki in 1957, directed by Nevanlinna, was the first post-war occasion when Soviet mathematicians could contact their Western colleagues in person. In 1965, Nevanlinna was an honorary guest at a function theory congress in Soviet Armenia. When the IMU in 1981 decided to create a prize, similar to the Fields Medal, in theoretical computer science and the funding for the prize was secured from Finland, the Union decided to give Nevanlinna's name to the prize; the Rolf Nevanlinna Prize is awarded every four years at the ICM. In 2018, the General Assembly of the IMU approved a resolution to remove Nevanlinna's name from the prize.
https://en.wikipedia.org/wiki?curid=26438
Red panda The red panda ("Ailurus fulgens") is a mammal species native to the eastern Himalayas and southwestern China. It is listed as Endangered on the IUCN Red List because the wild population is estimated at fewer than 10,000 mature individuals and continues to decline due to habitat loss and fragmentation, poaching, and inbreeding depression. Despite its name, it is not closely related to the giant panda. The red panda has reddish-brown fur, a long, shaggy tail, and a waddling gait due to its shorter front legs; it is roughly the size of a domestic cat, though with a longer body, and is somewhat heavier. It is arboreal and feeds mainly on bamboo, but also eats eggs, birds, and insects. It is a solitary animal, mainly active from dusk to dawn, and is largely sedentary during the day. It is also called the lesser panda, the red bear-cat, and the red cat-bear. The red panda is the only living member of the genus "Ailurus" and the family Ailuridae. It has previously been placed in the raccoon and bear families, but the results of phylogenetic analysis provide strong support for its taxonomic classification in its own family, Ailuridae, which is part of the superfamily Musteloidea, along with the weasel, raccoon and skunk families. Traditionally it was thought to consist of two subspecies. However, results of genetic analysis indicate that there are probably two distinct red panda species, the Chinese red panda and the Himalayan red panda, which genetically diverged . The red panda has long, soft, reddish-brown fur on the upper parts, blackish fur on the lower parts, and a light face with tear markings and white badges similar to those of a raccoon, but each individual can have distinctive markings. Its skull is roundish with medium-sized upright ears, its nose is black, and its eyes are blackish. Its teeth are robust. Its long, bushy tail with six alternating transverse ochre rings provide balance and excellent camouflage in a habitat with moss- and lichen-covered trees. The legs are black and short with thick fur on the soles of the paws. This fur serves as thermal insulation on snow-covered or icy surfaces and conceals scent glands, which are also present on the anus. The head and body length of a red panda measures , and its tail is long. Males weigh and females . The red panda is specialized as a bamboo feeder with strong, curved and sharp semi-retractile claws standing inward for grasping narrow tree branches, leaves, and fruit. Like the giant panda, it has a "false thumb", which is an extension of the wrist bone. When descending a tree head-first, the red panda rotates its ankle to control its descent, one of the few climbing species to do so. The red panda is endemic to the temperate forests of the Himalayas, and ranges from the foothills of western Nepal to China in the east. Its easternmost limit is the Qinling Mountains of the Shaanxi Province in China. Its range includes southern Tibet, Sikkim and Assam in India, Bhutan, the northern mountains of Burma, and in south-western China, in the Hengduan Mountains of Sichuan and the Gongshan Mountains in Yunnan. It may also live in south-west Tibet and northern Arunachal Pradesh, but this has not been documented. Locations with the highest density of red pandas include an area in the Himalayas that has been proposed as having been a refuge for a variety of endemic species in the Pleistocene. The distribution range of the red panda should be considered disjunct, rather than continuous. A disjunct population inhabits the Meghalaya Plateau of north-eastern India. The red panda lives between altitude, inhabiting areas of moderate temperature between with little annual change. It prefers mountainous mixed deciduous and conifer forests, especially with old trees and dense understories of bamboo. During a survey in the 1970s, signs of red pandas were found in Nepal's Dhorpatan Hunting Reserve. Their presence was confirmed in spring 2007 when four red pandas were sighted at elevations ranging from . Its westernmost distribution is in Rara National Park. In 2018, red pandas were sighted at elevations of in Nepal's Lamjung District. The red panda population in Sichuan Province is larger and more stable than the Yunnan population, suggesting a southward expansion from Sichuan into Yunnan in the Holocene. The red panda has become extirpated from the Chinese provinces of Guizhou, Gansu, Shaanxi, and Qinghai. The red panda is territorial; it is solitary except during mating season. It is generally quiet except for some twittering, tweeting, and whistling communication sounds. It has been reported to be both nocturnal and crepuscular, sleeping on tree branches or in tree hollows during the day and increasing its activity in the late afternoon and early evening hours. It sleeps stretched out on a branch with legs dangling when it is hot, and curled up with its tail over the face when it is cold. It is very heat-sensitive, with an optimal "well-being" temperature between . Shortly after waking, red pandas clean their fur somewhat like a cat would, licking their front paws and then rubbing their backs, torsos, and sides. They also rub their backs and bellies along the sides of trees or rocks. Then they patrol their territories, marking with urine and a weak musk-smelling secretion from their anal glands. They search for food running along the ground or through the trees. Red pandas may use their forepaws alternately to bring food to their mouths or place food directly into their mouths. Predators of the red panda include the snow leopard ("Panthera uncia"), mustelids, and humans. If they feel threatened or sense danger, they may try to escape by climbing a rock column or tree. If they can no longer flee, they stand on their hind legs to make themselves appear larger and use the sharp claws on their front paws to defend themselves. A red panda, Futa, became a visitor attraction in Japan for his ability to stand upright for ten seconds at a time. (See also: facultative biped) Red pandas are excellent climbers, and forage largely in trees. They eat mostly bamboo, and may eat small mammals, birds, eggs, flowers, and berries. In captivity, they were observed to eat birds, flowers, maple and mulberry leaves, and bark and fruits of maple, beech, and mulberry. Like the giant panda, they cannot digest cellulose, so they must consume a large volume of bamboo to survive. Their diets consist of about two-thirds bamboo, but they also eat mushrooms, roots, acorns, lichens, and grasses. Occasionally, they supplement their diets with fish and insects. They do little more than eat and sleep due to their low-calorie diets. Bamboo shoots are more easily digested than leaves, exhibiting the highest digestibility in summer and autumn, intermediate digestibility in the spring, and lowest digestibility in the winter. These variations correlate with the nutrient contents in the bamboo. Red pandas process bamboo poorly, especially the cellulose and cell wall components. This implies microbial digestion plays only a minor role in their digestive strategy. To survive on this poor-quality diet, they have to eat the high-quality sections of the bamboo plant, such as the tender leaves and shoots, in large quantities, over of fresh leaves and of fresh shoots daily. This food passes through the digestive tract fairly rapidly (about 2–4 hr) so as to maximize daily nutrient intake. Red pandas can taste artificial sweeteners, such as aspartame, and are the only nonprimates known to be able to do so. Red pandas are able to reproduce at around 18 months of age, and are fully mature at two to three years. Adults rarely interact in the wild except to mate. Both sexes may mate with more than one partner during the mating season from mid-January to early March. A few days before birth, females begin to collect material, such as brushwood, grass, and leaves; to build a nest, which is normally located in a hollow tree or a rock crevice. After a gestation period of 112 to 158 days, the female gives birth in mid-June to late July to one to four (usually 1–2) blind and deaf cubs weighing each. After birth, the mother cleans the cubs, and can then recognize each by their smell. At first, she spends 60% to 90% of her time with the cubs. After the first week, the mother starts spending more time outside the nest, returning every few hours to nurse and groom the cubs. She moves the young frequently among several nests, all of which she keeps clean. The cubs start to open their eyes at about 18 days of age. By about 90 days, they achieve full adult fur and coloring, and begin to venture out of the nest. They also start eating solid foods at this point, weaning at around six to eight months of age. The cubs stay with their mother until the next litter is born in the following summer. Males rarely help raise the young, and only if they live in pairs or in small groups. A red panda's lifespan ranges between eight and 10 years, but individuals have been known to reach 15 years. The primary threats to red pandas are direct harvest from the wild, live or dead, competition with domestic livestock resulting in habitat degradation, and deforestation resulting in habitat loss or fragmentation. The relative importance of these factors is different in each region, and is not well understood. For instance, in India, the biggest threat seems to be habitat loss followed by poaching, while in China, the biggest threat seems to be hunting and poaching. A 40% decrease in red panda populations has been reported in China over the last 50 years, and populations in western Himalayan areas are considered to be lower. Deforestation can inhibit the spread of red pandas and exacerbate the natural population subdivision by topography and ecology, leading to severe fragmentation of the remaining wild population. Fewer than 40 animals in four separate groups share resources with humans in Nepal's Langtang National Park, where only 6% of is preferred red panda habitat. Although direct competition for food with domestic livestock is not significant, livestock can depress bamboo growth by trampling. Small groups of animals with little opportunity for exchange between them face the risk of inbreeding, decreased genetic diversity, and even extinction. In addition, clearcutting for firewood or agriculture, including hillside terracing, removes old trees that provide maternal dens and decreases the ability of some species of bamboo to regenerate. In south-west China, red pandas are hunted for their fur, especially for the highly valued bushy tails, from which hats are produced. In these areas, the fur is often used for local cultural ceremonies. In weddings, the bridegroom traditionally carries the hide. The "good-luck charm" red panda-tail hats are also used by local newly-weds. This practice may be quite old, as the red panda seems to be depicted in a 13th-century Chinese pen-and-ink scroll showing a hunting scene. Little or no mention of the red panda is made in the culture and folklore of Nepal. In the past, red pandas were captured and sold to zoos. In an article appearing in the "International Zoo News" in 1969, one reported he personally had handled 350 red pandas in 17 years. Due to CITES, this zoo harvest has decreased substantially in recent years, but poaching continues, and red pandas are often sold to private collectors at exorbitant prices. In some parts of Nepal and India, red pandas are kept as pets. The red panda has a naturally low birth rate (usually one single or twin birth per year), and a high death rate in the wild. The red panda is listed as endangered on the IUCN Red List since 2008 because the global population is estimated at about 10,000 individuals, with a decreasing population trend; only about half of the total area of potential habitat of is actually being used by the species. Due to its shy and secretive nature, and its largely nocturnal habits, observation of red pandas is difficult. Therefore, population figures in the wild are determined by population density estimates and not direct counts. It is protected in all range countries, and hunting is illegal. It is listed in CITES Appendix I. Worldwide population estimates range from fewer than 2,500 to between 16,000 and 20,000 individuals. In 1999, the total population in China was estimated at between 3,000 and 7,000 individuals. In 2001, the wild population in India was estimated at between 5,000 and 6,000 individuals. Estimates for Nepal indicate only a few hundred individuals. Reliable population numbers are hard to find, partly because other animals have been mistaken for the red panda. For instance, one report from Myanmar stated that red pandas were still fairly common in some areas; however, the accompanying photographic proof of the "red panda" was in fact a viverrid species. Conservation efforts are highly variable between countries: A community-managed forest in Ilam District of eastern Nepal is home to 15 red pandas which generate household income through tourism activities, including homestays. Villagers in the high-altitude areas of Arunachal Pradesh have formed the Pangchen Red Panda Conservation Alliance comprising five villages with a community-conserved forest area of at an altitude of to over . The international red panda studbook is currently managed at Rotterdam Zoo in the Netherlands. In cooperation with the International Red Panda Management Group, they coordinate the Species Survival Plan in North America, the European Endangered Species Programme in Europe, and other captive-breeding programs in Australia, India, Japan, and China. As of 2006, more than 800 individuals were kept in zoos and parks around the world. Of these, 511 individuals of the Himalayan red panda were kept in 173 institutions and 306 individuals of Styan's red panda were kept in 81 institutions. Since 2009, the North American Red Panda Species Survival Plan is coordinated at the Knoxville Zoo, which by 2011 had 101 red panda births. Only the Rotterdam Zoo has had more captive births worldwide. The Padmaja Naidu Himalayan Zoological Park in Darjeeling successfully released four captive-bred red pandas to the wild in August and November 2003. The most often cited example of keeping red pandas as pets is the case of former Indian prime minister Indira Gandhi. Pandas were presented to her family as a gift, and they were then housed in "a special tree house". "Ailurus fulgens" was the scientific name proposed by Frédéric Cuvier in 1825, who described a zoological specimen sent by Alfred Duvaucel "from the mountains north of India". He was the first to also use the vernacular name "panda". In the 19th and 20th centuries, the following specimens were described: Pocock distinguished "A. f. styani" from "A. f. fulgens" by its longer winter coat and greater blackness of the pelage, bigger skull, more strongly curved forehead, and more robust teeth. His description is based on skulls and skins collected in Sichuan, Myitkyina District, close to the border of Yunnan, and Upper Burma. Two subspecies are usually recognised, although results from a recent genomic study has suggested that these should be considered separate species: The Brahmaputra River is often considered the natural barrier between the two subspecies, where it makes a curve around the eastern end of the Himalayas, although some authors suggest "A. f. fulgens" extends farther eastward into China. The name "Ailurus fulgens refulgens" is sometimes incorrectly used for "A. f. styani". This stems from a lapsus made by Henri Milne-Edwards in 1874. making "A. f. refulgens" a "nomen nudum". This has been corrected in later publications. At various times, the red panda was placed in the Procyonidae, Ursidae, with "Ailuropoda" (giant panda) in the Ailuropodinae (until this family was moved into the Ursidae), and into its own family, the Ailuridae. This uncertainty comes from difficulty in determining whether certain characteristics of "Ailurus" are phylogenetically conservative or are derived and convergent with species of similar ecological habits. Evidence based on the fossil record, serology, karyology, behavior, anatomy, and reproduction reflect closer affinities with Procyonidae than Ursidae. However, ecological and foraging specializations and distinct geographical distribution in relation to modern procyonids support classification in the separate family Ailuridae. Recent molecular systematic DNA research also places the red panda into its own family, Ailuridae, a part of the broad superfamily Musteloidea that also includes the mephitids (skunks), procyonids (raccoons), and mustelids (weasels). According to the most recent phylogenetic studies, the red panda's closest relatives within the Musteloidea superfamily are the procyonids and mustelids. The red panda is considered a living fossil and only distantly related to the giant panda ("Ailuropoda melanoleuca"), as it is naturally more closely related to the other members of the superfamily Musteloidea to which it belongs. The common ancestor of both pandas (which also was an ancestor for all living bears; pinnipeds like seals and walruses; and members of the family Musteloidea like weasels and otters) can be traced back to the Paleogene period tens of millions of years ago, with a wide distribution across Eurasia. Fossils of the extinct red panda "Parailurus anglicus" were excavated in sites from China in the east to Britain in the west. In 1977, a single tooth of "Parailurus" was discovered in the Pliocene Ringold Formation of Washington. This first North American record is almost identical to European specimens and indicates the immigration of this species from Asia. In 2004, a tooth from a red panda species never before recorded in North America was discovered at the Gray Fossil Site in Tennessee. The tooth dates from 4.5–7 million years ago. This species, described as "Pristinailurus bristoli", indicates that a second, more primitive ailurine lineage inhabited North America during the Miocene. Cladistic analysis suggests that "Parailurus" and "Ailurus" are sister taxa. Additional fossils of "Pristinailurus bristoli" were discovered at the Gray Fossil Site in 2010 and in 2012. The discovery in Spain of the postcranial remains of "Simocyon batalleri", a Miocene relative to the red panda, supports a sister-group relationship between red pandas and bears. The discovery suggests the red panda's "false thumb" was an adaptation to arboreal locomotion — independent of the giant panda's adaptation to manipulate bamboo — one of the most dramatic cases of convergent evolution among vertebrates. In 2020, results of a phylogenetic analysis of red panda samples showed that red pandas in China and the Himalayas were separated by a river about 250,000 years ago. Therefore, the two subspecies should be treated as distinct species. The analysed samples showed high levels of population structure across the red panda's range. However, the results of this research should be treated with caution because of the sampling gap of >500 km between the two proposed species, and the lack of isolation-by-distance and morphometric analyses. Additionally, the use of the phylogenetic species concept for species delimitation in mammals has been associated with the unnecessary splitting of subspecies into species. "Ailurus" is adopted from the ancient Greek word (""), meaning "cat". The specific epithet "fulgens" is Latin for "shining, bright". Panda is a Roman goddess of peace and travellers, who was called upon before starting a difficult journey. The Lepcha call it "sak nam". In Nepal, it is called "bhalu biralo" (bear-cat) and "habre". The Sherpa people of Nepal and Sikkim call it "ye niglva ponva" and "wah donka". The word "wậː" is Sunuwari meaning bear; in Tamang language, a small, red bear is called "tāwām". In the Kanchenjunga region of eastern Nepal, the Limbu people know red pandas as "kaala" (literally "dark") because of their underside pelage; villagers of Tibetan origin call them "hoptongar". Additionally, Pocock lists the vernacular names "ye" and ' (Nepal); "thokya" and "thongwa" (Limbu); "oakdonga" or "wakdonka" and "woker" (Bhotia); "saknam sunam" (Lepcha). ' may originate from the Nepali word ' or ', a small bamboo, "Arundinaria intermedia", but also refers to a kind of small leopard, or cat-bear. The word ' may originate from the Nepali ' ("claw") or "" ("paw"). 'Poonya' also means "eater of bamboo". The name panda could originate from "panjā". In modern Chinese, the red panda is called "xiăoxióngmāo" ( and , lesser or small panda, or literally "little bear cat"), or / ("hóngxióngmāo", red panda or literally "red bear cat"). In English, the red panda is also called the "lesser panda", "true panda" and "common panda". The first known written record of the red panda occurs in a 13th-century Chinese scroll depicting a hunting scene between hunters and the red panda. The red panda was recognized as the state animal of Sikkim in the early 1990s, and was the mascot of the Darjeeling Tea Festival. In 2005, Babu, a male red panda at Birmingham Nature Centre in Birmingham, England, escaped and briefly became a media celebrity, before being recaptured. He was subsequently voted "Brummie of the Year", the first animal to receive this honor. Rusty, a male red panda at the National Zoo in Washington, DC, similarly attracted media attention when he briefly escaped in 2013. The name of the open-source Firefox web browser is said to have been derived from a nickname of the red panda: "fire fox". An anthropomorphic red panda was featured as Master Shifu, the kung fu teacher, in the 2008 film "Kung Fu Panda", and its sequels "Kung Fu Panda 2" in 2011 and "Kung Fu Panda 3" in 2016. The red panda Futa inspired the character of Pabu, the so-called "fire ferret" animal companion (primarily of Bolin), in the U.S. animated TV series "The Legend of Korra". Jetstar Japan uses a red panda mascot character named "Jetta" (ジェッ太). An anthropomorphic red panda, Retsuko, is the main character of the TV anime and Netflix original series "Aggretsuko".
https://en.wikipedia.org/wiki?curid=26441
Roppongi The name "Roppongi", which appears to have been coined around 1660, literally means "six trees". Six very old and large zelkova trees used to mark the area; the first three were cleared, and the last were destroyed during World War II. Another legend has it that the name comes from the fact that six "daimyōs" lived nearby during the Edo period, each with the kanji character for "tree" or a kind of tree in their names. Roppongi was not extensively populated until after the Meiji Restoration, although the area was trafficked for centuries and served as the site of the cremation of Shōgun Tokugawa Hidetada's wife in 1626. In 1890, the Third Imperial Guard of the Imperial Japanese Army was moved to a site near Roppongi (now home to the Pacific bureau of "Stars and Stripes"). The influx of soldiers led to the area's rise as a nightlife district, briefly interrupted by the Great Kanto earthquake which flattened the area in 1923. Roppongi was administratively part of Azabu Ward from 1878 to 1947. After World War II, during which the area was again destroyed, this time by aerial bombing raids, the United States Army and Allied government officials occupied several facilities in the area, beginning Roppongi's reputation as a neighborhood with large numbers of non-Japanese. Several large US military installations were located in the nearby area, with Hardy Barracks probably the most significant (the US Embassy Housing Compound and Akasaka Press Center including Hardy Barracks Recreational Lodging, "Stars and Stripes" office and heliport are still there). Surrounding the military installations were many Japanese-owned restaurants, pool halls, bars, and brothels which catered to US military personnel but were also often frequented by Japanese customers. Starting in the late 1960s, Roppongi became popular among Japanese people and foreigners alike for its disco scene, which attracted many of Tokyo's entertainment elites. Contributing to the international scene was the location of several foreign embassies and foreign corporate offices in the Roppongi area. However, many dance clubs shut down in the recession following the market crash of 1989. The Roppongi area received a major economic boost in 2002–2003 when the Izumi Garden Tower and the Roppongi Hills high-rise complexes were completed. These projects brought high-end office and condominium space to Roppongi for the first time. The Tokyo Midtown project in neighbouring Akasaka, which was completed in 2006, and includes the first Tokyo Ritz-Carlton Hotel, continued this trend. The area features numerous bars, nightclubs, strip clubs, restaurants, hostess clubs, cabarets, and other forms of entertainment. Among the expatriate community, the area tends to be favored by business people, students, and off-duty US military personnel. Overall, the neighborhood caters to a younger crowd. Clubs can range from large, multi-level establishments, to smaller one-room clubs located in upper levels of buildings. In more recent times some of the larger venues with known Yakuza connections have closed. Restaurants in Roppongi vary from upscale Japanese fare to popular international restaurants. In the past, Roppongi had a reputation as an area with high Yakuza presence, whether as customers at Roppongi establishments, conducting business, or managing or owning clubs and bars in the area. Although still exerting some influence in Roppongi, in recent times they appear to have shifted much of their presence to other districts in the Tokyo area. In 2006, Nigerian immigrants to Japan began opening a number of bars and nightclubs in the area, following an earlier group of innovators who had been in business in Roppongi for many years. The Nigerians were noted for using visible, high-pressure tactics to draw customers to their bars. In 2009 and 2010 a series of drink-spiking incidents, in which customers reported being drugged and robbed, were linked to Nigerian-owned bars. The incidents resulted in the United States embassy in Japan warning US citizens to avoid certain bars and clubs in Roppongi. An investigation by "The Japan Times" in July 2011 found that though drink spiking occurred, most of the incidents did not involve criminal activity. Many customers claimed unusually severe hangovers after nights spent in Nigerian-run establishments. Similar complaints are often made about non-Nigerian bars in Roppongi that offer unlimited drink packages and often lace drinks with hard liquor to minimize customer consumption and increase profit. Mori Building Company and The Pokémon Company have their headquarters in the Roppongi Hills Mori Tower. Companies based in Roppongi include: Public elementary and middle schools are operated by the Minato City (the Minato Ward) Board of Education. Roppongi Junior High School is located at Imoarai-Zaka, in Roppongi. Public high schools are operated by the Tokyo Metropolitan Government Board of Education. Roppongi High School is located in Roppongi. Toyo Eiwa Jogakuin is private girls school, also located at Torii-Zaka in the district. The American School in Japan Early Learning Center is in Roppongi Hills.
https://en.wikipedia.org/wiki?curid=26442
Robert Louis Stevenson Robert Louis Stevenson (13 November 1850 – 3 December 1894) was a Scottish novelist, poet and travel writer, most noted for "Treasure Island", "Kidnapped", "Strange Case of Dr Jekyll and Mr Hyde", and "A Child's Garden of Verses". Born and educated in Edinburgh, Stevenson suffered from serious bronchial trouble for much of his life, but continued to write prolifically and travel widely in defiance of his poor health. As a young man, he mixed in London literary circles, receiving encouragement from Andrew Lang, Edmund Gosse, Leslie Stephen and W. E. Henley, the last of whom may have provided the model for Long John Silver in "Treasure Island". In 1890, he settled in Samoa, where he died in 1894. A celebrity in his lifetime, Stevenson's critical reputation has fluctuated since his death, though today his works are held in general acclaim. He is currently ranked as the 26th most translated author in the world. Stevenson was born at 8 Howard Place, Edinburgh, Scotland on 13 November 1850 to Thomas Stevenson (1818–1887), a leading lighthouse engineer, and his wife Margaret Isabella (born Balfour, 1829–1897). He was christened Robert Lewis Balfour Stevenson. At about age 18, he changed the spelling of "Lewis" to "Louis", and he dropped "Balfour" in 1873. Lighthouse design was the family's profession; Thomas's father (Robert's grandfather) was civil engineer Robert Stevenson, and Thomas's brothers (Robert's uncles) Alan and David were in the same field. Thomas's maternal grandfather Thomas Smith had been in the same profession. However, Robert's mother's family were gentry, tracing their lineage back to Alexander Balfour who had held the lands of Inchyra in Fife in the fifteenth century. His mother's father Lewis Balfour (1777–1860) was a minister of the Church of Scotland at nearby Colinton, and her siblings included physician George William Balfour and marine engineer James Balfour. Stevenson spent the greater part of his boyhood holidays in his maternal grandfather's house. "Now I often wonder what I inherited from this old minister," Stevenson wrote. "I must suppose, indeed, that he was fond of preaching sermons, and so am I, though I never heard it maintained that either of us loved to hear them." Lewis Balfour and his daughter both had weak chests, so they often needed to stay in warmer climates for their health. Stevenson inherited a tendency to coughs and fevers, exacerbated when the family moved to a damp, chilly house at 1 Inverleith Terrace in 1851. The family moved again to the sunnier 17 Heriot Row when Stevenson was six years old, but the tendency to extreme sickness in winter remained with him until he was 11. Illness was a recurrent feature of his adult life and left him extraordinarily thin. Contemporaneous views were that he had tuberculosis, but more recent views are that it was bronchiectasis or even sarcoidosis. Stevenson's parents were both devout Presbyterians, but the household was not strict in its adherence to Calvinist principles. His nurse Alison Cunningham (known as Cummy) was more fervently religious. Her mix of Calvinism and folk beliefs were an early source of nightmares for the child, and he showed a precocious concern for religion. But she also cared for him tenderly in illness, reading to him from John Bunyan and the Bible as he lay sick in bed and telling tales of the Covenanters. Stevenson recalled this time of sickness in "The Land of Counterpane" in "A Child's Garden of Verses" (1885), dedicating the book to his nurse. Stevenson was an only child, both strange-looking and eccentric, and he found it hard to fit in when he was sent to a nearby school at age 6, a problem repeated at age 11 when he went on to the Edinburgh Academy; but he mixed well in lively games with his cousins in summer holidays at Colinton. His frequent illnesses often kept him away from his first school, so he was taught for long stretches by private tutors. He was a late reader, learning at age 7 or 8, but even before this he dictated stories to his mother and nurse, and he compulsively wrote stories throughout his childhood. His father was proud of this interest; he had also written stories in his spare time until his own father found them and told him to "give up such nonsense and mind your business." He paid for the printing of Robert's first publication at 16, entitled "The Pentland Rising: A Page of History, 1666". It was an account of the Covenanters' rebellion which was published in 1866, the 200th anniversary of the event. In September 1857, Stevenson went to "Mr Henderson's School" in India Street, Edinburgh, but because of poor health stayed only a few weeks and did not return until October 1859. During his many absences, he was taught by private tutors. In October 1861, he went to Edinburgh Academy, an independent school for boys, and stayed there sporadically for about fifteen months. In the autumn of 1863, he spent one term at an English boarding school at Spring Grove in Isleworth in Middlesex (now an urban area of West London). In October 1864, following an improvement to his health, he was sent to Robert Thomson's private school in Frederick Street, Edinburgh, where he remained until he went to university. In November 1867, Stevenson entered the University of Edinburgh to study engineering. He showed from the start no enthusiasm for his studies and devoted much energy to avoiding lectures. This time was more important for the friendships he made with other students in the Speculative Society (an exclusive debating club), particularly with Charles Baxter, who would become Stevenson's financial agent, and with a professor, Fleeming Jenkin, whose house staged amateur drama in which Stevenson took part, and whose biography he would later write. Perhaps most important at this point in his life was a cousin, Robert Alan Mowbray Stevenson (known as "Bob"), a lively and light-hearted young man who, instead of the family profession, had chosen to study art. Each year during vacations, Stevenson travelled to inspect the family's engineering works—to Anstruther and Wick in 1868, with his father on his official tour of Orkney and Shetland islands lighthouses in 1869, and for three weeks to the island of Erraid in 1870. He enjoyed the travels more for the material they gave for his writing than for any engineering interest. The voyage with his father pleased him because a similar journey of Walter Scott with Robert Stevenson had provided the inspiration for Scott's 1822 novel "The Pirate". In April 1871, Stevenson notified his father of his decision to pursue a life of letters. Though the elder Stevenson was naturally disappointed, the surprise cannot have been great, and Stevenson's mother reported that he was "wonderfully resigned" to his son's choice. To provide some security, it was agreed that Stevenson should read Law (again at Edinburgh University) and be called to the Scottish bar. In his 1887 poetry collection "Underwoods", Stevenson muses on his having turned from the family profession: Say not of me that weakly I declined The labours of my sires, and fled the sea, The towers we founded and the lamps we lit, To play at home with paper like a child. But rather say: "In the afternoon of time" "A strenuous family dusted from its hands" "The sand of granite, and beholding far" "Along the sounding coast its pyramids" "And tall memorials catch the dying sun," "Smiled well content, and to this childish task" "Around the fire addressed its evening hours." In other respects too, Stevenson was moving away from his upbringing. His dress became more Bohemian; he already wore his hair long, but he now took to wearing a velveteen jacket and rarely attended parties in conventional evening dress. Within the limits of a strict allowance, he visited cheap pubs and brothels. More importantly, he had come to reject Christianity and declared himself an atheist. In January 1873, his father came across the constitution of the LJR (Liberty, Justice, Reverence) Club, of which Stevenson and his cousin Bob were members, which began: "Disregard everything our parents have taught us". Questioning his son about his beliefs, he discovered the truth, leading to a long period of dissension with both parents: What a "damned" curse I am to my parents! As my father said, "You have rendered my whole life a failure". As my mother said, "This is the heaviest affliction that has ever befallen me". O Lord, what a pleasant thing it is to have damned the happiness of (probably) the only two people who care a damn about you in the world. Stevenson was visiting a cousin in England in late 1873 when he met two people who became very important to him: Sidney Colvin and Fanny (Frances Jane) Sitwell. Sitwell was a 34-year-old woman with a son, who was separated from her husband. She attracted the devotion of many who met her, including Colvin, who married her in 1901. Stevenson was also drawn to her, and they kept up a warm correspondence over several years in which he wavered between the role of a suitor and a son (he addressed her as "Madonna"). Colvin became Stevenson's literary adviser and was the first editor of his letters after his death. He placed Stevenson's first paid contribution in "The Portfolio", an essay entitled "Roads". Stevenson was soon active in London literary life, becoming acquainted with many of the writers of the time, including Andrew Lang, Edmund Gosse, and Leslie Stephen, the editor of the "Cornhill Magazine" who took an interest in Stevenson's work. Stephen took Stevenson to visit a patient at the Edinburgh Infirmary named William Ernest Henley, an energetic and talkative man with a wooden leg. Henley became a close friend and occasional literary collaborator, until a quarrel broke up the friendship in 1888, and he is often considered to be the model for Long John Silver in "Treasure Island". Stevenson was sent to Menton on the French Riviera in November 1873 to recuperate after his health failed. He returned in better health in April 1874 and settled down to his studies, but he returned to France several times after that. He made long and frequent trips to the neighborhood of the Forest of Fontainebleau, staying at Barbizon, Grez-sur-Loing, and Nemours and becoming a member of the artists' colonies there. He also traveled to Paris to visit galleries and the theatres. He qualified for the Scottish bar in July 1875, and his father added a brass plate to the Heriot Row house reading "R.L. Stevenson, Advocate". His law studies did influence his books, but he never practised law; all his energies were spent in travel and writing. One of his journeys was a canoe voyage in Belgium and France with Sir Walter Simpson, a friend from the Speculative Society, a frequent travel companion, and the author of "The Art of Golf" (1887). This trip was the basis of his first travel book "An Inland Voyage" (1878). The canoe voyage with Simpson brought Stevenson to Grez in September 1876 where he met Fanny Van de Grift Osbourne (1840–1914), born in Indianapolis. She had married at age 17 and moved to Nevada to rejoin husband Samuel after his participation in the American Civil War. Their children were Isobel (or "Belle"), Lloyd, and Hervey (who died in 1875). But anger over her husband's infidelities led to a number of separations. In 1875, she had taken her children to France where she and Isobel studied art. Stevenson returned to Britain shortly after this first meeting, but Fanny apparently remained in his thoughts, and he wrote the essay "On falling in love" for the "Cornhill Magazine". They met again early in 1877 and became lovers. Stevenson spent much of the following year with her and her children in France. In August 1878, she returned to San Francisco and Stevenson remained in Europe, making the walking trip that formed the basis for "Travels with a Donkey in the Cévennes" (1879). But he set off to join her in August 1879, against the advice of his friends and without notifying his parents. He took second-class passage on the steamship "Devonia", in part to save money but also to learn how others traveled and to increase the adventure of the journey. He then traveled overland by train from New York City to California. He later wrote about the experience in "The Amateur Emigrant". It was good experience for his writing, but it broke his health. He was near death when he arrived in Monterey, California, where some local ranchers nursed him back to health. He stayed for a time at the French Hotel located at 530 Houston Street, now a museum dedicated to his memory called the "Stevenson House". While there, he often dined "on the cuff," as he said, at a nearby restaurant run by Frenchman Jules Simoneau, which stood at what is now Simoneau Plaza; several years later, he sent Simoneau an inscribed copy of his novel "Strange Case of Dr Jekyll and Mr Hyde" (1886), writing that it would be a stranger case still if Robert Louis Stevenson ever forgot Jules Simoneau. While in Monterey, he wrote an evocative article about "the Old Pacific Capital" of Monterey. By December 1879, Stevenson had recovered his health enough to continue to San Francisco where he struggled "all alone on forty-five cents a day, and sometimes less, with quantities of hard work and many heavy thoughts," in an effort to support himself through his writing. But by the end of the winter, his health was broken again and he found himself at death's door. Fanny was now divorced and recovered from her own illness, and she came to his bedside and nursed him to recovery. "After a while," he wrote, "my spirit got up again in a divine frenzy, and has since kicked and spurred my vile body forward with great emphasis and success." When his father heard of his condition, he cabled him money to help him through this period. Fanny and Robert were married in May 1880, although he said that he was "a mere complication of cough and bones, much fitter for an emblem of mortality than a bridegroom." He travelled with his new wife and her son Lloyd north of San Francisco to Napa Valley and spent a summer honeymoon at an abandoned mining camp on Mount Saint Helena (today designated Robert Louis Stevenson State Park). He wrote about this experience in "The Silverado Squatters". He met Charles Warren Stoddard, co-editor of the "Overland Monthly" and author of "South Sea Idylls", who urged Stevenson to travel to the South Pacific, an idea which returned to him many years later. In August 1880, he sailed with Fanny and Lloyd from New York to Britain and found his parents and his friend Sidney Colvin on the wharf at Liverpool, happy to see him return home. Gradually, his wife was able to patch up differences between father and son and make herself a part of the family through her charm and wit. Stevenson searched in vain between 1880 and 1887 for a residence suitable to his health. He spent his summers at various places in Scotland and England, including Westbourne, Dorset, a residential area in Bournemouth. It was during his time in Bournemouth that he wrote the story "Strange Case of Dr Jekyll and Mr Hyde", naming the character Mr. Poole after the town of Poole which is situated next to Bournemouth. In Westbourne, he named his house "Skerryvore" after the tallest lighthouse in Scotland, which his uncle Alan had built (1838–44). A novel of Stevenson's life while residing in Westbourne was written by Adelaide A. Boodle, who had met him there. In the wintertime, Stevenson travelled to France and lived at Davos Platz and the Chalet de Solitude at Hyères, where he was very happy for a time. "I have so many things to make life sweet for me," he wrote, "it seems a pity I cannot have that other one thing—health. But though you will be angry to hear it, I believe, for myself at least, what is is best." In spite of his ill health, he produced the bulk of his best-known work during these years. "Treasure Island" was published under the pseudonym "Captain George North" and became his first widely popular book; he wrote it during this time, along with "Kidnapped", "Strange Case of Dr Jekyll and Mr Hyde" (which established his wider reputation), "", "A Child's Garden of Verses", and "Underwoods". He gave a copy of "Kidnapped" to his friend and frequent Skerryvore visitor Henry James. His father died in 1887 and Stevenson felt free to follow the advice of his physician to try a complete change of climate, so he headed for Colorado with his mother and family. But after landing in New York, they decided to spend the winter in the Adirondacks at a cure cottage now known as Stevenson Cottage at Saranac Lake, New York. During the intensely cold winter, Stevenson wrote some of his best essays, including "Pulvis et Umbra". He also began "The Master of Ballantrae" and lightheartedly planned a cruise to the southern Pacific Ocean for the following summer. Stevenson believed in Conservatism for most of his life. His cousin and biographer Sir Graham Balfour said that "he probably throughout life would, if compelled to vote, have always supported the Conservative candidate." In 1866, Stevenson voted for Benjamin Disraeli, future Conservative Prime Minister of the United Kingdom, over Thomas Carlyle for the Lord Rectorship of the University of Edinburgh. During his college years, he briefly identified himself as a "red-hot socialist". He wrote at age 26: "I look back to the time when I was a Socialist with something like regret…. Now I know that in thus turning Conservative with years, I am going through the normal cycle of change and travelling in the common orbit of men's opinions." In June 1888, Stevenson chartered the yacht "Casco" and set sail with his family from San Francisco. The vessel "plowed her path of snow across the empty deep, far from all track of commerce, far from any hand of help." The sea air and thrill of adventure for a time restored his health, and for nearly three years he wandered the eastern and central Pacific, stopping for extended stays at the Hawaiian Islands, where he became a good friend of King Kalākaua. He befriended the king's niece Princess Victoria Kaiulani, who also had Scottish heritage. He spent time at the Gilbert Islands, Tahiti, New Zealand, and the Samoan Islands. During this period, he completed "The Master of Ballantrae", composed two ballads based on the legends of the islanders, and wrote "The Bottle Imp". He preserved the experience of these years in his various letters and in his "In the South Seas" (which was published posthumously). He made a voyage in 1889 with Lloyd on the trading schooner "Equator", visiting Butaritari, Mariki, Apaiang, and Abemama in the Gilbert Islands. They spent several months on Abemama with tyrant-chief Tem Binoka, whom Stevenson described in "In the South Seas". Stevenson left Sydney, Australia, on the "Janet Nicoll" in April 1890 for his third and final voyage among the South Seas islands. He intended to produce another book of travel writing to follow his earlier book "In the South Seas", but it was his wife who eventually published her journal of their third voyage. (Fanny misnames the ship in her account "The Cruise of the Janet Nichol".) A fellow passenger was Jack Buckland, whose stories of life as an island trader became the inspiration for the character of Tommy Hadden in "The Wrecker" (1892), which Stevenson and Lloyd Osbourne wrote together. Buckland visited the Stevensons at Vailima in 1894. In 1890, Stevenson purchased a tract of about 400 acres (1.6 km²) in Upolu, an island in Samoa where he established himself on his estate in the village of Vailima after two aborted attempts to visit Scotland. He took the native name Tusitala (Samoan for "Teller of Tales"). His influence spread among the Samoans, who consulted him for advice, and he soon became involved in local politics. He was convinced that the European officials who had been appointed to rule the Samoans were incompetent, and he published "" after many futile attempts to resolve the matter. This was such a stinging protest against existing conditions that it resulted in the recall of two officials, and Stevenson feared for a time that it would result in his own deportation. He wrote to Colvin, "I used to think meanly of the plumber; but how he shines beside the politician!" He also found time to work at his writing, although he felt that "there was never any man had so many irons in the fire". He wrote "The Beach of Falesa", "Catriona" (titled "David Balfour" in the US), "The Ebb-Tide", and the "Vailima Letters" during this period. Stevenson grew depressed and wondered if he had exhausted his creative vein, as he had been "overworked bitterly" and that the best he could write was "ditch-water". He even feared that he might again become a helpless invalid. He rebelled against this idea: "I wish to die in my boots; no more Land of Counterpane for me. To be drowned, to be shot, to be thrown from a horse — ay, to be hanged, rather than pass again through that slow dissolution." He then suddenly had a return of energy and he began work on "Weir of Hermiston". "It's so good that it frightens me," he is reported to have exclaimed. He felt that this was the best work he had done. On 3 December 1894, Stevenson was talking to his wife and straining to open a bottle of wine when he suddenly exclaimed, "What's that?", asked his wife "does my face look strange?", and collapsed. He died within a few hours, probably of a cerebral haemorrhage. He was 44 years old. The Samoans insisted on surrounding his body with a watch-guard during the night and on bearing him on their shoulders to nearby Mount Vaea, where they buried him on a spot overlooking the sea on land donated by British Acting Vice Consul Thomas Trood. Stevenson had always wanted his Requiem inscribed on his tomb: Stevenson was loved by the Samoans, and his tombstone epigraph was translated to a Samoan song of grief. Half of Stevenson's original manuscripts are lost, including those of "Treasure Island", "The Black Arrow," and "The Master of Ballantrae". His heirs sold his papers during World War I, and many Stevenson documents were auctioned off in 1918. Stevenson was a celebrity in his own time, being admired by many other writers, including Jorge Luis Borges, Bertolt Brecht, Marcel Proust, Arthur Conan Doyle, Henry James, Cesare Pavese, Emilio Salgari, Ernest Hemingway, Rudyard Kipling, Jack London, Vladimir Nabokov, J. M. Barrie, and G. K. Chesterton, who said that Stevenson "seemed to pick the right word up on the point of his pen, like a man playing spillikins." Stevenson was seen for much of the 20th century as a second-class writer. He became relegated to children's literature and horror genres, condemned by literary figures such as Virginia Woolf (daughter of his early mentor Leslie Stephen) and her husband Leonard Woolf, and he was gradually excluded from the canon of literature taught in schools. His exclusion reached its nadir in the 1973 2,000-page "Oxford Anthology of English Literature" where he was entirely unmentioned, and "The Norton Anthology of English Literature" excluded him from 1968 to 2000 (1st–7th editions), including him only in the 8th edition (2006). The late 20th century brought a re-evaluation of Stevenson as an artist of great range and insight, a literary theorist, an essayist and social critic, a witness to the colonial history of the Pacific Islands, and a humanist. He was praised by Roger Lancelyn Green, one of the Oxford Inklings, as a writer of a consistently high level of "literary skill or sheer imaginative power" and a pioneer of the Age of the Story Tellers along with H. Rider Haggard. He is now evaluated as a peer of authors such as Joseph Conrad (whom Stevenson influenced with his South Seas fiction) and Henry James, with new scholarly studies and organisations devoted to him. Throughout the vicissitudes of his scholarly reception, Stevenson has remained popular worldwide. According to the Index Translationum, Stevenson is ranked the 26th most translated author in the world, ahead of Oscar Wilde and Edgar Allan Poe. On the subject of Stevenson's modern reputation, American film critic Roger Ebert wrote in 1996, The Writers' Museum near Edinburgh's Royal Mile devotes a room to Stevenson, containing some of his personal possessions from childhood through to adulthood. The Stevenson House at 530 Houston Street in Monterey, California, formerly the French Hotel, memorializes Stevenson's 1879 stay in "the Old Pacific Capital", as he was crossing the United States to join his future wife, Fanny Osbourne. The Stevenson House museum is graced with a bas-relief depicting the sickly author writing in bed. The Robert Louis Stevenson Museum in St. Helena, California, is home to over 11,000 objects and artifacts, the majority of which belonged to Stevenson. Opened in 1969, the museum houses such treasures as his childhood rocking chair, writing desk, toy soldiers, and personal writings among many other items. The museum is free to the public and serves as an academic archive for students, writers, and Stevenson enthusiasts. Stevenson's former home in Vailima, Samoa, is now a museum dedicated to the later years of his life. The museum collection includes several original items belonging to Stevenson and his family. The path to Stevenson's grave at the top of Mt Vaea commences from the museum. A bronze relief memorial to Stevenson, designed by the American sculptor Augustus Saint-Gaudens in 1904, is mounted in the Moray Aisle of St Giles' Cathedral, Edinburgh. Saint-Gaudens' scaled-down version of this relief is in the collection of the Montclair Art Museum. Another small version depicting Stevenson with a cigarette in his hand rather than the pen he holds in the St. Giles memorial is displayed in the Nichols House Museum in Beacon Hill, Boston. Another memorial in Edinburgh stands in West Princes Street Gardens below Edinburgh Castle; it is a simple upright stone inscribed with "RLS – A Man of Letters 1850–1894" by sculptor Ian Hamilton Finlay in 1987. In 2013, a statue of Stevenson as a child with his dog was unveiled by the author Ian Rankin outside Colinton Parish Church. The sculptor of the statue was Alan Herriot, and the money to erect it was raised by the Colinton Community Conservation Trust. A plaque above the door of a house in Castleton of Braemar states "Here R.L. Stevenson spent the Summer of 1881 and wrote Treasure Island, his first great work". A garden was designed by the Bournemouth Corporation in 1957 as a memorial to Stevenson, on the site of his Westbourne house, "Skerryvore", which he occupied from 1885 to 1887. A statue of the Skerryvore lighthouse is present on the site. In 1966, the Canadian actor Lloyd Bochner played Stevenson in the episode "Jolly Roger and Wells Fargo" of the syndicated American television series, "Death Valley Days", hosted by Robert Taylor and directed by Denver Pyle. In an earlier "Death Valley Days" episode from 1958, "The Great Amulet", hosted by Stanley Andrews, the actor Don Reardon (died 2004) played the role of Stevenson. In the story line, Stevenson falls in love with Fanny Osbourne, played by Aline Towne (1919-1996), the mother of two children in a loveless marriage in San Francisco. The couple met in France where Stevenson was recuperating from health issues and moved to San Francisco, where Stevenson worked tirelessly despite lingering health matters in the production of his large volume of literary works. "The Great Amulet" is revealed at the conclusion of the episode. In 1994, to mark the 100th anniversary of Stevenson's death, the Royal Bank of Scotland issued a series of commemorative £1 notes which featured a quill pen and Stevenson's signature on the obverse, and Stevenson's face on the reverse side. Alongside Stevenson's portrait are scenes from some of his books and his house in Western Samoa. Two million notes were issued, each with a serial number beginning "RLS". The first note to be printed was sent to Samoa in time for their centenary celebrations on 3 December 1994. At least six US elementary schools are named after Stevenson, in the Upper West Side of New York City, in Fridley, Minnesota, in Burbank, California, in Grandview Heights, Ohio (suburb of Columbus), in San Francisco, California, and in Merritt Island, Florida. There is an R. L. Stevenson middle school in Honolulu, Hawaii and in Saint Helena, California. Stevenson School in Pebble Beach, California, was established in 1952 and still exists as a college preparatory boarding school. Robert Louis Stevenson State Park near Calistoga, California, contains the location where he and Fanny spent their honeymoon in 1880. A street in Honolulu's Waikiki District, where Stevenson lived while in the Hawaiian Islands, was named after his Samoan moniker: Tusitala. This was also (until 2014) the name of a restaurant on Buckstone Terrace, Edinburgh on the route of a favourite walk that Stevenson often took to the village of Swanston in the Pentland Hills. In 2011, Stevenson's open letter defending Father Damien from Rev. Dr. Charles McEwen Hyde influenced the founding of the Saint Damien Advocates in Hawaii. The Chemin de Stevenson (GR 70) is a popular long-distance footpath in France that approximately follows Stevenson's route as described in "Travels with a Donkey in the Cévennes". There are numerous monuments and businesses named after him along the route, including a fountain in the town of Saint-Jean-du-Gard where Stevenson sold his donkey Modestine and took a stagecoach to Alès. The Robert Louis Stevenson Memorial is an outdoor memorial in Portsmouth Square, San Francisco, California. A memorial by Gutzon Borglum was unveiled, in 1915, at Baker Cottage, Saranac Lake, New York. List of short stories sorted chronologically. Note: does not include collaborations with Fanny found in "More New Arabian Nights: The Dynamiter". Although not well known, his island fiction and non-fiction is among the most valuable and collected of the 19th century body of work that addresses the Pacific area.
https://en.wikipedia.org/wiki?curid=26444
Recreational mathematics Recreational mathematics is mathematics carried out for recreation (entertainment) rather than as a strictly research and application-based professional activity. Although it is not necessarily limited to being an endeavor for amateurs, many topics in this field require no knowledge of advanced mathematics. Recreational mathematics involves mathematical puzzles and games, often appealing to children and untrained adults, inspiring their further study of the subject. The Mathematical Association of America (MAA) includes Recreational Mathematics as one of its seventeen Special Interest Groups, commenting: Mathematical competitions (such as those sponsored by mathematical associations) are also categorized under recreational mathematics. Some of the more well-known topics in recreational mathematics are Rubik's Cubes, magic squares, fractals, logic puzzles and mathematical chess problems, but this area of mathematics includes the aesthetics and culture of mathematics, peculiar or amusing stories and coincidences about mathematics, and the personal lives of mathematicians. Mathematical games are multiplayer games whose rules, strategies, and outcomes can be studied and explained using mathematics. The players of the game may not need to use explicit mathematics in order to play mathematical games. For example, Mancala is studied in the mathematical field of combinatorial game theory, but no mathematics is necessary in order to play it. Mathematical puzzles require mathematics in order to solve them. They have specific rules, as do multiplayer games, but mathematical puzzles don't usually involve competition between two or more players. Instead, in order to solve such a puzzle, the solver must find a solution that satisfies the given conditions. Logic puzzles and classical ciphers are common examples of mathematical puzzles. Cellular automata and fractals are also considered mathematical puzzles, even though the solver only interacts with them by providing a set of initial conditions. As they often include or require game-like features or thinking, mathematical puzzles are sometimes also called mathematical games. Other curiosities and pastimes of non-trivial mathematical interest include: There are many online blogs devoted to recreational mathematics. Among them Prominent practitioners and advocates of recreational mathematics have included:
https://en.wikipedia.org/wiki?curid=26446
Resurrection Resurrection or anastasis is the concept of coming back to life after death. In a number of religions, a dying-and-rising god is a deity which dies and resurrects. The resurrection of the dead is a standard eschatological belief in the Abrahamic religions. As a religious concept, it is used in two distinct respects: a belief in the resurrection of individual souls that is current and ongoing (Christian idealism, realized eschatology), or else a belief in a singular resurrection of the dead at the end of the world. Some believe the soul is the actual vehicle by which people are resurrected. The death and resurrection of Jesus is a central focus of Christianity. Christian theological debate ensues with regard to what kind of resurrection is factual – either a "spiritual" resurrection with a spirit body into Heaven, or a material resurrection with a restored human body. While most Christians believe Jesus' resurrection from the dead and ascension to Heaven was in a material body, a very small minority believes it was spiritual. Resurrection, from the Latin noun "resurrectio -onis", from the verb "rego", "to make straight, rule" + preposition "sub", "under", altered to "subrigo" and contracted to "surgo, surrexi, surrectum" ("to rise", "get up", "stand up") + preposition "re-", "again", thus literally "a straightening from under again". The concept of resurrection is found in the writings of some ancient non-Abrahamic religions in the Middle East. A few extant Egyptian and Canaanite writings allude to dying and rising gods such as Osiris and Baal. Sir James Frazer in his book "The Golden Bough" relates to these dying and rising gods, but many of his examples, according to various scholars, distort the sources. Taking a more positive position, Tryggve Mettinger argues in his recent book that the category of rise and return to life is significant for Ugaritic Baal, Melqart, Adonis, Eshmun, Osiris and Dumuzi. In ancient Greek religion a number of men and women became physically immortal as they were resurrected from the dead. Asclepius was killed by Zeus, only to be resurrected and transformed into a major deity. Achilles, after being killed, was snatched from his funeral pyre by his divine mother Thetis and resurrected, brought to an immortal existence in either Leuce, the Elysian plains or the Islands of the Blessed. Memnon, who was killed by Achilles, seems to have received a similar fate. Alcmene, Castor, Heracles, and Melicertes, were also among the figures sometimes considered to have been resurrected to physical immortality. According to Herodotus's "Histories", the seventh century BC sage Aristeas of Proconnesus was first found dead, after which his body disappeared from a locked room. Later he found not only to have been resurrected but to have gained immortality. Many other figures, like a great part of those who fought in the Trojan and Theban wars, Menelaus, and the historical pugilist Cleomedes of Astupalaea, were also believed to have been made physically immortal, but without having died in the first place. Indeed, in Greek religion, immortality originally always included an eternal union of body and soul. The philosophical idea of an immortal soul was a later invention, which, although influential, never had a breakthrough in the Greek world. As may be witnessed even into the Christian era, not least by the complaints of various philosophers over popular beliefs, traditional Greek believers maintained the conviction that certain individuals were resurrected from the dead and made physically immortal and that for the rest of us, we could only look forward to an existence as disembodied and dead souls. Greek philosophers generally denied this traditional religious belief in physical immortality. Writing his "Lives of Illustrious Men" (Parallel Lives) in the first century, the Middle Platonic philosopher Plutarch in his chapter on Romulus gave an account of the mysterious disappearance and subsequent deification of this first king of Rome, comparing it to traditional Greek beliefs such as the resurrection and physical immortalization of Alcmene and Aristeas the Proconnesian, "for they say Aristeas died in a fuller's work-shop, and his friends coming to look for him, found his body vanished; and that some presently after, coming from abroad, said they met him traveling towards Croton". Plutarch openly scorned such beliefs held in traditional ancient Greek religion, writing, "many such improbabilities do your fabulous writers relate, deifying creatures naturally mortal." Alcestis undergoes resurrection over a three-day period of time, but without achieving immortality. The parallel between these traditional beliefs and the later resurrection of Jesus was not lost on the early Christians, as Justin Martyr argued: "when we say ... Jesus Christ, our teacher, was crucified and died, and rose again, and ascended into heaven, we propose nothing different from what you believe regarding those whom you consider sons of Zeus." ("1 Apol." 21). There are stories in Buddhism where the power of resurrection was allegedly demonstrated in Chan or Zen tradition. One is the legend of Bodhidharma, the Indian master who brought the Ekayana school of India that subsequently became Chan Buddhism to China. The other is the passing of Chinese Chan master Puhua (Japanese:Jinshu Fuke) and is recounted in the Record of Linji (Japanese: Rinzai Gigen). Puhua was known for his unusual behavior and teaching style so it is no wonder that he is associated with an event that breaks the usual prohibition on displaying such powers. Here is the account from Irmgard Schloegl's "The Zen Teaching of Rinzai". In Christianity, resurrection most critically concerns the resurrection of Jesus, but also includes the resurrection of Judgment Day known as the resurrection of the dead by those Christians who subscribe to the Nicene Creed (which is the majority or mainstream Christianity), as well as the resurrection miracles done by Jesus and the prophets of the Old Testament. In the New Testament, Jesus is said to have raised several persons from death. These resurrections included the daughter of Jairus shortly after death, a young man in the midst of his own funeral procession, and Lazarus of Bethany, who had been buried for four days. During the Ministry of Jesus on earth, before his death, Jesus commissioned his Twelve Apostles to, among other things, raise the dead. Similar resurrections are credited to the apostles and Catholic saints. In the Acts of the Apostles, Saint Peter raised a woman named Dorcas (also called Tabitha), and Paul the Apostle revived a man named Eutychus who had fallen asleep and fell from a window to his death. According to the Gospel of Matthew, after Jesus's resurrection, many of those previously dead came out of their tombs and entered Jerusalem, where they appeared to many. Following the Apostolic Age, many saints were said to resurrect the dead, as recorded in Orthodox Christian hagiographies. St Columba supposedly raised a boy from the dead in the land of Picts. Christians regard the resurrection of Jesus as the central doctrine in Christianity. Others take the incarnation of Jesus to be more central; however, it is the miracles – and particularly his resurrection – which provide validation of his incarnation. According to Paul, the entire Christian faith hinges upon the centrality of the resurrection of Jesus and the hope for a life after death. The Apostle Paul wrote in his first letter to the Corinthians: Christianity started as a religious movement within 1st-century Judaism (late Second Temple Judaism), and it retains what the New Testament itself claims was the Pharisaic belief in the afterlife and resurrection of the dead. Whereas this belief was only one of many beliefs held about the world to come in Second Temple Judaism, and was notably rejected by the Sadducees, but accepted by the Pharisees (cf. Acts 23:6-8). Belief in the resurrection became dominant within Early Christianity and already in the Gospels of Luke and John included an insistence on the resurrection of the flesh. Most modern Christian churches continue to uphold the belief that there will be a final resurrection of the dead and world to come. Belief in the resurrection of the dead, and Jesus' role as judge, is codified in the Apostles' Creed, which is the fundamental creed of Christian baptismal faith. The Book of Revelation also makes many references about the Day of Judgment when the dead will be raised. The emphasis on the literal resurrection of the flesh remained strong in the medieval ages, and still remains so in Orthodox churches. In modern Western Christianity, especially "from the 17th to the 19th century, the language of popular piety no longer evoked the resurrection of the soul but everlasting life. Although theological textbooks still mentioned resurrection, they dealt with it as a speculative question more than as an existential problem." In Platonic philosophy and other Greek philosophical thought, at death the soul was said to leave the inferior body behind. The idea that Jesus was resurrected spiritually rather than physically even gained popularity among some Christian teachers, whom the author of 1 John declared to be antichrists. Similar beliefs appeared in the early church as Gnosticism. However, in Luke 24:39, the resurrected Jesus expressly states "behold my hands and my feet, that it is I myself. Handle me and see, for a spirit does not have flesh and bones as you see I have." There are folklore, stories, and extractions from certain holy texts that refer to resurrections. One major folklore is that of Savitri saving her husband's life from Yamraj. In the Ramayana, after Ravana was slayed by Rama in a great battle between good and evil, Rama requests the king of Gods, Indra, to restore the lives of all the monkeys who died in the great battle. Belief in the "Day of Resurrection" ("Yawm al-Qiyāmah"; ) is also crucial for Muslims. They believe the time of "Qiyāmah" is preordained by God but unknown to man. The trials and tribulations preceding and during the "Qiyāmah" are described in the Qur'an and the hadith, and also in the commentaries of scholars. The Quran emphasizes bodily resurrection, a break from the pre-Islamic Arabian understanding of death.
https://en.wikipedia.org/wiki?curid=26447
Robert Parr Robert Ghormley Parr (September 22, 1921 – March 27, 2017) was an American theoretical chemist who was a Professor of Chemistry at the University of North Carolina at Chapel Hill. Parr received an A. B. degree "magna cum laude" from Brown University in 1942, and then entered the University of Minnesota, receiving a Ph.D. in physical chemistry in 1947. He joined the faculty at Minnesota upon receiving his Ph.D. and remained there one year. In 1948 he moved to the Carnegie Institute of Technology (now Carnegie Mellon University) in Pittsburgh, Pennsylvania, becoming a full professor in 1957. In 1962 he moved to Johns Hopkins University in Baltimore, Maryland, and in 1974 to the University of North Carolina at Chapel Hill, where he received appointment to an endowed professorship in 1990 and where he last taught. Working with DuPont chemist Rudolph Pariser, Parr developed a method of computing approximate molecular orbitals for pi electron systems, published in 1953. Since an identical procedure was derived by John A. Pople the same year, it is generally referred to as the Pariser–Parr–Pople method or PPP method. The PPP method differed from existing structural chemistry thinking (which advocated "maximum overlap principle") by advancing the concept of "zero differential overlap approximation". By 1978 Parr had realized that density functional theory (DFT) would be extremely useful in quantitative calculations of chemical and biological systems, especially those with high molecular weights. In 1988 Parr, Weitao Yang and Chengteh Lee produced an improved DFT method which could approximate the correlation energy of systems. The LYP functional theory is now one of the most-often cited papers in the chemical literature. In 1963 Parr published "Quantum Theory of Molecular Electronic Structure", one of the first books to apply quantum theory to chemical systems. In 1989 he and Yang published "Density Functional Theory of Atoms and Molecules", now considered the basic textbook on DFT.
https://en.wikipedia.org/wiki?curid=26451
Riesz representation theorem Riesz representation theorem, sometimes called Riesz–Fréchet representation theorem, named after Frigyes Riesz and Maurice René Fréchet, establishes an important connection between a Hilbert space and its continuous dual space. If the underlying field is the real numbers, the two are isometrically isomorphic; if the underlying field is the complex numbers, the two are isometrically anti-isomorphic. The (anti-) isomorphism is a particular natural one as will be described next; a natural isomorphism. Let "H" be a Hilbert space, and let "H*" denote its dual space, consisting of all bounded operators from "H" into the field formula_1 or formula_2. If formula_3 is an element of "H", then the function formula_4 for all formula_5 in "H" defined by formula_6 where formula_7 denotes the inner product of the Hilbert space, is an element of "H*". The Riesz representation theorem states that "every" element of "H*" can be written uniquely in this form. Let formula_8 be a Hilbert space and formula_9. Then there exists formula_10 such that for any formula_11, formula_12. Moreover formula_13 Let formula_14. Clearly formula_15 is closed subspace of formula_8. If formula_17, then we can trivially choose formula_18. Now assume formula_19. Then formula_20 is one-dimensional. Indeed, let formula_21 be nonzero vectors in formula_20. Then there is nonzero real number formula_23, such that formula_24. Observe that formula_25 and formula_26, so formula_27. This means that formula_28. Now let formula_29 be unit vector in formula_20. For arbitrary formula_11, let formula_32 be the orthogonal projection of formula_3 onto formula_20. Then formula_35 and formula_36 (from the properties of orthogonal projections), so that formula_37 and formula_38. Thus formula_39. Hence formula_40. We also see formula_41. From the Cauchy-Bunyakovsky-Schwartz inequality formula_42, thus for formula_3 with unit norm formula_44. This implies that formula_45. Given any continuous linear functional "g" in "H*", the corresponding element formula_46 can be constructed uniquely by formula_47, where formula_48 is an orthonormal basis of "H", and the value of formula_49 does not vary by choice of basis. Thus, if formula_50, then formula_51 The mapping formula_52: "H" → "H*" defined by formula_53 = formula_54 is an isometric (anti-) isomorphism, meaning that: The inverse map of formula_52 can be described as follows. Given a non-zero element formula_68 of "H*", the orthogonal complement of the kernel of formula_68 is a one-dimensional subspace of "H". Take a non-zero element "z" in that subspace, and set formula_70. Then formula_53 = formula_68. Historically, the theorem is often attributed simultaneously to Riesz and Fréchet in 1907 (see references). In the mathematical treatment of quantum mechanics, the theorem can be seen as a justification for the popular bra–ket notation. The theorem says that, every bra formula_73 has a corresponding ket formula_74, and the latter is unique.
https://en.wikipedia.org/wiki?curid=26452
Romano Scarpa Romano Scarpa (September 27, 1927, Venice – April 23, 2005, Málaga) was one of the most famous Italian creators of Disney comics. Growing up in Venice he developed a particular love for American cartoons and Disney comics, that, at the time, were published in the big format of the Topolino Giornale which was then printing now classic Floyd Gottfredson's stories. In the Forties he opened an Animation Studio in Venice in which he produced his first works: some commercials, a short titled "E poi venne il diluvio" and another one titled "La piccola fiammiferaia" (1953, based on Hans Christian Andersen's "The Little Match Girl"), distributed in Italy together with Robert Aldrich's "Attack!" (1956). Right after that he stopped working in animation for a while and dedicated wholly to creating Disney comics. When in 1956 Italian editors had no more new Floyd Gottfredson's stories to reprint, he was given the responsibility to continue Gottfredson's stories about Mickey Mouse. Also influenced by Carl Barks in the late Fifties and up to about 1963 he wrote and penciled stories like "Topolino e la collana Chirikawa" (1960) or "The Flying Scot" (1957) that have, later, been translated in many different languages throughout the world. Many of these stories have their backgrounds in movies, for example "Topolino nel favoloso regno di Shan Grillà" (1961) is based upon Frank Capra's "Lost Horizon" (1937); not to talk about all the stories starring Snow White or the Seven Dwarfs, obviously based on "Snow White and the Seven Dwarfs" (1937). Sometimes the exact opposite happened; the Italian movie "Riusciranno i nostri eroi a ritrovare l'amico misteriosamente scomparso in Africa?" (1968) is based on Scarpa's story "Topolino e il Pippotarzan" (1957). Around 1963, Scarpa stopped writing for 6 or 7 years. In the seventies, he moved to Spain and started working for a different publisher. Among the last things he made while he was still in Italy, at the end of the Eighties and at beginning of the Nineties, there are the so-called "Paperolimpiadi" (a long story about the 1988 Seoul Olympic games) and some strip stories, the same kind of stories that he loved when he was a child. One of these, "Topolino e l'enigma di Brigaboom" (1989) was partially based on "Brigadoon" (1954). In the meanwhile he has had time enough for some more animation, so we have "Aihnoo degli Icebergs" (1972), "The Fourth King" (1977) and a new TV series, "The Adventures of Marco and Gina" ("Sopra i tetti di Venezia") (2001). Mainly Scarpa worked on Disney comics, but many years ago he used to do something non-Disney once in a while, so he did one (Rolf Kauka's) "Lupo" story and one (Hannah and Barbera's) "Yogi Bear" story. In the 1950s he also drew some "Angelino" story, and Italian character. Since 1988 some of his comic stories have been published in the USA by Gladstone Publishing; it was the first time that this happened to an Italian Disney author. Later, when Disney Comics took Gladstone's place, they published some more of his stories, and in 2003, the same happened with Gemstone Publishing, that is publishing his stories in the US at the moment. He has influenced many younger creators (Giorgio Cavazzano was his inker during the Sixties) and many have attempted to imitate his style. In his career Scarpa created many Disney characters that are now accepted by some as part of the Disney Universe. Those include, but are not limited to: In 2017 Fantagraphics Books published a collection containing four stories of Scarpa's Snow White comics, titled "The Return of Snow-White and the Seven Dwarfs", ISBN . In 2018 Fantagraphics Books began publishing a hardcover series titled "Disney Masters", in which Romano Scarpa has to date (October 2019) had three volumes dedicated to his Disney works. This is an index of all Romano Scarpa comics published in the US. Only Duck universe and Mouse universe are listed. Chip and Dale comics are not listed.
https://en.wikipedia.org/wiki?curid=26455
Rosa Parks Rosa Louise McCauley Parks (February 4, 1913 – October 24, 2005) was an American activist in the civil rights movement best known for her pivotal role in the Montgomery bus boycott. The United States Congress has called her "the first lady of civil rights" and "the mother of the freedom movement". On December 1, 1955, in Montgomery, Alabama, Parks rejected bus driver James F. Blake's order to relinquish her seat in the "colored section" to a white passenger, after the whites-only section was filled. Parks was not the first person to resist bus segregation, but the National Association for the Advancement of Colored People (NAACP) believed that she was the best candidate for seeing through a court challenge after her arrest for civil disobedience in violating Alabama segregation laws. Parks' prominence in the community and her willingness to become a controversial figure inspired the black community to boycott the Montgomery buses for over a year, the first major direct action campaign of the post-war civil rights movement. Her case became bogged down in the state courts, but the federal Montgomery bus lawsuit "Browder v. Gayle" succeeded in November 1956. Parks' act of defiance and the Montgomery bus boycott became important symbols of the movement. She became an international icon of resistance to racial segregation. She organized and collaborated with civil rights leaders, including Edgar Nixon, president of the local chapter of the NAACP; and Martin Luther King Jr., a new minister in Montgomery who gained national prominence in the civil rights movement and went on to win a Nobel Peace Prize. Though employed as a seamstress at a local department store at the time, Parks was also secretary of the Montgomery chapter of the NAACP. She had recently attended the Highlander Folk School, a Tennessee center for training activists for workers' rights and racial equality. She acted as a private citizen "tired of giving in". Although widely honored in later years, she also suffered for her act; she was fired from her job, and received death threats for years afterwards. Shortly after the boycott, she moved to Detroit, where she briefly found similar work. From 1965 to 1988, she served as secretary and receptionist to John Conyers, an African-American US Representative. She was also active in the Black Power movement and the support of political prisoners in the US. After retirement, Parks wrote her autobiography and continued to insist that the struggle for justice was not over and there was more work to be done. In her final years, she suffered from dementia. Parks received national recognition, including the NAACP's 1979 Spingarn Medal, the Presidential Medal of Freedom, the Congressional Gold Medal, and a posthumous statue in the United States Capitol's National Statuary Hall. Upon her death in 2005, she was the first woman to lie in honor in the Capitol Rotunda, becoming the thirty-first person to receive this honor. California and Missouri commemorate Rosa Parks Day on her birthday, February 4, while Ohio and Oregon commemorate the occasion on the anniversary of the day she was arrested, December 1. Rosa Parks was born Rosa Louise McCauley in Tuskegee, Alabama, on February 4, 1913, to Leona (née Edwards), a teacher, and James McCauley, a carpenter. In addition to African ancestry, one of Parks' great-grandfathers was Scots-Irish and one of her great-grandmothers a part-Native American slave. She was small as a child and suffered poor health with chronic tonsillitis. When her parents separated, she moved with her mother to Pine Level, just outside the state capital, Montgomery. She grew up on a farm with her maternal grandparents, mother, and younger brother Sylvester. They all were members of the African Methodist Episcopal Church (AME), a century-old independent black denomination founded by free blacks in Philadelphia, Pennsylvania, in the early nineteenth century. McCauley attended rural schools until the age of eleven. As a student at the Industrial School for Girls in Montgomery, she took academic and vocational courses. Parks went on to a laboratory school set up by the Alabama State Teachers College for Negroes for secondary education, but dropped out in order to care for her grandmother and later her mother, after they became ill. Around the turn of the 20th century, the former Confederate states had adopted new constitutions and electoral laws that effectively disenfranchised black voters and, in Alabama, many poor white voters as well. Under the white-established Jim Crow laws, passed after Democrats regained control of southern legislatures, racial segregation was imposed in public facilities and retail stores in the South, including public transportation. Bus and train companies enforced seating policies with separate sections for blacks and whites. School bus transportation was unavailable in any form for black schoolchildren in the South, and black education was always underfunded. Parks recalled going to elementary school in Pine Level, where school buses took white students to their new school and black students had to walk to theirs: I'd see the bus pass every day ... But to me, that was a way of life; we had no choice but to accept what was the custom. The bus was among the first ways I realized there was a black world and a white world. Although Parks' autobiography recounts early memories of the kindness of white strangers, she could not ignore the racism of her society. When the Ku Klux Klan marched down the street in front of their house, Parks recalls her grandfather guarding the front door with a shotgun. The Montgomery Industrial School, founded and staffed by white northerners for black children, was burned twice by arsonists. Its faculty was ostracized by the white community. Repeatedly bullied by white children in her neighborhood, Parks often fought back physically. She later said: "As far back as I remember, I could never think in terms of accepting physical abuse without some form of retaliation if possible." In 1932, Rosa married Raymond Parks, a barber from Montgomery. He was a member of the NAACP, which at the time was collecting money to support the defense of the Scottsboro Boys, a group of black men falsely accused of raping two white women. Rosa took numerous jobs, ranging from domestic worker to hospital aide. At her husband's urging, she finished her high school studies in 1933, at a time when less than 7% of African Americans had a high-school diploma. In December 1943, Parks became active in the civil rights movement, joined the Montgomery chapter of the NAACP, and was elected secretary at a time when this was considered a woman's job. She later said, "I was the only woman there, and they needed a secretary, and I was too timid to say no." She continued as secretary until 1957. She worked for the local NAACP leader Edgar Nixon, even though he maintained that "Women don't need to be nowhere but in the kitchen." When Parks asked, "Well, what about me?", he replied: "I need a secretary and you are a good one." In 1944, in her capacity as secretary, she investigated the gang-rape of Recy Taylor, a black woman from Abbeville, Alabama. Parks and other civil rights activists organized "The Committee for Equal Justice for Mrs. Recy Taylor", launching what the "Chicago Defender" called "the strongest campaign for equal justice to be seen in a decade." Although never a member of the Communist Party, she attended meetings with her husband. The notorious Scottsboro case had been brought to prominence by the Communist Party. In the 1940s, Parks and her husband were members of the League of Women Voters. Sometime soon after 1944, she held a brief job at Maxwell Air Force Base, which, despite its location in Montgomery, Alabama, did not permit racial segregation because it was federal property. She rode on its integrated trolley. Speaking to her biographer, Parks noted, "You might just say Maxwell opened my eyes up." Parks worked as a housekeeper and seamstress for Clifford and Virginia Durr, a white couple. Politically liberal, the Durrs became her friends. They encouraged—and eventually helped sponsor—Parks in the summer of 1955 to attend the Highlander Folk School, an education center for activism in workers' rights and racial equality in Monteagle, Tennessee. There Parks was mentored by the veteran organizer Septima Clark. In 1945, despite the Jim Crow laws and discrimination by registrars, she succeeded in registering to vote on her third try. In August 1955, black teenager Emmett Till was brutally murdered after reportedly flirting with a young white woman while visiting relatives in Mississippi. On November 27, 1955, four days before she would make her stand on the bus, Rosa Parks attended a mass meeting at Dexter Avenue Baptist Church in Montgomery that addressed this case, as well as the recent murders of the activists George W. Lee and Lamar Smith. The featured speaker was T. R. M. Howard, a black civil rights leader from Mississippi who headed the Regional Council of Negro Leadership. Howard brought news of the recent acquittal of the two men who had murdered Till. Parks was deeply saddened and angry at the news, particularly because Till's case had garnered much more attention than any of the cases she and the Montgomery NAACP had worked on—and yet, the two men still walked free. In 1900, Montgomery had passed a city ordinance to segregate bus passengers by race. Conductors were empowered to assign seats to achieve that goal. According to the law, no passenger would be required to move or give up their seat and stand if the bus was crowded and no other seats were available. Over time and by custom, however, Montgomery bus drivers adopted the practice of requiring black riders to move when there were no white-only seats left. The first four rows of seats on each Montgomery bus were reserved for whites. Buses had "colored" sections for black people generally in the rear of the bus, although blacks composed more than 75% of the ridership. The sections were not fixed but were determined by placement of a movable sign. Black people could sit in the middle rows until the white section filled; if more whites needed seats, blacks were to move to seats in the rear, stand, or, if there was no room, leave the bus. Black people could not sit across the aisle in the same row as white people. The driver could move the "colored" section sign, or remove it altogether. If white people were already sitting in the front, black people had to board at the front to pay the fare, then disembark and reenter through the rear door. For years, the black community had complained that the situation was unfair. Parks said, "My resisting being mistreated on the bus did not begin with that particular arrest. I did a lot of walking in Montgomery." One day in 1943, Parks boarded a bus and paid the fare. She then moved to a seat, but driver James F. Blake told her to follow city rules and enter the bus again from the back door. When Parks exited the vehicle, Blake drove off without her. Parks waited for the next bus, determined never to ride with Blake again. After working all day, Parks boarded the Cleveland Avenue bus, a General Motors Old Look bus belonging to the Montgomery City Lines, around 6 p.m., Thursday, December 1, 1955, in downtown Montgomery. She paid her fare and sat in an empty seat in the first row of back seats reserved for blacks in the "colored" section. Near the middle of the bus, her row was directly behind the ten seats reserved for white passengers. Initially, she did not notice that the bus driver was the same man, James F. Blake, who had left her in the rain in 1943. As the bus traveled along its regular route, all of the white-only seats in the bus filled up. The bus reached the third stop in front of the Empire Theater, and several white passengers boarded. Blake noted that two or three white passengers were standing, as the front of the bus had filled to capacity. He moved the "colored" section sign behind Parks and demanded that four black people give up their seats in the middle section so that the white passengers could sit. Years later, in recalling the events of the day, Parks said, "When that white driver stepped back toward us, when he waved his hand and ordered us up and out of our seats, I felt a determination cover my body like a quilt on a winter night." By Parks' account, Blake said, "Y'all better make it light on yourselves and let me have those seats." Three of them complied. Parks said, "The driver wanted us to stand up, the four of us. We didn't move at the beginning, but he says, 'Let me have these seats.' And the other three people moved, but I didn't." The black man sitting next to her gave up his seat. Parks moved, but toward the window seat; she did not get up to move to the redesignated colored section. Parks later said about being asked to move to the rear of the bus, "I thought of Emmett Till – a 14-year-old African American who was lynched in Mississippi in 1955, after being accused of offending a white woman in her family's grocery store, whose killers were tried and acquitted – and I just couldn't go back." Blake said, "Why don't you stand up?" Parks responded, "I don't think I should have to stand up." Blake called the police to arrest Parks. When recalling the incident for "Eyes on the Prize", a 1987 public television series on the Civil Rights Movement, Parks said, "When he saw me still sitting, he asked if I was going to stand up, and I said, 'No, I'm not.' And he said, 'Well, if you don't stand up, I'm going to have to call the police and have you arrested.' I said, 'You may do that.'" During a 1956 radio interview with Sydney Rogers in West Oakland several months after her arrest, Parks said she had decided, "I would have to know for once and for all what rights I had as a human being and a citizen." In her autobiography, "My Story", she said: When Parks refused to give up her seat, a police officer arrested her. As the officer took her away, she recalled that she asked, "Why do you push us around?" She remembered him saying, "I don't know, but the law's the law, and you're under arrest." She later said, "I only knew that, as I was being arrested, that it was the very last time that I would ever ride in humiliation of this kind. ... " Parks was charged with a violation of Chapter 6, Section 11 segregation law of the Montgomery City code, although technically she had not taken a white-only seat; she had been in a colored section. Edgar Nixon, president of the Montgomery chapter of the NAACP and leader of the Pullman Porters Union, and her friend Clifford Durr bailed Parks out of jail that evening. Parks did not originate the idea of protesting segregation with a bus sit-in. Those preceding her included Bayard Rustin in 1942, Irene Morgan in 1946, Lillie Mae Bradford in 1951, Sarah Louise Keys in 1952, and the members of the ultimately successful "Browder v. Gayle" 1956 lawsuit (Claudette Colvin, Aurelia Browder, Susie McDonald, and Mary Louise Smith) who were arrested in Montgomery for not giving up their bus seats months before Parks. Nixon conferred with Jo Ann Robinson, an Alabama State College professor and member of the Women's Political Council (WPC), about the Parks case. Robinson believed it important to seize the opportunity and stayed up all night mimeographing over 35,000 handbills announcing a bus boycott. The Women's Political Council was the first group to officially endorse the boycott. On Sunday, December 4, 1955, plans for the Montgomery bus boycott were announced at black churches in the area, and a front-page article in the "Montgomery Advertiser" helped spread the word. At a church rally that night, those attending agreed unanimously to continue the boycott until they were treated with the level of courtesy they expected, until black drivers were hired, and until seating in the middle of the bus was handled on a first-come basis. The next day, Parks was tried on charges of disorderly conduct and violating a local ordinance. The trial lasted 30 minutes. After being found guilty and fined $10, plus $4 in court costs (combined total ), Parks appealed her conviction and formally challenged the legality of racial segregation. In a 1992 interview with National Public Radio's Lynn Neary, Parks recalled: On the day of Parks' trial—December 5, 1955—the WPC distributed the 35,000 leaflets. The handbill read, We are ... asking every Negro to stay off the buses Monday in protest of the arrest and trial ... You can afford to stay out of school for one day. If you work, take a cab, or walk. But please, children and grown-ups, don't ride the bus at all on Monday. Please stay off the buses Monday. It rained that day, but the black community persevered in their boycott. Some rode in carpools, while others traveled in black-operated cabs that charged the same fare as the bus, 10 cents (). Most of the remainder of the 40,000 black commuters walked, some as far as . That evening after the success of the one-day boycott, a group of 16 to 18 people gathered at the Mt. Zion AME Zion Church to discuss boycott strategies. At that time, Parks was introduced but not asked to speak, despite a standing ovation and calls from the crowd for her to speak; when she asked if she should say something, the reply was, "Why, you've said enough." The group agreed that a new organization was needed to lead the boycott effort if it were to continue. Rev. Ralph Abernathy suggested the name "Montgomery Improvement Association" (MIA). The name was adopted, and the MIA was formed. Its members elected as their president Martin Luther King Jr., a relative newcomer to Montgomery, who was a young and mostly unknown minister of the Dexter Avenue Baptist Church. That Monday night, 50 leaders of the African-American community gathered to discuss actions to respond to Parks' arrest. Edgar Nixon, the president of the NAACP, said, "My God, look what segregation has put in my hands!" Parks was considered the ideal plaintiff for a test case against city and state segregation laws, as she was seen as a responsible, mature woman with a good reputation. She was securely married and employed, was regarded as possessing a quiet and dignified demeanor, and was politically savvy. King said that Parks was regarded as "one of the finest citizens of Montgomery—not one of the finest Negro citizens, but one of the finest citizens of Montgomery." Parks' court case was being slowed down in appeals through the Alabama courts on their way to a Federal appeal and the process could have taken years. Holding together a boycott for that length of time would have been a great strain. In the end, black residents of Montgomery continued the boycott for 381 days. Dozens of public buses stood idle for months, severely damaging the bus transit company's finances, until the city repealed its law requiring segregation on public buses following the US Supreme Court ruling in "Browder v. Gayle" that it was unconstitutional. Parks was not included as a plaintiff in the Browder decision because the attorney Fred Gray concluded the courts would perceive they were attempting to circumvent her prosecution on her charges working their way through the Alabama state court system. Parks played an important part in raising international awareness of the plight of African Americans and the civil rights struggle. King wrote in his 1958 book "Stride Toward Freedom" that Parks' arrest was the catalyst rather than the cause of the protest: "The cause lay deep in the record of similar injustices." He wrote, "Actually, no one can understand the action of Mrs. Parks unless he realizes that eventually the cup of endurance runs over, and the human personality cries out, 'I can take it no longer.'" After her arrest, Parks became an icon of the Civil Rights Movement but suffered hardships as a result. Due to economic sanctions used against activists, she lost her job at the department store. Her husband quit his job after his boss forbade him to talk about his wife or the legal case. Parks traveled and spoke extensively about the issues. In 1957, Raymond and Rosa Parks left Montgomery for Hampton, Virginia; mostly because she was unable to find work. She also disagreed with King and other leaders of Montgomery's struggling civil rights movement about how to proceed, and was constantly receiving death threats. In Hampton, she found a job as a hostess in an inn at Hampton Institute, a historically black college. Later that year, at the urging of her brother and sister-in-law in Detroit, Sylvester and Daisy McCauley, Rosa and Raymond Parks and her mother moved north to join them. The City of Detroit attempted to cultivate a progressive reputation, but Parks encountered numerous signs of discrimination against African-Americans. Schools were effectively segregated, and services in black neighborhoods substandard. In 1964, Parks told an interviewer that, "I don't feel a great deal of difference here ... Housing segregation is just as bad, and it seems more noticeable in the larger cities." She regularly participated in the movement for open and fair housing. Parks rendered crucial assistance in the first campaign for Congress by John Conyers. She persuaded Martin Luther King (who was generally reluctant to endorse local candidates) to appear with Conyers, thereby boosting the novice candidate's profile. When Conyers was elected, he hired her as a secretary and receptionist for his congressional office in Detroit. She held this position until she retired in 1988. In a telephone interview with CNN on October 24, 2005, Conyers recalled, "You treated her with deference because she was so quiet, so serene—just a very special person ... There was only one Rosa Parks." Doing much of the daily constituent work for Conyers, Parks often focused on socio-economic issues including welfare, education, job discrimination, and affordable housing. She visited schools, hospitals, senior citizen facilities, and other community meetings and kept Conyers grounded in community concerns and activism. Parks participated in activism nationally during the mid-1960s, traveling to support the Selma-to-Montgomery Marches, the Freedom Now Party, and the Lowndes County Freedom Organization. She also befriended Malcolm X, who she regarded as a personal hero. Like many Detroit blacks, Parks remained particularly concerned about housing issues. She herself lived in a neighborhood, Virginia Park, which had been compromised by highway construction and urban renewal. By 1962, these policies had destroyed 10,000 structures in Detroit, displacing 43,096 people, 70 percent of them African-American. Parks lived just a mile from the center of the riot that took place in Detroit in 1967, and she considered housing discrimination a major factor that provoked the disorder. In the aftermath Parks collaborated with members of the League of Revolutionary Black Workers and the Republic of New Afrika in raising awareness of police abuse during the conflict. She served on a "people's tribunal" on August 30, 1967, investigating the killing of three young men by police during the 1967 Detroit uprising, in what came to be known as the Algiers Motel incident. She also helped form the Virginia Park district council to help rebuild the area. The council facilitated the building of the only black-owned shopping center in the country. Parks took part in the black power movement, attending the Philadelphia Black Power conference, and the Black Political Convention in Gary, Indiana. She also supported and visited the Black Panther school in Oakland. In the 1970s, Parks organized for the freedom of political prisoners in the United States, particularly cases involving issues of self-defense. She helped found the Detroit chapter of the Joann Little Defense Committee, and also worked in support of the Wilmington 10, the RNA 11, and Gary Tyler. Following national outcry around her case, Little succeeded in her defense that she used deadly force to resist sexual assault and was acquitted. Gary Tyler was finally released in April 2016 after 41 years in prison. The 1970s were a decade of loss for Parks in her personal life. Her family was plagued with illness; she and her husband had suffered stomach ulcers for years and both required hospitalization. In spite of her fame and constant speaking engagements, Parks was not a wealthy woman. She donated most of the money from speaking to civil rights causes, and lived on her staff salary and her husband's pension. Medical bills and time missed from work caused financial strain that required her to accept assistance from church groups and admirers. Her husband died of throat cancer on August 19, 1977, and her brother, her only sibling, died of cancer that November. Her personal ordeals caused her to become removed from the civil rights movement. She learned from a newspaper of the death of Fannie Lou Hamer, once a close friend. Parks suffered two broken bones in a fall on an icy sidewalk, an injury which caused considerable and recurring pain. She decided to move with her mother into an apartment for senior citizens. There she nursed her mother Leona through the final stages of cancer and geriatric dementia until she died in 1979 at the age of 92. In 1980, Parks—widowed and without immediate family—rededicated herself to civil rights and educational organizations. She co-founded the Rosa L. Parks Scholarship Foundation for college-bound high school seniors, to which she donated most of her speaker fees. In February 1987, she co-founded, with Elaine Eason Steele, the Rosa and Raymond Parks Institute for Self Development, an institute that runs the "Pathways to Freedom" bus tours which introduce young people to important civil rights and Underground Railroad sites throughout the country. Parks also served on the Board of Advocates of Planned Parenthood. Though her health declined as she entered her seventies, Parks continued to make many appearances and devoted considerable energy to these causes. In 1992, Parks published "Rosa Parks: My Story", an autobiography aimed at younger readers, which recounts her life leading to her decision to keep her seat on the bus. A few years later, she published "Quiet Strength" (1995), her memoir, which focuses on her faith. At age 81, Parks was robbed and assaulted in her home in central Detroit on August 30, 1994. The assailant, Joseph Skipper, broke down the door but claimed he had chased away an intruder. He requested a reward and when Parks paid him, he demanded more. Parks refused and he attacked her. Hurt and badly shaken, Parks called a friend, who called the police. A neighborhood manhunt led to Skipper's capture and reported beating. Parks was treated at Detroit Receiving Hospital for facial injuries and swelling on the right side of her face. Parks said about the attack on her by the African-American man, "Many gains have been made ... But as you can see, at this time we still have a long way to go." Skipper was sentenced to 8 to 15 years and was transferred to prison in another state for his own safety. Suffering anxiety upon returning to her small central Detroit house following the ordeal, Parks moved into Riverfront Towers, a secure high-rise apartment building. Learning of Parks' move, Little Caesars owner Mike Ilitch offered to pay for her housing expenses for as long as necessary. In 1994, the Ku Klux Klan applied to sponsor a portion of United States Interstate 55 in St. Louis County and Jefferson County, Missouri, near St. Louis, for cleanup (which allowed them to have signs stating that this section of highway was maintained by the organization). Since the state could not refuse the KKK's sponsorship, the Missouri legislature voted to name the highway section the "Rosa Parks Highway". When asked how she felt about this honor, she is reported to have commented, "It is always nice to be thought of." In 1999, Parks filmed a cameo appearance for the television series "Touched by an Angel". It was her last appearance on film; Parks began to suffer from health problems due to old age. In 2002, Parks received an eviction notice from her $1,800 per month () apartment for non-payment of rent. Parks was incapable of managing her own financial affairs by this time due to age-related physical and mental decline. Her rent was paid from a collection taken by Hartford Memorial Baptist Church in Detroit. When her rent became delinquent and her impending eviction was highly publicized in 2004, executives of the ownership company announced they had forgiven the back rent and would allow Parks, by then 91 and in extremely poor health, to live rent-free in the building for the remainder of her life. Elaine Steele, manager of the nonprofit Rosa and Raymond Parks Institute, told the newspaper that Parks got proper care, and that eviction notices were sent in error in 2002. Her heirs and various interest organizations alleged at the time that her financial affairs had been mismanaged. In 2016, Parks's former residence in Detroit was threatened with demolition. A Berlin-based American artist, Ryan Mendoza, arranged to have the house disassembled, moved to his garden in Germany, and partly restored. It currently serves as a museum honoring Rosa Parks. Parks died of natural causes on October 24, 2005, at the age of 92, in her apartment on the east side of Detroit. She and her husband never had children and she outlived her only sibling. She was survived by her sister-in-law (Raymond's sister), 13 nieces and nephews and their families, and several cousins, most of them residents of Michigan or Alabama. City officials in Montgomery and Detroit announced on October 27, 2005, that the front seats of their city buses would be reserved with black ribbons in honor of Parks until her funeral. Parks' coffin was flown to Montgomery and taken in a horse-drawn hearse to the St. Paul African Methodist Episcopal (AME) church, where she lay in repose at the altar on October 29, 2005, dressed in the uniform of a church deaconess. A memorial service was held there the following morning. One of the speakers, United States Secretary of State Condoleezza Rice, said that if it had not been for Parks, she would probably have never become the Secretary of State. In the evening the casket was transported to Washington, D.C. and transported by a bus similar to the one in which she made her protest, to lie in honor in the rotunda of the U.S. Capitol. Since the founding of the practice in 1852, Parks was the 31st person, the first American who had not been a U.S. government official, and the second private person (after the French planner Pierre L'Enfant) to be honored in this way. She was the first woman and the second black person to lie in honor in the Capitol. An estimated 50,000 people viewed the casket there, and the event was broadcast on television on October 31, 2005. A memorial service was held that afternoon at Metropolitan AME Church in Washington, DC. With her body and casket returned to Detroit, for two days, Parks lay in repose at the Charles H. Wright Museum of African American History. Her funeral service was seven hours long and was held on November 2, 2005, at the Greater Grace Temple Church in Detroit. After the service, an honor guard from the Michigan National Guard laid the U.S. flag over the casket and carried it to a horse-drawn hearse, which was intended to carry it, in daylight, to the cemetery. As the hearse passed the thousands of people who were viewing the procession, many clapped, cheered loudly and released white balloons. Parks was interred between her husband and mother at Detroit's Woodlawn Cemetery in the chapel's mausoleum. The chapel was renamed the Rosa L. Parks Freedom Chapel in her honor. Parks had previously prepared and placed a headstone on the selected location with the inscription "Rosa L. Parks, wife, 1913–". Multimedia and interviews Others
https://en.wikipedia.org/wiki?curid=26458
Ringworld (role-playing game) The Ringworld science fiction role-playing game was published by Chaosium in 1984, using the Basic Role-Playing system for its rules and Larry Niven's "Ringworld" novels as a setting. The setting is a distant future based on extrapolation of as much hard science as Niven had available. Specifically, it's the 29th century. "Known Space" (also the commonly used title for Larry Niven's future history science fiction series) is about 80 light years in diameter with 10,000 stars, including Human Space (40 light years diameter, 524 stars in 357 systems, 30 billion humans, ⅔ on Earth), as well as neighbouring Alien civilisations. Important Alien civilisations include the Puppeteers, paranoid pacifist herbivore centaurs, and the Kzinti, carnivorous warlike felines, who fought multiple wars over hundreds of years against the Humans, being defeated each time. Human allies include intelligent dolphins and orcas. "Known Space" only serves as a background for the game. The game is intended to be set on the Ringworld itself, an enormous single world discovered at the far reaches of Known Space, a ring around a sun at approximately the orbit of the Earth. It is 997,000 miles wide, about 125 Earth-diameters. The total inner surface of the ring is equal to that of 3 million Earths. The ring is spun at a speed to provide 0.992G of gravity on the innerside, while 20 giant shadow squares at about the orbit of Mercury occlude the Sun to provide night. It was constructed by the Pak Protectors, now mostly extinct, who had a common origin with humans. The Ringworld is home to some 30 trillion sentient inhabitants from up to 2000 hominid species. The world is described in a series of novels by Niven, "Ringworld", "The Ringworld Engineers", and, after the game's publication, "The Ringworld Throne" and "Ringworld's Children". The role-playing game contains a great deal of technical details about the setting, more than the fiction the setting is based on. The game setting details are complete enough that some "Ringworld" fans not interested in role-playing buy the game just for the background material. Information from the RPG, along with notes composed by RPG author Hewitt with Niven, were later used to form the "Bible" given to authors writing in the "Man-Kzin Wars" series. Niven himself recommended that Hewitt write one of the stories for the original two MKW books, although this never came to pass. The players initially play explorers from Known Space, sent as scouts to the Ringworld. They can be anthropologists, artists, doctors, police, or even zealots, who will explore the mysteries of this huge artificial world and its inhabitants. Basic characters can be humans from a dozen planets of Human Space, Puppeteers, or Kzin. Later play can see characters from Ringworld species, such as the (so-called) Ghouls, Vampires, Giants, Sea People, and others. This Ringworld focus has been a criticism of the game. The "Ringworld" role-playing game is not a 'full' science fiction RPG, like "Traveller", including, for example, rules for starship construction, space combat, travel to different planets and systems, and so forth. Instead, the game and rules focused on parties of characters exploring the Ringworld itself, and, despite its vast size (with a surface area larger than that of all of Known Space's inhabited planets put together), many who bought the game felt limited by this one world setting. A character is initially defined by his species or world of origin, which affects characteristics (for example, by determining the gravity to which it is accustomed). Then the players roll randomly for a certain number of defects, character age, and characteristics. The system used is Chaosium's Basic Role-Playing, with eight basic characteristics: Strength, Constitution, Mass (equivalent to Size in other BRP games), Intelligence, Power, Dexterity, Appearance and Education determining secondary attributes such damage modifiers, hit points, and skill rolls. At creation, each character gets to spend a number of points (based mainly on age, Education, and Intelligence) on skills determined by interests or career choice. Each of the three playable races has specific tables for the creation of characters. Character Skills are based on percentages. To succeed in a skill, the player must roll under the relevant skill with modifiers on percentile dice. Another critique of the game system has been the large effect of character age on skills, usually considered the most important character attributes. In Niven's future world, the deterioration of age has been largely reversed, so humans live hundreds of years. Therefore, a 200-year-old character will have vastly more skill points than a 20-year-old, with little compensatory advantages for the younger one. Only two publications were published, the "Ringworld" role-playing game box set itself, and the "Ringworld Companion", both in 1984 by Chaosium. The magazine "Different Worlds", issue 37, featured a "Ringworld" adventure, "Louis Wu & His Motley Crew." The article "The Dolphins of Known Space: A new race for the Ringworld Game" appeared in "Dragon Magazine" issue 95. The "Ringworld" role-playing game box set was titled "Larry Niven's Ringworld: Roleplaying Adventure Beneath the Great Arch", referring to the way the Ringworld looked from its interior surface. The authors are credited as Greg Stafford, John Hewitt, Sherman Kahn, Lynn Willis, Sandy Petersen, Rudy Kraft, Charlie Krank, Ed Gore, and Jeff Okamoto. It came in a box set with four books: the Explorer Book, Technology Book, Gamemasters Book, and Creatures Book, a sheet of cardboard miniatures, reference and character sheets, and a set of dice: 2d20 (actually dice with two sets of digits 0 to 9), 1d8, and 2d6. This book begins with a character sheet. It introduces role-playing games, then covers character creation, skill use, and combat. It presents a detailed history of humanity between the 20th and 29th centuries. It then describes eleven human worlds: Belt (the asteroid belt), Canyon, Down, Gummidgy, Home, Jinx, Margrave, Plateau, Silvereyes, "We Made It" and Wunderland. Finally, it gives rules for non-human, Kzin or Puppeteer, player characters, and a glossary. The Gamemaster Book begins with technical essays on the Ringworld, from physical construction, to life on the ring, with diagrams. There is a section on the "City Builders"—a Ringworld race that dominated the Ringworld, built floating cities, and sent spaceships to explore other worlds, until a mysterious technological virus destroyed their empire. Another section lists unanswered questions about the Ringworld. There are suggestions for creating scenarios and campaigns, and information on technology of various humanoid species of the Ringworld, and additional rules, including gravity, Credit Rating, and psionics. There is also an introductory scenario ("The Journey of the Catseye") intended to begin a Ringworld campaign. The characters are hired by Captain Gregor Lopez, famous explorer, for a journey to the Ringworld that does not go completely as planned. The Technology Book gives rules and descriptions of the equipment employed by the explorers of the 29th century, categorized into generators, computers, medical equipment, tools, vehicles, weapons and defenses. The Creature Book gives rules and descriptions for creatures, divided into Aliens, Humanoids, Animals and Plants. Many races get specialized hit location tables, characteristic maxima and minima, skills and traits. This supplement was published not long after the box set. The authors are credited as Greg Stafford, John Hewitt, Sherman Kahn, Lynn Willis, Sandy Petersen, Rudy Kraft, and Charlie Krank. The book starts with a diagram of the Ringworld and its star, EC-1752, new humanoids, aliens, plants and animals, technological objects, and original errata. There is some information on spaceships (Human and City Builder), hyperspace, a map of Human Space, and statistics for vehicles used on the Ringworld. Then there is a new race, the "Agamans", desert nomads, and a scenario involving them, "The Sand Eaters". Finally, there is a three part scenario named "The Kaladians", about the defense of travelling merchants. Both scenarios can be integrated into the campaign given in the basic set. None of these three additional races appear in any of the "Ringworld" novels. Steve Peterson reviewed "Ringworld" in "Space Gamer" No. 71. Peterson commented that "Niven fans should buy it for the essays and background materials. Role-players should be prepared to do some work on scenarios; but if you do, you'll have some terrific roleplaying in a beautifully detailed world. Science-fiction gamers who want to use it for source material probably won't get their money's worth." Steve Nutt reviewed "Ringworld" for "Imagine" magazine, and stated that "Altogether, Ringworld's advantages and disadvantages stem from its campaign setting. The actual mechanics of the game are top quality, yet background and atmosphere are what make or break a campaign, and in Ringworld this aspect could be somewhat daunting to the uninitiated."
https://en.wikipedia.org/wiki?curid=26459
Rigel Rigel , designated β Orionis (Latinized to Beta Orionis, abbreviated Beta Ori, β Ori), is a blue supergiant star in the constellation of Orion, approximately from Earth. Rigel is the brightest and most massive component—and the eponym—of a star system of at least four stars that appear as a single blue-white point of light to the naked eye. A star of spectral type B8Ia, Rigel is calculated to be anywhere from 61,500 to 363,000 times as luminous as the Sun, and 18 to 24 times as massive, depending on the method and assumptions used. Its radius is over 70 times that of the Sun, and its surface temperature is . Rigel's mass-loss due to its stellar wind is estimated to be 10 million times more than that of the Sun. With an estimated age of 7 to 9 million years, Rigel has exhausted its core hydrogen fuel, expanded and cooled to become a supergiant. It will end its life as a type II supernova. Rigel varies slightly in brightness, its apparent magnitude ranging from 0.05 to 0.18. It is classified as an Alpha Cygni variable due to the amplitude and periodicity of its brightness variation, as well as its spectral type. Its intrinsic variability is caused by pulsations in its unstable atmosphere. Rigel is generally the seventh-brightest star in the night sky and the brightest star in Orion, though it is occasionally outshone by Betelgeuse, which varies over a larger range. A triple-star system is separated from Rigel by . It has an apparent magnitude of 6.7, making it 1/400th as bright as Rigel. Two stars in the system can be resolved by large telescopes, and the brighter of the two is a spectroscopic binary. These three stars are all blue-white main sequence stars, each three to four times as massive as the Sun. Rigel and the triple system orbit a common center of gravity with a period estimated to be 24,000 years. The inner stars of the triple system orbit each other every 10 days, and the outer star orbits the inner pair every 63 years. A much fainter star, separated from Rigel and the others by nearly an arc minute, may be part of the same star system. In 2016, the International Astronomical Union (IAU) included the name "Rigel" in the IAU Catalog of Star Names. According to the IAU, this proper name applies only to the primary component A of the Rigel system. In historical astronomical catalogs, the system is listed variously as H II 33, Σ 668, β 555, or ADS 3823. For simplicity, Rigel's companions are referred to as Rigel B, C, and D; the IAU describes such names as "useful nicknames" that are "unofficial". In modern comprehensive catalogs, the whole multiple star system is known as WDS 05145-0812 or CCDM 05145–0812. The designation of Rigel as β Orionis (Latinized to Beta Orionis) was made by Johann Bayer in 1603. The "beta" designation is commonly given to the second-brightest star in each constellation, but Rigel is almost always brighter than α Orionis (Betelgeuse). Astronomer James B. Kaler has speculated that Rigel was designated by Bayer during a rare period when it was outshone by the variable star Betelgeuse, resulting in the latter star being designated "alpha" and Rigel designated "beta". Bayer did not strictly order the stars by brightness, instead grouping them by magnitude. Rigel and Betelgeuse were both considered to be of the first magnitude class, and in Orion the stars of each class are thought to have been ordered north to south. Rigel is included in the General Catalogue of Variable Stars, but since it already has a Bayer designation it has no separate variable star designation. Rigel has many other stellar designations taken from various catalogs, including the Flamsteed designation 19 Orionis (19 Ori), the Bright Star Catalogue entry HR 1713, and the Henry Draper Catalogue number HD 34085. These designations frequently appear in the scientific literature, but rarely in popular writing. Rigel is an intrinsic variable star with an apparent magnitude ranging from 0.05 to 0.18. It is typically the seventh-brightest star in the celestial sphere, excluding the Sun, although occasionally fainter than Betelgeuse. It is fainter than Capella, which may also vary slightly in brightness. Rigel appears slightly blue-white and has a B-V color index of −0.06. It contrasts strongly with reddish Betelgeuse. Culminating every year at midnight on 12 December, and at 9 pm on 24 January, Rigel is visible on winter evenings in the Northern Hemisphere and on summer evenings in the Southern Hemisphere. In the Southern Hemisphere, Rigel is the first bright star of Orion visible as the constellation rises. The star is a vertex of the "Winter Hexagon", an asterism that includes Aldebaran, Capella, Pollux, Procyon, and Sirius. Rigel is a prominent equatorial navigation star, being easily located and readily visible in all the world's oceans (the exception is the area north of the 82nd parallel north). Rigel's spectral type is a defining point of the classification sequence for supergiants. The overall spectrum is typical for a late B class star, with strong absorption lines of the hydrogen Balmer series as well as neutral helium lines and some of heavier elements such as oxygen, calcium, and magnesium. The luminosity class for B8 stars is estimated from the strength and narrowness of the hydrogen spectral lines, and Rigel is assigned to the bright supergiant class Ia. Variations in the spectrum have resulted in the assignment of different classes to Rigel, such as B8 Ia, B8 Iab, and B8 Iae. As early as 1888, the heliocentric radial velocity of Rigel, as estimated from the Doppler shifts of its spectral lines, was seen to vary. This was confirmed and interpreted at the time as being due to a spectroscopic companion with a period of about 22 days. The radial velocity has since been measured to vary by about around a mean of . In 1933, the Hα line in Rigel's spectrum was seen to be unusually weak and shifted towards shorter wavelengths, while there was a narrow emission spike about to the long wavelength side of the main absorption line. This is now known as a P Cygni profile after a star that shows this feature strongly in its spectrum. It is associated with mass loss where there is simultaneously emission from a dense wind close to the star and absorption from circumstellar material expanding away from the star. The unusual Hα line profile is observed to vary unpredictably. Around a third of the time it is a normal absorption line. About a quarter of the time it is a double-peaked line, that is, an absorption line with an emission core or an emission line with an absorption core. About a quarter of the time it has a P Cygni profile; most of the rest of the time the line has an inverse P Cygni profile, where the emission component is on the short wavelength side of the line. Rarely, there is a pure emission Hα line. The line profile changes are interpreted as variations in the quantity and velocity of material being expelled from the star. Occasional very high-velocity outflows have been inferred, and, more rarely, infalling material. The overall picture is one of large looping structures arising from the photosphere and driven by magnetic fields. Rigel has been known to vary in brightness since at least 1930. The small amplitude of Rigel's brightness variation requires photoelectric or CCD photometry to be reliably detected. This brightness variation has no obvious period. Observations over 18 nights in 1984 showed variations at red, blue, and yellow wavelengths of up to 0.13 magnitudes on timescales of a few hours to several days, but again no clear period. Rigel's color index varies slightly, but this is not significantly correlated with its brightness variations. From analysis of "Hipparcos" satellite photometry, Rigel is identified as belonging to the Alpha Cygni class of variable stars, defined as "non-radially pulsating supergiants of the Bep–AepIa spectral types". In those spectral types, the 'e' indicates that it displays emission lines in its spectrum, while the 'p' means it has an unspecified spectral peculiarity. Alpha Cygni type variables are generally considered to be irregular or have quasi-periods. Rigel was added to the General Catalogue of Variable Stars in the 74th name-list of variable stars on the basis of the Hipparcos photometry, which showed variations with a photographic amplitude of 0.039 magnitudes and a possible period of 2.075 days. Rigel was observed with the Canadian MOST satellite for nearly 28 days in 2009. Milli-magnitude variations were observed, and gradual changes in flux suggest the presence of long-period pulsation modes. From observations of the variable Hα spectral line, Rigel's mass-loss rate due to stellar wind is estimated be solar masses per year (/yr)—around 10 million times more than the mass-loss rate from the Sun. More detailed optical and K band infrared spectroscopic observations, together with VLTI interferometry, were taken from 2006 to 2010. Analysis of the Hα and Hγ line profiles, and measurement of the regions producing the lines, show that Rigel's stellar wind varies greatly in structure and strength. Loop and arm structures were also detected within the wind. Calculations of mass loss from the Hγ line give in 2006-7 and in 2009–10. Calculations using the Hα line give lower results, around . The terminal wind velocity is . It is estimated that Rigel has lost around 3 solar masses () since beginning life as a star of 7 to 9 million years ago. Rigel's distance from the Sun is somewhat uncertain, different estimates being obtained by different methods. The 2007 Hipparcos new reduction of Rigel's parallax is , giving a distance of with a margin of error of about 9%. Rigel B, usually considered to be physically associated with Rigel and at the same distance, has a Gaia Data Release 2 parallax of , suggesting a distance around . However, the measurements for this object may be unreliable. Indirect distance estimation methods have also been employed. For example, Rigel is believed to be in a region of nebulosity, its radiation illuminating several nearby clouds. Most notable of these is the 5°-long IC 2118 (Witch Head Nebula), located at an angular separation of 2.5° from the star, or a projected distance of away. From measures of other nebula-embedded stars, IC 2118's distance is estimated to be . Rigel is an outlying member of the Orion OB1 Association, which is located at a distance of up to from Earth. It is a member of the loosely defined Taurus-Orion R1 Association, somewhat closer at . Rigel is thought to be considerably closer than most of the members of Orion OB1 and the Orion Nebula. Betelgeuse and Saiph lie at a similar distance to Rigel, although Betelgeuse is a runaway star with a complex history and might have originally formed in the main body of the association. Hierarchical scheme for Rigel's components The star system of which Rigel is a part has at least four components. Rigel (sometimes called Rigel A to distinguish from the other components) has a visual companion, which is likely a close triple-star system. A fainter star at a wider separation might be a fifth component of the Rigel system. William Herschel discovered Rigel to be a visual double star on 1 October 1781, cataloguing it as star 33 in the "second class of double stars" in his Catalogue of Double Stars, usually abbreviated to H II 33, or as H 2 33 in the Washington Double Star Catalogue. Friedrich Georg Wilhelm von Struve first measured the relative position of the companion in 1822, cataloguing the visual pair as Σ 668. The secondary star is often referred to as Rigel B or β Orionis B. The angular separation of Rigel B from Rigel A is 9.5 arc seconds to its south along position angle 204°. Although not particularly faint at visual magnitude 6.7, the overall difference in brightness from Rigel A (about 6.6 magnitudes or 440 times fainter) makes it a challenging target for telescope apertures smaller than . At Rigel's estimated distance, Rigel B's projected separation from Rigel A is over 2,200 astronomical units (AU). Since its discovery, there has been no sign of orbital motion, although both stars share a similar common proper motion. The pair would have an estimated orbital period of 24,000 years. Gaia Data Release 2 (DR2) contains a somewhat unreliable parallax for Rigel B, placing it at about , further away than the Hipparcos distance for Rigel, but similar to the Taurus-Orion R1 association. There is no parallax for Rigel in Gaia DR2. The Gaia DR2 proper motions for Rigel B and the Hipparcos proper motions for Rigel are both small, although not quite the same. In 1871, Sherburne Wesley Burnham suspected Rigel B to be a binary system, and in 1878, he resolved it into two components. This visual companion is designated as component C (Rigel C), with a measured separation from component B that varies from less than to around . In 2009, speckle interferometry showed the two almost identical components separated by , with visual magnitudes of 7.5 and 7.6, respectively. Their estimated orbital period is 63 years. Burnham listed the Rigel multiple system as β 555 in his double star catalog or BU 555 in modern use. Component B is a double-lined spectroscopic binary system, which shows two sets of spectral lines combined within its single stellar spectrum. Periodic changes observed in relative positions of these lines indicate an orbital period of 9.86 days. The two spectroscopic components Rigel Ba and Rigel Bb cannot be resolved in optical telescopes but are known to both be hot stars of spectral type around B9. This spectroscopic binary, together with the close visual component Rigel C, is likely a physical triple-star system, although Rigel C cannot be detected in the spectrum, which is inconsistent with its observed brightness. In 1878, Burnham found another possibly associated star of approximately 13th magnitude. He listed it as component D of β 555, although it is unclear whether it is physically related or a coincidental alignment. Its 2017 separation from Rigel was , almost due north at a position angle of 1°. Gaia DR2 finds it to be a 12th magnitude sunlike star at approximately the same distance as Rigel. Likely a K-type main-sequence star, this star would have an orbital period of around 250,000 years, if it is part of the Rigel system. A spectroscopic companion to Rigel was reported on the basis of radial velocity variations, and its orbit was even calculated, but subsequent work suggests that the star does not exist and that observed pulsations are intrinsic to Rigel itself. Rigel is a blue supergiant that has exhausted the hydrogen fuel in its core, expanded and cooled as it moved away from the main sequence across the upper part of the Hertzsprung–Russell diagram. When it was on the main sequence, its effective temperature would have been around . Rigel's complex variability at visual wavelengths is caused by stellar pulsations similar to those of Deneb. Further observations of radial velocity variations indicate that it simultaneously oscillates in at least 19 non-radial modes with periods ranging from about 1.2 to 74 days. Estimation of many physical characteristics of blue supergiant stars, including Rigel, is challenging due to their rarity and uncertainty about how far they are from the Sun. As such, their characteristics are mainly estimated from theoretical stellar evolution models. Its effective temperature can be estimated from the spectral type and color to be around . A mass of at an age of million years has been estimated by comparing evolutionary tracks, while atmospheric modeling from the spectrum gives a mass of . Although Rigel is often considered the most luminous star within 1,000 light-years of the Sun, its energy output is poorly known. Using the Hipparcos distance of , the estimated relative luminosity for Rigel is about 120,000 times that of the Sun (), but another recently published distance of suggests an even higher luminosity of . Other calculations based on theoretical stellar evolutionary models of Rigel's atmosphere give luminosities anywhere between and , while summing the spectral energy distribution from historical photometry with the Hipparcos distance suggests a luminosity as low as . A 2018 study using the Navy Precision Optical Interferometer measured the angular diameter as . After correcting for limb darkening, the angular diameter is found to be , yielding a radius of . An older measurement of the angular diameter gives , equivalent to a radius of at . These radii are calculated assuming the Hipparcos distance of ; adopting a distance of leads to a significantly larger size. Due to their closeness to each other and ambiguity of the spectrum, little is known about the intrinsic properties of the members of the Rigel BC triple system. All three stars seem to be near equally hot B-type main-sequence stars that are 3 to 4 times as massive as the Sun. Stellar evolution models suggest that the pulsations of Rigel are powered by nuclear reactions in a hydrogen-burning shell that is at least partially non-convective. These pulsations are stronger and more numerous in stars that have evolved through a red supergiant phase and then increased in temperature to again become a blue supergiant. This is due to the decreased mass and increased levels of fusion products at the surface of the star. Rigel is likely to be fusing helium in its core. Due to strong convection of helium produced in the core while Rigel was on the main sequence and in the hydrogen-burning shell since it became a supergiant, the fraction of helium at the surface has increased from 26.6% when the star formed to 32% now. The surface abundances of carbon, nitrogen, and oxygen seen in the spectrum are compatible with a post-red supergiant star only if its internal convection zones are modeled using non-homogeneous chemical conditions known as the Ledoux Criteria. Rigel is expected to eventually end its stellar life as a type II supernova. It is one of the closest known potential supernova progenitors to Earth, and would be expected to have a maximum apparent magnitude of around (about the same brightness as a quarter Moon or around 300 times brighter than Venus ever gets.) The earliest known recording of the name "Rigel" is in the "Alfonsine tables" of 1521. It is derived from the Arabic name ', "the left leg (foot) of Jauzah" (i.e. "rijl" meaning "leg, foot"), which can be traced to the 10th century. "Jauzah" was a proper name for Orion; an alternative Arabic name was ', "the foot of the great one", from which stems the rarely used variant names "Algebar" or "Elgebar". The "Alphonsine tables" saw its name split into "Rigel" and "Algebar", with the note, "et dicitur Algebar. Nominatur etiam Rigel." . Alternate spellings from the 17th century include "Regel" by Italian astronomer Giovanni Battista Riccioli, "Riglon" by German astronomer Wilhelm Schickard, and "Rigel Algeuze" or "Algibbar" by English scholar Edmund Chilmead. With the constellation representing the mythological Greek huntsman Orion, Rigel is his knee or (as its name suggests) foot; with the nearby star Beta Eridani marking Orion's footstool. Rigel is presumably the star known as "Aurvandil's toe" in Norse mythology. In the Caribbean, Rigel represented the severed leg of the folkloric figure "Trois Rois", himself represented by the three stars of Orion's Belt. The leg had been severed with a cutlass by the maiden "Bįhi" (Sirius). The Lacandon people of southern Mexico knew it as "tunsel" ("little woodpecker"). Rigel was known as "Yerrerdet-kurrk" to the Wotjobaluk koori of southeastern Australia, and held to be the mother-in-law of "Totyerguil" (Altair). The distance between them signified the taboo preventing a man from approaching his mother-in-law. The indigenous Boorong people of northwestern Victoria named Rigel as "Collowgullouric Warepil". The Wardaman people of northern Australia know Rigel as the Red Kangaroo Leader "Unumburrgu" and chief conductor of ceremonies in a songline when Orion is high in the sky. Eridanus, the river, marks a line of stars in the sky leading to it, and the other stars of Orion are his ceremonial tools and entourage. Betelgeuse is "Ya-jungin" "Owl Eyes Flicking", watching the ceremonies. The Māori people of New Zealand named Rigel as "Puanga", said to be a daughter of "Rehua" (Antares), the chief of all-stars. Its heliacal rising presages the appearance of "Matariki" (the Pleiades) in the dawn sky, marking the Māori New Year in late May or early June. The Moriori people of the Chatham Islands, as well as some Maori groups in New Zealand, mark the start of their New Year with Rigel rather than the Pleiades. "Puaka" is a southern name variant used in the South Island. In Japan, the Minamoto or Genji clan chose Rigel and its white color as its symbol, calling the star "Genji-boshi" (), while the Taira or Heike clan adopted Betelgeuse and its red color. The two powerful families fought the Genpei War; the stars were seen as facing off against each other and only kept apart by the three stars of Orion's Belt. The MS "Rigel" was originally a Norwegian ship, built in Copenhagen in 1924. It was requisitioned by the Germans during World War II and sunk in 1944 while being used to transport prisoners of war. Two US Navy ships have borne the name USS "Rigel". The SSM-N-6 Rigel was a cruise missile program for the US Navy that was cancelled in 1953 before reaching deployment. The Rigel Skerries are a chain of small islands in Antarctica, renamed after originally being called Utskjera. They were given their current name as Rigel was used as an astrofix. Mount Rigel, elevation , is in Antarctica.
https://en.wikipedia.org/wiki?curid=26463
Special relativity In physics, special relativity (also known as the special theory of relativity) is the generally accepted and experimentally confirmed physical theory regarding the relationship between space and time. In Albert Einstein's original treatment, it is based on two postulates: Some of the work of Albert Einstein in special relativity is built on the earlier work by Hendrik Lorentz. Special relativity was originally proposed by Albert Einstein in a paper published on 26 September 1905 titled "On the Electrodynamics of Moving Bodies". The incompatibility of Newtonian mechanics with Maxwell's equations of electromagnetism and, experimentally, the Michelson-Morley null result (and subsequent similar experiments) demonstrated that the historically hypothesized luminiferous aether did not exist. This led to Einstein's development of special relativity, which corrects mechanics to handle situations involving all motions and especially those at a speed close to that of light (known as ""). Today, special relativity is proven to be the most accurate model of motion at any speed when gravitational effects are negligible. Even so, the Newtonian model is still valid as a simple and accurate approximation at low velocities (relative to the speed of light), for example, everyday motions on Earth. Special relativity has a wide range of consequences. These have been experimentally verified, and include length contraction, time dilation, relativistic mass, a universal speed limit, mass–energy equivalence, the speed of causality and relativity of simultaneity. It has, for example, replaced the conventional notion of an absolute universal time with the notion of a time that is dependent on reference frame and spatial position. Rather than an invariant time interval between two events, there is an invariant spacetime interval. Combined with other laws of physics, the two postulates of special relativity predict the equivalence of mass and energy, as expressed in the mass–energy equivalence formula "E" = "mc"2 ("c" is the speed of light in a vacuum). A defining feature of special relativity is the replacement of the Galilean transformations of Newtonian mechanics with the Lorentz transformations. Time and space cannot be defined separately from each other (as was earlier thought to be the case). Rather, space and time are interwoven into a single continuum known as "spacetime". Events that occur at the same time for one observer can occur at different times for another. Until Einstein developed general relativity, introducing a curved spacetime to incorporate gravity, the phrase "special relativity" was not used. A translation sometimes used is "restricted relativity"; "special" really means "special case". The theory is "special" in that it only applies in the special case where the spacetime is "flat", that is, the curvature of spacetime, described by the energy-momentum tensor and causing gravity, is negligible. In order to correctly accommodate gravity, Einstein formulated general relativity in 1915. Special relativity, contrary to some historical descriptions, does accommodate accelerations as well as accelerating frames of reference. Just as Galilean relativity is now accepted to be an approximation of special relativity that is valid for low speeds, special relativity is considered an approximation of general relativity that is valid for weak gravitational fields, that is, at a sufficiently small scale (e.g., for tidal forces) and in conditions of free fall. General relativity, however, incorporates noneuclidean geometry in order to represent gravitational effects as the geometric curvature of spacetime. Special relativity is restricted to the flat spacetime known as Minkowski space. As long as the universe can be modeled as a pseudo-Riemannian manifold, a Lorentz-invariant frame that abides by special relativity can be defined for a sufficiently small neighborhood of each point in this curved spacetime. Galileo Galilei had already postulated that there is no absolute and well-defined state of rest (no privileged reference frames), a principle now called Galileo's principle of relativity. Einstein extended this principle so that it accounted for the constant speed of light, a phenomenon that had been observed in the Michelson–Morley experiment. He also postulated that it holds for all the laws of physics, including both the laws of mechanics and of electrodynamics. Einstein discerned two fundamental propositions that seemed to be the most assured, regardless of the exact validity of the (then) known laws of either mechanics or electrodynamics. These propositions were the constancy of the speed of light in a vacuum and the independence of physical laws (especially the constancy of the speed of light) from the choice of inertial system. In his initial presentation of special relativity in 1905 he expressed these postulates as: The constancy of the speed of light was motivated by Maxwell's theory of electromagnetism and the lack of evidence for the luminiferous ether. There is conflicting evidence on the extent to which Einstein was influenced by the null result of the Michelson–Morley experiment. In any case, the null result of the Michelson–Morley experiment helped the notion of the constancy of the speed of light gain widespread and rapid acceptance. The derivation of special relativity depends not only on these two explicit postulates, but also on several tacit assumptions (made in almost all theories of physics), including the isotropy and homogeneity of space and the independence of measuring rods and clocks from their past history. Following Einstein's original presentation of special relativity in 1905, many different sets of postulates have been proposed in various alternative derivations. However, the most common set of postulates remains those employed by Einstein in his original paper. A more mathematical statement of the Principle of Relativity made later by Einstein, which introduces the concept of simplicity not mentioned above is: Henri Poincaré provided the mathematical framework for relativity theory by proving that Lorentz transformations are a subset of his Poincaré group of symmetry transformations. Einstein later derived these transformations from his axioms. Many of Einstein's papers present derivations of the Lorentz transformation based upon these two principles. Reference frames play a crucial role in relativity theory. The term reference frame as used here is an observational perspective in space which is not undergoing any change in motion (acceleration), from which a position can be measured along 3 spatial axes (so, at rest or constant velocity). In addition, a reference frame has the ability to determine measurements of the time of events using a 'clock' (any reference device with uniform periodicity). An event is an occurrence that can be assigned a single unique moment and location in space relative to a reference frame: it is a "point" in spacetime. Since the speed of light is constant in relativity irrespective of reference frame, pulses of light can be used to unambiguously measure distances and refer back the times that events occurred to the clock, even though light takes time to reach the clock after the event has transpired. For example, the explosion of a firecracker may be considered to be an "event". We can completely specify an event by its four spacetime coordinates: The time of occurrence and its 3-dimensional spatial location define a reference point. Let's call this reference frame "S". In relativity theory, we often want to calculate the coordinates of an event from differing reference frames. The equations that relate measurements made in different frames are called "transformation equations". To gain insight in how the spacetime coordinates measured by observers in different reference frames compare with each other, it is useful to work with a simplified setup with frames in a "standard configuration." With care, this allows simplification of the math with no loss of generality in the conclusions that are reached. In Fig. 2‑1, two Galilean reference frames (i.e., conventional 3-space frames) are displayed in relative motion. Frame S belongs to a first observer O, and frame S′ (pronounced "S prime" or "S dash") belongs to a second observer O′. Since there is no absolute reference frame in relativity theory, a concept of 'moving' doesn't strictly exist, as everything may be moving with respect to some other reference frame. Instead, any two frames that move at the same speed in the same direction are said to be "comoving". Therefore, "S" and "S"′ are not "comoving". The principle of relativity, which states that physical laws have the same form in each inertial reference frame, dates back to Galileo, and was incorporated into Newtonian physics. However, in the late 19th century, the existence of electromagnetic waves led some physicists to suggest that the universe was filled with a substance that they called "aether", which, they postulated, would act as the medium through which these waves, or vibrations, propagated (in many respects similar to the way sound propagates through air). The aether was thought to be an absolute reference frame against which all speeds could be measured, and could be considered fixed and motionless relative to Earth or some other fixed reference point. The aether supposedly possessed some wonderful properties: it was sufficiently elastic to support electromagnetic waves, and those waves could interact with matter, yet it offered no resistance to bodies passing through it (its one property was that it allowed electromagnetic waves to propagate). The results of various experiments, including the Michelson–Morley experiment in 1887 (subsequently verified with more accurate and innovative experiments), led to the theory of special relativity, by showing that the aether did not exist. Einstein's solution was to discard the notion of an aether and the absolute state of rest. In relativity, any reference frame moving with uniform motion will observe the same laws of physics. In particular, the speed of light in vacuum is always measured to be "c", even when measured by multiple systems that are moving at different (but constant) velocities. From the principle of relativity alone without assuming the constancy of the speed of light (i.e., using the isotropy of space and the symmetry implied by the principle of special relativity) it can be shown that the spacetime transformations between inertial frames are either Euclidean, Galilean, or Lorentzian. In the Lorentzian case, one can then obtain relativistic interval conservation and a certain finite limiting speed. Experiments suggest that this speed is the speed of light in vacuum. Einstein consistently based the derivation of Lorentz invariance (the essential core of special relativity) on just the two basic principles of relativity and light-speed invariance. He wrote: Thus many modern treatments of special relativity base it on the single postulate of universal Lorentz covariance, or, equivalently, on the single postulate of Minkowski spacetime. Rather than considering universal Lorentz covariance to be a derived principle, this article considers it to be the fundamental postulate of special relativity. The traditional two-postulate approach to special relativity is presented in innumerable college textbooks and popular presentations. Textbooks starting with the single postulate of Minkowski spacetime include those by Taylor and Wheeler and by Callahan. This is also the approach followed by the Wikipedia articles Spacetime and Minkowski diagram. Define an event to have spacetime coordinates in system "S" and in a reference frame moving at a velocity v with respect to that frame, "S"′. Then the Lorentz transformation specifies that these coordinates are related in the following way: where is the Lorentz factor and "c" is the speed of light in vacuum, and the velocity "v" of "S"′, relative to "S", is parallel to the "x"-axis. For simplicity, the "y" and "z" coordinates are unaffected; only the "x" and "t" coordinates are transformed. These Lorentz transformations form a one-parameter group of linear mappings, that parameter being called rapidity. Solving the four transformation equations above for the unprimed coordinates yields the inverse Lorentz transformation: Enforcing this "inverse" Lorentz transformation to coincide with the Lorentz transformation from the primed to the unprimed system, shows the unprimed frame as moving with the velocity "v′" = −"v", as measured in the primed frame. There is nothing special about the "x"-axis. The transformation can apply to the "y"- or "z"-axis, or indeed in any direction parallel to the motion (which are warped by the "γ" factor) and perpendicular; see the article Lorentz transformation for details. A quantity invariant under Lorentz transformations is known as a Lorentz scalar. Writing the Lorentz transformation and its inverse in terms of coordinate differences, where one event has coordinates and , another event has coordinates and , and the differences are defined as we get If we take differentials instead of taking differences, we get Spacetime diagrams (Minkowski diagrams) are an extremely useful aid to visualizing how coordinates transform between different reference frames. Although it is not as easy to perform exact computations using them as directly invoking the Lorentz transformations, their main power is their ability to provide an intuitive grasp of the results of a relativistic scenario. To draw a spacetime diagram, begin by considering two Galilean reference frames, S and S', in standard configuration, as shown in Fig. 2‑1. Fig. 3‑1a. Draw the formula_14 and formula_15 axes of frame S. The formula_14 axis is horizontal and the formula_15 (actually formula_18) axis is vertical, which is the opposite of the usual convention in kinematics. The formula_18 axis is scaled by a factor of formula_20 so that both axes have common units of length. In the diagram shown, the gridlines are spaced one unit distance apart. The 45° diagonal lines represent the "worldlines" of two photons passing through the origin at time formula_21 The slope of these worldlines is 1 because the photons advance one unit in space per unit of time. Two events, formula_22 and formula_23 have been plotted on this graph so that their coordinates may be compared in the S and S' frames. Fig. 3‑1b. Draw the formula_24 and formula_25 axes of frame S'. The formula_25 axis represents the worldline of the origin of the S' coordinate system as measured in frame S. In this figure, formula_27 Both the formula_25 and formula_24 axes are tilted from the unprimed axes by an angle formula_30 where formula_31 The primed and unprimed axes share a common origin because frames S and S' had been set up in standard configuration, so that formula_32 when formula_33 Fig. 3‑1c. Units in the primed axes have a different scale from units in the unprimed axes. From the Lorentz transformations, we observe that formula_34 coordinates of formula_35 in the primed coordinate system transform to formula_36 in the unprimed coordinate system. Likewise, formula_34 coordinates of formula_38 in the primed coordinate system transform to formula_39 in the unprimed system. Draw gridlines parallel with the formula_25 axis through points formula_41 as measured in the unprimed frame, where formula_42 is an integer. Likewise, draw gridlines parallel with the formula_24 axis through formula_44 as measured in the unprimed frame. Using the Pythagorean theorem, we observe that the spacing between formula_25 units equals formula_46 times the spacing between formula_18 units, as measured in frame S. This ratio is always greater than 1, and ultimately it approaches infinity as formula_48 Fig. 3‑1d. Since the speed of light is an invariant, the "worldlines" of two photons passing through the origin at time formula_49 still plot as 45° diagonal lines. The primed coordinates of formula_22 and formula_51 are related to the unprimed coordinates through the Lorentz transformations and "could" be approximately measured from the graph (assuming that it has been plotted accurately enough), but the real merit of a Minkowski diagram is its granting us a geometric view of the scenario. For example, in this figure, we observe that the two timelike-separated events that had different x-coordinates in the unprimed frame are now at the same position in space. While the unprimed frame is drawn with space and time axes that meet at right angles, the primed frame is drawn with axes that meet at acute or obtuse angles. The frames are actually equivalent. The asymmetry is due to unavoidable distortions in how spacetime coordinates map onto a Cartesian plane. By analogy, planar maps of the world are unavoidably distorted, but with experience and intuition, one learns to mentally account for these distortions. The consequences of special relativity can be derived from the Lorentz transformation equations. These transformations, and hence special relativity, lead to different physical predictions than those of Newtonian mechanics at all relative velocities, and most pronounced when relative velocities become comparable to the speed of light. The speed of light is so much larger than anything most humans encounter that some of the effects predicted by relativity are initially counterintuitive. In Galilean relativity, length (formula_52) and temporal separation between two events (formula_53) are independent invariants, the values of which do not change when observed from different frames of reference. In special relativity, however, the interweaving of spatial and temporal coordinates generates the concept of an invariant interval, denoted as formula_54: The interweaving of space and time revokes the implicitly assumed concepts of absolute simultaneity and synchronization across non-comoving frames. The form of formula_57 being the "difference" of the squared time lapse and the squared spatial distance, demonstrates a fundamental discrepancy between Euclidean and spacetime distances. The invariance of this interval is a property of the "general" Lorentz transform (also called the Poincaré transformation), making it an isometry of spacetime. The general Lorentz transform extends the standard Lorentz transform (which deals with translations without rotation, that is, Lorentz boosts, in the x-direction) with all other translations, reflections, and rotations between any Cartesian inertial frame. In the analysis of simplified scenarios, such as spacetime diagrams, a reduced-dimensionality form of the invariant interval is often employed: Demonstrating that the interval is invariant is straightforward for the reduced-dimensionality case and with frames in standard configuration: The value of formula_54 is hence independent of the frame in which it is measured. In considering the physical significance of formula_54, there are three cases to note: Consider two events happening in two different locations that occur simultaneously in the reference frame of one inertial observer. They may occur non-simultaneously in the reference frame of another inertial observer (lack of absolute simultaneity). From (the forward Lorentz transformation in terms of coordinate differences) It is clear that the two events that are simultaneous in frame "S" (satisfying ), are not necessarily simultaneous in another inertial frame "S"′ (satisfying ). Only if these events are additionally co-local in frame "S" (satisfying ), will they be simultaneous in another frame "S"′. The Sagnac effect can be considered a manifestation of the relativity of simultaneity. Since relativity of simultaneity is a first order effect in formula_70, instruments based on the Sagnac effect for their operation, such as ring laser gyroscopes and fiber optic gyroscopes, are capable of extreme levels of sensitivity. The time lapse between two events is not invariant from one observer to another, but is dependent on the relative speeds of the observers' reference frames (e.g., the twin paradox which concerns a twin who flies off in a spaceship traveling near the speed of light and returns to discover that the non-traveling twin sibling has aged much more, the paradox being that at constant velocity we are unable to discern which twin is non-traveling and which twin travels). Suppose a clock is at rest in the unprimed system "S". The location of the clock on two different ticks is then characterized by . To find the relation between the times between these ticks as measured in both systems, can be used to find: This shows that the time (Δ"t"′) between the two ticks as seen in the frame in which the clock is moving ("S"′), is "longer" than the time (Δ"t") between these ticks as measured in the rest frame of the clock ("S"). Time dilation explains a number of physical phenomena; for example, the lifetime of high speed muons created by the collision of cosmic rays with particles in the Earth's outer atmosphere and moving towards the surface is greater than the lifetime of slowly moving muons, created and decaying in a laboratory. The dimensions (e.g., length) of an object as measured by one observer may be smaller than the results of measurements of the same object made by another observer (e.g., the ladder paradox involves a long ladder traveling near the speed of light and being contained within a smaller garage). Similarly, suppose a measuring rod is at rest and aligned along the "x"-axis in the unprimed system "S". In this system, the length of this rod is written as Δ"x". To measure the length of this rod in the system "S"′, in which the rod is moving, the distances "x"′ to the end points of the rod must be measured simultaneously in that system "S"′. In other words, the measurement is characterized by , which can be combined with to find the relation between the lengths Δ"x" and Δ"x"′: This shows that the length (Δ"x"′) of the rod as measured in the frame in which it is moving ("S"′), is "shorter" than its length (Δ"x") in its own rest frame ("S"). Time dilation and length contraction are not merely appearances. Time dilation is explicitly related to our way of measuring "time intervals" between events which occur at the same place in a given coordinate system (called "co-local" events). These time intervals (which can be, and are, actually measured experimentally by relevant observers) are "different" in another coordinate system moving with respect to the first, unless the events, in addition to being co-local, are also simultaneous. Similarly, length contraction relates to our measured distances between separated but simultaneous events in a given coordinate system of choice. If these events are not co-local, but are separated by distance (space), they will "not" occur at the same "spatial distance" from each other when seen from another moving coordinate system. Consider two frames "S" and "S′ " in standard configuration. A particle in "S" moves in the x direction with velocity vector formula_97 What is its velocity formula_98 in frame "S′ "? We can write Substituting expressions for formula_101 and formula_102 from into followed by straightforward mathematical manipulations and back-substitution from yields the Lorentz transformation of the speed formula_103 to formula_104: The inverse relation is obtained by interchanging the primed and unprimed symbols and replacing formula_107 with formula_108 For formula_110 not aligned along the x-axis, we write: The forward and inverse transformations for this case are: We note the following points: There is nothing special about the "x" direction in the standard configuration. The above formalism applies to any direction; and three orthogonal directions allow dealing with all directions in space by decomposing the velocity vectors to their components in these directions. See Velocity-addition formula for details. The composition of two non-collinear Lorentz boosts (i.e., two non-collinear Lorentz transformations, neither of which involve rotation) results in a Lorentz transformation that is not a pure boost but is the composition of a boost and a rotation. Thomas rotation results from the relativity of simultaneity. In Fig. 4‑2a, a rod of length formula_125 in its rest frame (i.e., having a proper length of formula_125) rises vertically along the y‑axis in the ground frame. In Fig. 4‑2b, the same rod is observed from the frame of a rocket moving at speed formula_70 to the right. If we imagine two clocks situated at the left and right ends of the rod that are synchronized "in the frame of the rod," relativity of simultaneity causes the observer in the rocket frame to observe (not "see") the clock at the right end of the rod as being advanced in time by formula_128 and the rod is correspondingly observed as tilted. Unlike second-order relativistic effects such as length contraction or time dilation, this effect becomes quite significant even at fairly low velocities. For example, this can be seen in the spin of moving particles, where Thomas precession is a relativistic correction that applies to the spin of an elementary particle or the rotation of a macroscopic gyroscope, relating the angular velocity of the spin of a particle following a curvilinear orbit to the angular velocity of the orbital motion. Thomas rotation provides the resolution to the well-known "meter stick and hole paradox". In Fig. 4‑3, the time interval between the events A (the "cause") and B (the "effect") is 'time-like'; that is, there is a frame of reference in which events A and B occur at the "same location in space", separated only by occurring at different times. If A precedes B in that frame, then A precedes B in all frames accessible by a Lorentz transformation. It is possible for matter (or information) to travel (below light speed) from the location of A, starting at the time of A, to the location of B, arriving at the time of B, so there can be a causal relationship (with A the cause and B the effect). The interval AC in the diagram is 'space-like'; that is, there is a frame of reference in which events A and C occur simultaneously, separated only in space. There are also frames in which A precedes C (as shown) and frames in which C precedes A. However, there are no frames accessible by a Lorentz transformation, in which events A and C occur at the same location. If it were possible for a cause-and-effect relationship to exist between events A and C, then paradoxes of causality would result. For example, if signals could be sent faster than light, then signals could be sent into the sender's past (observer B in the diagrams). A variety of causal paradoxes could then be constructed. Consider the spacetime diagrams in Fig. 4‑4. A and B stand alongside a railroad track, when a high speed train passes by, with C riding in the last car of the train and D riding in the leading car. The world lines of A and B are vertical ("ct"), distinguishing the stationary position of these observers on the ground, while the world lines of C and D are tilted forwards ("ct′"), reflecting the rapid motion of the observers C and D stationary in their train, as observed from the ground. It is not necessary for signals to be instantaneous to violate causality. Even if the signal from D to C were slightly shallower than the formula_24 axis (and the signal from A to B slightly steeper than the formula_14 axis), it would still be possible for B to receive his message before he had sent it. By increasing the speed of the train to near light speeds, the formula_25 and formula_24 axes can be squeezed very close to the dashed line representing the speed of light. With this modified setup, it can be demonstrated that even signals only "slightly" faster than the speed of light will result in causality violation. Therefore, if causality is to be preserved, one of the consequences of special relativity is that no information signal or material object can travel faster than light in vacuum. This is not to say that "all" faster than light speeds are impossible. Various trivial situations can be described where some "things" (not actual matter or energy) move faster than light. For example, the location where the beam of a search light hits the bottom of a cloud can move faster than light when the search light is turned rapidly (although this does not violate causality or any other relativistic phenomenon). In 1850, Hippolyte Fizeau and Léon Foucault independently established that light travels more slowly in water than in air, thus validating a prediction of Fresnel's wave theory of light and invalidating the corresponding prediction of Newton's corpuscular theory. The speed of light was measured in still water. What would be the speed of light in flowing water? In 1851, Fizeau conducted an experiment to answer this question, a simplified representation of which is illustrated in Fig. 5‑1. A beam of light is divided by a beam splitter, and the split beams are passed in opposite directions through a tube of flowing water. They are recombined to form interference fringes, indicating a difference in optical path length, that an observer can view. The experiment demonstrated that dragging of the light by the flowing water caused displacement of the fringes, showing that the motion of the water had affected the speed of the light. According to the theories prevailing at the time, light traveling through a moving medium would be a simple sum of its speed "through" the medium plus the speed "of" the medium. Contrary to expectation, Fizeau found that although light appeared to be dragged by the water, the magnitude of the dragging was much lower than expected. If formula_135 is the speed of light in still water, and formula_70 is the speed of the water, and formula_137 is the water-bourne speed of light in the lab frame with the flow of water adding to or subtracting from the speed of light, then Fizeau's results, although consistent with Fresnel's earlier hypothesis of partial aether dragging, were extremely disconcerting to physicists of the time. Among other things, the presence of an index of refraction term meant that, since formula_139 depends on wavelength, "the aether must be capable of sustaining different motions at the same time." A variety of theoretical explanations were proposed to explain Fresnel's dragging coefficient that were completely at odds with each other. Even before the Michelson–Morley experiment, Fizeau's experimental results were among a number of observations that created a critical situation in explaining the optics of moving bodies. From the point of view of special relativity, Fizeau's result is nothing but an approximation to , the relativistic formula for composition of velocities. Because of the finite speed of light, if the relative motions of a source and receiver include a transverse component, then the direction from which light arrives at the receiver will be displaced from the geometric position in space of the source relative to the receiver. The classical calculation of the displacement takes two forms and makes different predictions depending on whether the receiver, the source, or both are in motion with respect to the medium. (1) If the receiver is in motion, the displacement would be the consequence of the aberration of light. The incident angle of the beam relative to the receiver would be calculable from the vector sum of the receiver's motions and the velocity of the incident light. (2) If the source is in motion, the displacement would be the consequence of light-time correction. The displacement of the apparent position of the source from its geometric position would be the result of the source's motion during the time that its light takes to reach the receiver. The classical explanation failed experimental test. Since the aberration angle depends on the relationship between the velocity of the receiver and the speed of the incident light, passage of the incident light through a refractive medium should change the aberration angle. In 1810, Arago used this expected phenomenon in a failed attempt to measure the speed of light, and in 1870, George Airy tested the hypothesis using a water-filled telescope, finding that, against expectation, the measured aberration was identical to the aberration measured with an air-filled telescope. A "cumbrous" attempt to explain these results used the hypothesis of partial aether-drag, but was incompatible with the results of the Michelson–Morley experiment, which apparently demanded "complete" aether-drag. Assuming inertial frames, the relativistic expression for the aberration of light is applicable to both the receiver moving and source moving cases. A variety of trigonometrically equivalent formulas have been published. Expressed in terms of the variables in Fig. 5‑2, these include The classical Doppler effect depends on whether the source, receiver, or both are in motion with respect to the medium. The relativistic Doppler effect is independent of any medium. Nevertheless, relativistic Doppler shift for the longitudinal case, with source and receiver moving directly towards or away from each other, can be derived as if it were the classical phenomenon, but modified by the addition of a time dilation term, and that is the treatment described here. Assume the receiver and the source are moving "away" from each other with a relative speed formula_147 as measured by an observer on the receiver or the source (The sign convention adopted here is that formula_147 is "negative" if the receiver and the source are moving "towards" each other). Assume that the source is stationary in the medium. Then where formula_150 is the speed of sound. For light, and with the receiver moving at relativistic speeds, clocks on the receiver are time dilated relative to clocks at the source. The receiver will measure the received frequency to be where An identical expression for relativistic Doppler shift is obtained when performing the analysis in the reference frame of the "receiver" with a moving source. The transverse Doppler effect is one of the main novel predictions of the special theory of relativity. Classically, one might expect that if source and receiver are moving transversely with respect to each other with no longitudinal component to their relative motions, that there should be no Doppler shift in the light arriving at the receiver. Special relativity predicts otherwise. Fig. 5‑3 illustrates two common variants of this scenario. Both variants can be analyzed using simple time dilation arguments. In Fig. 5‑3a, the receiver observes light from the source as being blueshifted by a factor of formula_155. In Fig. 5‑3b, the light is redshifted by the same factor. Time dilation and length contraction are not optical illusions, but genuine effects. Measurements of these effects are not an artifact of Doppler shift, nor are they the result of neglecting to take into account the time it takes light to travel from an event to an observer. Scientists make a fundamental distinction between "measurement" or "observation" on the one hand, versus "visual appearance", or what one "sees". The measured shape of an object is a hypothetical snapshot of all of the object's points as they exist at a single moment in time. The visual appearance of an object, however, is affected by the varying lengths of time that light takes to travel from different points on the object to one's eye. For many years, the distinction between the two had not been generally appreciated, and it had generally been thought that a length contracted object passing by an observer would in fact actually be "seen" as length contracted. In 1959, James Terrell and Roger Penrose independently pointed out that differential time lag effects in signals reaching the observer from the different parts of a moving object result in a fast moving object's visual appearance being quite different from its measured shape. For example, a receding object would "appear" contracted, an approaching object would "appear" elongated, and a passing object would have a skew appearance that has been likened to a rotation. A sphere in motion retains the appearance of a sphere, although images on the surface of the sphere will appear distorted. Fig. 5‑4 illustrates a cube viewed from a distance of four times the length of its sides. At high speeds, the sides of the cube that are perpendicular to the direction of motion appear hyperbolic in shape. The cube is actually not rotated. Rather, light from the rear of the cube takes longer to reach one's eyes compared with light from the front, during which time the cube has moved to the right. This illusion has come to be known as "Terrell rotation" or the "Terrell–Penrose effect". Another example where visual appearance is at odds with measurement comes from the observation of apparent superluminal motion in various radio galaxies, BL Lac objects, quasars, and other astronomical objects that eject relativistic-speed jets of matter at narrow angles with respect to the viewer. An apparent optical illusion results giving the appearance of faster than light travel. In Fig. 5‑5, galaxy M87 streams out a high-speed jet of subatomic particles almost directly towards us, but Penrose–Terrell rotation causes the jet to appear to be moving laterally in the same manner that the appearance of the cube in Fig. 5‑4 has been stretched out. Section "Consequences derived from the Lorentz transformation" dealt strictly with kinematics, the study of the motion of points, bodies, and systems of bodies without considering the forces that caused the motion. This section discusses masses, forces, energy and so forth, and as such requires consideration of physical effects beyond those encompassed by the Lorentz transformation itself. As an object's speed approaches the speed of light from an observer's point of view, its relativistic mass increases thereby making it more and more difficult to accelerate it from within the observer's frame of reference. The energy content of an object at rest with mass "m" equals "mc"2. Conservation of energy implies that, in any reaction, a decrease of the sum of the masses of particles must be accompanied by an increase in kinetic energies of the particles after the reaction. Similarly, the mass of an object can be increased by taking in kinetic energies. In addition to the papers referenced above—which give derivations of the Lorentz transformation and describe the foundations of special relativity—Einstein also wrote at least four papers giving heuristic arguments for the equivalence (and transmutability) of mass and energy, for . Mass–energy equivalence is a consequence of special relativity. The energy and momentum, which are separate in Newtonian mechanics, form a four-vector in relativity, and this relates the time component (the energy) to the space components (the momentum) in a non-trivial way. For an object at rest, the energy–momentum four-vector is : it has a time component which is the energy, and three space components which are zero. By changing frames with a Lorentz transformation in the x direction with a small value of the velocity v, the energy momentum four-vector becomes . The momentum is equal to the energy multiplied by the velocity divided by "c"2. As such, the Newtonian mass of an object, which is the ratio of the momentum to the velocity for slow velocities, is equal to "E"/"c"2. The energy and momentum are properties of matter and radiation, and it is impossible to deduce that they form a four-vector just from the two basic postulates of special relativity by themselves, because these don't talk about matter or radiation, they only talk about space and time. The derivation therefore requires some additional physical reasoning. In his 1905 paper, Einstein used the additional principles that Newtonian mechanics should hold for slow velocities, so that there is one energy scalar and one three-vector momentum at slow velocities, and that the conservation law for energy and momentum is exactly true in relativity. Furthermore, he assumed that the energy of light is transformed by the same Doppler-shift factor as its frequency, which he had previously shown to be true based on Maxwell's equations. The first of Einstein's papers on this subject was "Does the Inertia of a Body Depend upon its Energy Content?" in 1905. Although Einstein's argument in this paper is nearly universally accepted by physicists as correct, even self-evident, many authors over the years have suggested that it is wrong. Other authors suggest that the argument was merely inconclusive because it relied on some implicit assumptions. Einstein acknowledged the controversy over his derivation in his 1907 survey paper on special relativity. There he notes that it is problematic to rely on Maxwell's equations for the heuristic mass–energy argument. The argument in his 1905 paper can be carried out with the emission of any massless particles, but the Maxwell equations are implicitly used to make it obvious that the emission of light in particular can be achieved only by doing work. To emit electromagnetic waves, all you have to do is shake a charged particle, and this is clearly doing work, so that the emission is of energy. Since one can not travel faster than light, one might conclude that a human can never travel farther from Earth than 40 light years if the traveler is active between the ages of 20 and 60. One would easily think that a traveler would never be able to reach more than the very few solar systems which exist within the limit of 20–40 light years from the earth. But that would be a mistaken conclusion. Because of time dilation, a hypothetical spaceship can travel thousands of light years during the pilot's 40 active years. If a spaceship could be built that accelerates at a constant 1"g", it will, after a little less than a year, be travelling at almost the speed of light as seen from Earth. This is described by: where "v"("t") is the velocity at a time "t", "a" is the acceleration of 1"g" and "t" is the time as measured by people on Earth. Therefore, after one year of accelerating at 9.81 m/s2, the spaceship will be travelling at "v" = 0.77"c" relative to Earth. Time dilation will increase the travellers life span as seen from the reference frame of the Earth to 2.7 years, but his lifespan measured by a clock travelling with him will not change. During his journey, people on Earth will experience more time than he does. A 5-year round trip for him will take 6.5 Earth years and cover a distance of over 6 light-years. A 20-year round trip for him (5 years accelerating, 5 decelerating, twice each) will land him back on Earth having travelled for 335 Earth years and a distance of 331 light years. A full 40-year trip at 1"g" will appear on Earth to last 58,000 years and cover a distance of 55,000 light years. A 40-year trip at 1.1"g" will take 148,000 Earth years and cover about 140,000 light years. A one-way 28 year (14 years accelerating, 14 decelerating as measured with the astronaut's clock) trip at 1"g" acceleration could reach 2,000,000 light-years to the Andromeda Galaxy. This same time dilation is why a muon travelling close to "c" is observed to travel much farther than "c" times its half-life (when at rest). Theoretical investigation in classical electromagnetism led to the discovery of wave propagation. Equations generalizing the electromagnetic effects found that finite propagation speed of the E and B fields required certain behaviors on charged particles. The general study of moving charges forms the Liénard–Wiechert potential, which is a step towards special relativity. The Lorentz transformation of the electric field of a moving charge into a non-moving observer's reference frame results in the appearance of a mathematical term commonly called the magnetic field. Conversely, the "magnetic" field generated by a moving charge disappears and becomes a purely "electrostatic" field in a comoving frame of reference. Maxwell's equations are thus simply an empirical fit to special relativistic effects in a classical model of the Universe. As electric and magnetic fields are reference frame dependent and thus intertwined, one speaks of "electromagnetic" fields. Special relativity provides the transformation rules for how an electromagnetic field in one inertial frame appears in another inertial frame. Maxwell's equations in the 3D form are already consistent with the physical content of special relativity, although they are easier to manipulate in a manifestly covariant form, that is, in the language of tensor calculus. "Special" relativity can be combined with quantum mechanics to form relativistic quantum mechanics and quantum electrodynamics. It is an unsolved problem in physics how "general" relativity and quantum mechanics can be unified; quantum gravity and a "theory of everything", which require a unification including general relativity too, are active and ongoing areas in theoretical research. The early Bohr–Sommerfeld atomic model explained the fine structure of alkali metal atoms using both special relativity and the preliminary knowledge on quantum mechanics of the time. In 1928, Paul Dirac constructed an influential relativistic wave equation, now known as the Dirac equation in his honour, that is fully compatible both with special relativity and with the final version of quantum theory existing after 1926. This equation not only describe the intrinsic angular momentum of the electrons called "spin", it also led to the prediction of the antiparticle of the electron (the positron), and fine structure could only be fully explained with special relativity. It was the first foundation of "relativistic quantum mechanics". On the other hand, the existence of antiparticles leads to the conclusion that relativistic quantum mechanics is not enough for a more accurate and complete theory of particle interactions. Instead, a theory of particles interpreted as quantized fields, called "quantum field theory", becomes necessary; in which particles can be created and destroyed throughout space and time. Special relativity in its Minkowski spacetime is accurate only when the absolute value of the gravitational potential is much less than "c"2 in the region of interest. In a strong gravitational field, one must use general relativity. General relativity becomes special relativity at the limit of a weak field. At very small scales, such as at the Planck length and below, quantum effects must be taken into consideration resulting in quantum gravity. However, at macroscopic scales and in the absence of strong gravitational fields, special relativity is experimentally tested to extremely high degree of accuracy (10−20) and thus accepted by the physics community. Experimental results which appear to contradict it are not reproducible and are thus widely believed to be due to experimental errors. Special relativity is mathematically self-consistent, and it is an organic part of all modern physical theories, most notably quantum field theory, string theory, and general relativity (in the limiting case of negligible gravitational fields). Newtonian mechanics mathematically follows from special relativity at small velocities (compared to the speed of light) – thus Newtonian mechanics can be considered as a special relativity of slow moving bodies. See classical mechanics for a more detailed discussion. Several experiments predating Einstein's 1905 paper are now interpreted as evidence for relativity. Of these it is known Einstein was aware of the Fizeau experiment before 1905, and historians have concluded that Einstein was at least aware of the Michelson–Morley experiment as early as 1899 despite claims he made in his later years that it played no role in his development of the theory. Particle accelerators routinely accelerate and measure the properties of particles moving at near the speed of light, where their behavior is completely consistent with relativity theory and inconsistent with the earlier Newtonian mechanics. These machines would simply not work if they were not engineered according to relativistic principles. In addition, a considerable number of modern experiments have been conducted to test special relativity. Some examples: Special relativity uses a 'flat' 4-dimensional Minkowski space – an example of a spacetime. Minkowski spacetime appears to be very similar to the standard 3-dimensional Euclidean space, but there is a crucial difference with respect to time. In 3D space, the differential of distance (line element) "ds" is defined by where are the differentials of the three spatial dimensions. In Minkowski geometry, there is an extra dimension with coordinate "X"0 derived from time, such that the distance differential fulfills where are the differentials of the four spacetime dimensions. This suggests a deep theoretical insight: special relativity is simply a rotational symmetry of our spacetime, analogous to the rotational symmetry of Euclidean space (see Fig. 10‑1). Just as Euclidean space uses a Euclidean metric, so spacetime uses a Minkowski metric. Basically, special relativity can be stated as the "invariance of any spacetime interval" (that is the 4D distance between any two events) when viewed from "any inertial reference frame". All equations and effects of special relativity can be derived from this rotational symmetry (the Poincaré group) of Minkowski spacetime. The actual form of "ds" above depends on the metric and on the choices for the "X"0 coordinate. To make the time coordinate look like the space coordinates, it can be treated as imaginary: (this is called a Wick rotation). According to Misner, Thorne and Wheeler (1971, §2.3), ultimately the deeper understanding of both special and general relativity will come from the study of the Minkowski metric (described below) and to take , rather than a "disguised" Euclidean metric using "ict" as the time coordinate. Some authors use , with factors of "c" elsewhere to compensate; for instance, spatial coordinates are divided by "c" or factors of "c"±2 are included in the metric tensor. These numerous conventions can be superseded by using natural units where . Then space and time have equivalent units, and no factors of "c" appear anywhere. If we reduce the spatial dimensions to 2, so that we can represent the physics in a 3D space we see that the null geodesics lie along a dual-cone (see Fig. 10‑2) defined by the equation; or simply  which is the equation of a circle of radius "c dt". If we extend this to three spatial dimensions, the null geodesics are the 4-dimensional cone: so As illustrated in Fig. 10‑3, the null geodesics can be visualized as a set of continuous concentric spheres with radii = "c dt". This null dual-cone represents the "line of sight" of a point in space. That is, when we look at the stars and say "The light from that star which I am receiving is X years old", we are looking down this line of sight: a null geodesic. We are looking at an event a distance formula_164 away and a time "d/c" in the past. For this reason the null dual cone is also known as the 'light cone'. (The point in the lower left of the Fig. 10‑2 represents the star, the origin represents the observer, and the line represents the null geodesic "line of sight".) The cone in the −"t" region is the information that the point is 'receiving', while the cone in the +"t" section is the information that the point is 'sending'. The geometry of Minkowski space can be depicted using Minkowski diagrams, which are useful also in understanding many of the thought-experiments in special relativity. Note that, in 4d spacetime, the concept of the center of mass becomes more complicated, see center of mass (relativistic). Above, the Lorentz transformation for the time coordinate and three space coordinates illustrates that they are intertwined. This is true more generally: certain pairs of "timelike" and "spacelike" quantities naturally combine on equal footing under the same Lorentz transformation. The Lorentz transformation in standard configuration above, that is, for a boost in the "x" direction, can be recast into matrix form as follows: In Newtonian mechanics, quantities which have magnitude and direction are mathematically described as 3d vectors in Euclidean space, and in general they are parametrized by time. In special relativity, this notion is extended by adding the appropriate timelike quantity to a spacelike vector quantity, and we have 4d vectors, or "four vectors", in Minkowski spacetime. The components of vectors are written using tensor index notation, as this has numerous advantages. The notation makes it clear the equations are manifestly covariant under the Poincaré group, thus bypassing the tedious calculations to check this fact. In constructing such equations, we often find that equations previously thought to be unrelated are, in fact, closely connected being part of the same tensor equation. Recognizing other physical quantities as tensors simplifies their transformation laws. Throughout, upper indices (superscripts) are contravariant indices rather than exponents except when they indicate a square (this should be clear from the context), and lower indices (subscripts) are covariant indices. For simplicity and consistency with the earlier equations, Cartesian coordinates will be used. The simplest example of a four-vector is the position of an event in spacetime, which constitutes a timelike component "ct" and spacelike component , in a contravariant position four vector with components: where we define so that the time coordinate has the same dimension of distance as the other spatial dimensions; so that space and time are treated equally. Now the transformation of the contravariant components of the position 4-vector can be compactly written as: where there is an implied summation on formula_168 from 0 to 3, and formula_169 is a matrix. More generally, all contravariant components of a four-vector formula_170 transform from one frame to another frame by a Lorentz transformation: Examples of other 4-vectors include the four-velocity formula_172 defined as the derivative of the position 4-vector with respect to proper time: where the Lorentz factor is: The relativistic energy formula_175 and relativistic momentum formula_176 of an object are respectively the timelike and spacelike components of a contravariant four momentum vector: where "m" is the invariant mass. The four-acceleration is the proper time derivative of 4-velocity: The transformation rules for "three"-dimensional velocities and accelerations are very awkward; even above in standard configuration the velocity equations are quite complicated owing to their non-linearity. On the other hand, the transformation of "four"-velocity and "four"-acceleration are simpler by means of the Lorentz transformation matrix. The four-gradient of a scalar field φ transforms covariantly rather than contravariantly: which is the transpose of: only in Cartesian coordinates. It's the covariant derivative which transforms in manifest covariance, in Cartesian coordinates this happens to reduce to the partial derivatives, but not in other coordinates. More generally, the "co"variant components of a 4-vector transform according to the "inverse" Lorentz transformation: where formula_182 is the reciprocal matrix of formula_169. The postulates of special relativity constrain the exact form the Lorentz transformation matrices take. More generally, most physical quantities are best described as (components of) tensors. So to transform from one frame to another, we use the well-known tensor transformation law where formula_185 is the reciprocal matrix of formula_186. All tensors transform by this rule. An example of a four dimensional second order antisymmetric tensor is the relativistic angular momentum, which has six components: three are the classical angular momentum, and the other three are related to the boost of the center of mass of the system. The derivative of the relativistic angular momentum with respect to proper time is the relativistic torque, also second order antisymmetric tensor. The electromagnetic field tensor is another second order antisymmetric tensor field, with six components: three for the electric field and another three for the magnetic field. There is also the stress–energy tensor for the electromagnetic field, namely the electromagnetic stress–energy tensor. The metric tensor allows one to define the inner product of two vectors, which in turn allows one to assign a magnitude to the vector. Given the four-dimensional nature of spacetime the Minkowski metric "η" has components (valid in any inertial reference frame) which can be arranged in a matrix: which is equal to its reciprocal, formula_188, in those frames. Throughout we use the signs as above, different authors use different conventions – see Minkowski metric alternative signs. The Poincaré group is the most general group of transformations which preserves the Minkowski metric: and this is the physical symmetry underlying special relativity. The metric can be used for raising and lowering indices on vectors and tensors. Invariants can be constructed using the metric, the inner product of a 4-vector "T" with another 4-vector "S" is: Invariant means that it takes the same value in all inertial frames, because it is a scalar (0 rank tensor), and so no Λ appears in its trivial transformation. The magnitude of the 4-vector "T" is the positive square root of the inner product with itself: One can extend this idea to tensors of higher order, for a second order tensor we can form the invariants: similarly for higher order tensors. Invariant expressions, particularly inner products of 4-vectors with themselves, provide equations that are useful for calculations, because one doesn't need to perform Lorentz transformations to determine the invariants. The coordinate differentials transform also contravariantly: so the squared length of the differential of the position four-vector "dXμ" constructed using is an invariant. Notice that when the line element "dX2 is negative that is the differential of proper time, while when "dX2 is positive, is differential of the proper distance. The 4-velocity "U"μ has an invariant form: which means all velocity four-vectors have a magnitude of "c". This is an expression of the fact that there is no such thing as being at coordinate rest in relativity: at the least, you are always moving forward through time. Differentiating the above equation by "τ" produces: So in special relativity, the acceleration four-vector and the velocity four-vector are orthogonal. The invariant magnitude of the momentum 4-vector generates the energy–momentum relation: We can work out what this invariant is by first arguing that, since it is a scalar, it doesn't matter in which reference frame we calculate it, and then by transforming to a frame where the total momentum is zero. We see that the rest energy is an independent invariant. A rest energy can be calculated even for particles and systems in motion, by translating to a frame in which momentum is zero. The rest energy is related to the mass according to the celebrated equation discussed above: The mass of systems measured in their center of momentum frame (where total momentum is zero) is given by the total energy of the system in this frame. It may not be equal to the sum of individual system masses measured in other frames. To use Newton's third law of motion, both forces must be defined as the rate of change of momentum with respect to the same time coordinate. That is, it requires the 3D force defined above. Unfortunately, there is no tensor in 4D which contains the components of the 3D force vector among its components. If a particle is not traveling at "c", one can transform the 3D force from the particle's co-moving reference frame into the observer's reference frame. This yields a 4-vector called the four-force. It is the rate of change of the above energy momentum four-vector with respect to proper time. The covariant version of the four-force is: In the rest frame of the object, the time component of the four force is zero unless the "invariant mass" of the object is changing (this requires a non-closed system in which energy/mass is being directly added or removed from the object) in which case it is the negative of that rate of change of mass, times "c". In general, though, the components of the four force are not equal to the components of the three-force, because the three force is defined by the rate of change of momentum with respect to coordinate time, that is, "dp"/"dt" while the four force is defined by the rate of change of momentum with respect to proper time, that is, "dp"/"d"τ. In a continuous medium, the 3D "density of force" combines with the "density of power" to form a covariant 4-vector. The spatial part is the result of dividing the force on a small cell (in 3-space) by the volume of that cell. The time component is −1/"c" times the power transferred to that cell divided by the volume of the cell. This will be used below in the section on electromagnetism.
https://en.wikipedia.org/wiki?curid=26962
Saab AB Saab AB (originally , later SAAB and Saab Group) is a Swedish aerospace and defence company, founded in 1937. From 1947 to 1990 it was the parent company of automobile manufacturer Saab Automobile. Between 1968 and 1995 the company was in a merger with commercial vehicle manufacturer Scania-Vabis, known as Saab-Scania. The two were de-merged in 1995 by the new owners, Investor AB. Despite the demerger, both Saab and Scania share the right to use the griffin logo, which originates from the coat of arms of the Swedish region of Scania. "Svenska Aeroplan AB (aktiebolag)" (Swedish for "Swedish Aeroplane Company Limited") (SAAB) was founded in 1937 in Trollhättan, with the merger of Svenska Aero AB (SAAB) and Linköping based ASJA the headquarters moved to Linköping. The style "Saab" replaced "SAAB" around 1950. The basic initial development was the problems for the Swedish government to get quality military aircraft delivered at the beginning of the Second World War. The final trigger was the inability to get a large number of Seversky P-35 delivered from the United States. From then on the Swedish government focused on establishing domestic production and development of military aircraft and Saab took that challenge. A policy that has continued to this day. Originally manufacturing aircraft, the company sought ways in which to diversify its business. Before the Second World War most cars in Sweden were imported from the United States. The US car manufacturers have been producing tanks during the war and the US domestic market took all the US car production in the late 1940s. So there was a huge supply need of private cars in Europe and Sweden. Buyers were facing waiting lists of years on new cars. In the late 1940s Saab began manufacturing cars at its Saab Automobile division, based in Trollhättan. The first car was the Saab 92; full-scale production started 12 December 1949, based on the prototype Ursaab. In the late 1950s Saab ventured into the computer market with Datasaab. The company was a result partly of the need to make a computer that would be small enough to mount in an aeroplane as navigational equipment. During the 1960s several computers were developed and sold to European countries, for uses such as banking. The aircraft computer (CK 37) was used in 1971 in the Viggen. The company was sold in 1975 to Sperry UNIVAC, while Saab retained its flight computer development. In May 1965, the company name was changed to Saab AB to reflect its broad range of activities. In 1968 Saab AB merged with the Swedish lorry, bus and heavy-duty diesel engine manufacturer Scania-Vabis, and became Saab-Scania AB. In 1990 General Motors bought 51% of the car division Saab Automobile, and acquired the rest a decade later. In 1991 Investor AB completed a leveraged buyout of Saab-Scania AB. Investor AB acquired all the outstanding shares in Saab-Scania for approximately SEK 21 billion. Saab-Scania became a wholly owned subsidiary of Investor AB and the company was de-listed. In 1995 Saab-Scania was divided by Investor AB into two independent companies, de-merging into Scania AB and Saab AB. The intention by Investor AB was to broaden ownership in the two companies later. Following the sale of 50% of the car division Saab Automobile AB to General Motors, the main reason behind the merger with lorry manufacturer Scania-Vabis in 1968 had disappeared. Saab Military Aircraft and British Aerospace (now BAE Systems) formed in 1995 the joint venture company Saab-BAe Gripen AB, to manufacture, market and support Gripen internationally. This co-operation was extended in 2001 with the formation of Gripen International for the same purpose. From 1998 until 2005 the largest shareholder in Saab was the British aerospace company BAE Systems, following its acquisition of a 35% stake from Investor AB by its predecessor, British Aerospace. In January 2005, BAE Systems reduced its shareholding to 20%. Investor AB maintained a 20% share. 16 November 1999, Saab announced its intention is to purchase Celsius AB and the acquisition was concluded by early March 2000. In September 2000 United Defense Industries (UDI) purchased Bofors Weapon Systems from Saab (the autocannon and tube artillery interests), while Saab retained the missile interests. Later BAE Systems acquired United Defense Industries. In December 2005 Saab joined the Dassault nEUROn project as a major partner. In October 2008 the company announced its intention to merge its operations with that of Simrad Optronics. The new unit will develop high-tech optronics products and will be headquartered in Norway, although other details of the new arrangement have not been finalized. In 2010 the company restructured from fifteen business units into five business areas; Aeronautics, Dynamics, Electronic Defence Systems, Security and Defence Solutions, and Support and Services. According to Saab the restructuring was undertaken to become more market and customer oriented. In March 2010, BAE Systems sold half of its 20% stake in the company to Investor AB, which then became the major shareholder. In June 2011, the British company eventually sold its remaining stake bringing its 16-year involvement in Saab to an end. As of June 2012, Investor AB owns a 30% stake in the company (39.5% of the voting rights) and is the majority owner. The main focus of aircraft production is fighter aircraft. Saab has been making aircraft since the 1930s, and the jet predecessors of the JAS 39 Gripen were the Tunnan, the Lansen, the Draken and the Viggen. The last civilian models made by Saab were the Saab 340 and Saab 2000. Both were mid-range turboprop-powered airliners. The development and the manufacturing of these aircraft is undertaken in Linköping. In May 2019, Saab announced plans to locate a new U.S. manufacturing operation in Discovery Park District Aerospace on the west side of the Purdue University campus. The facility will do the final assembly of the T-X advanced jet trainer, which is a plane developed by Boeing and Saab for the United States Air Force. Aeronautics offers airborne systems, related subsystems, Unmanned Aerial Systems (UAS) and aerostructures. The business area Aeronautics is responsible for airframe structures for JAS 39 Gripen, and whole sections for Airbus, Boeing and NH90; & system development of the JAS 39 Gripen and the Skeldar VTOL UAV. Aeronautics is also partner in the European joint UAV-project Dassault nEUROn, where Saab develop Avionics and is responsible for the overall architecture and design. Marketing and support of the JAS 39 Gripen fighter jet is also included in the Aeronautics business area. Dynamics offers ground combat weapons, missile systems, torpedoes, sensor systems, unmanned underwater vehicles and signature management systems, remotely operated vehicles for armed forces as well as civil security applications. Short range weapons offered include Carl-Gustaf, AT4/AT4 CS, STRIX and MBT LAW. Missile systems offered are RBS 70, RBS 23 and RBS 15. Surveillance offers airborne surveillance solutions (including GlobalEye, Saab 2000 Erieye) AEW&C and fighter radar, ground-based and naval radar (including the Giraffe radar range), electronic warfare (including IDAS and ESTL) and combat systems and C4I solutions. Industrial Products and Services was established on 1 January 2015 and comprises the business units Combitech, Avionics Systems, Aerostructures, Traffic Management, Vricon as well as the development of product ideas that fall outside of Saab’s core business. The business units within Industrial Products and Services differ from Saab’s other operations by their focus on business-to-business (B2B) customers or because they are not dependent on Saab’s principal end-customers. Other business areas within Saab have a customer base largely consisting of public authorities. With different customer groups come different management strategies and priorities. Opportunities to strengthen these operations in the long term are greater in the new organisation. Industrial Products and Services will work with individual growth strategies for each business unit. Support and Services offer maintenance, integrated support solutions, field facilities, logistics and regional aircraft maintenance. Saab Aircraft Leasing leases and resells Saab aircraft to airlines. It completed 30 transactions in 2010. The Saab Barracuda LLC facility in Lillington, North Carolina, manufactures signature management products and provides customized services. Foremost among the camouflage, concealment and deception products is the Ultra Lightweight Camouflage Net System (ULCANS) which provides multi-spectral protection against visual, near infrared, thermal infrared and broadband radar detection. ULCANS is fielded with the U.S. Army and other Department of Defense organizations and is available in both woodland and desert versions. Saab Barracuda is one of only two qualified suppliers of ULCANS in North America, and currently has a competed $US1.76 billion contract, along with GMA Cover Corp. Submarine division Kockums acquisition.
https://en.wikipedia.org/wiki?curid=26968
Škoda Auto ŠKODA AUTO (), commonly called Škoda, is a Czech automobile manufacturer founded in 1895 as Laurin & Klement and headquartered in Mladá Boleslav, Czech Republic. In 1925 Laurin & Klement was acquired by the industrial conglomerate Škoda Works, which itself became state owned in 1948. After 1991 it was gradually privatized and in 2000 Škoda became a wholly owned subsidiary of the German Volkswagen Group. Škoda automobiles are sold in over 100 countries and in 2018, total global sales reached 1.25 million units, an increase of 4.4% from the previous year. The operating profit was €1.6 billion in 2017, an increase of 34.6% over the previous year. As of 2017, Škoda's profit margin was the second highest of all VW Group brands after Porsche. The Škoda Works were established as an arms manufacturer in 1859. ŠKODA AUTO (and its predecessor Laurin & Klement) is fifth oldest company producing cars and has an unbroken history alongside Daimler, Opel, Peugeot and Tatra. As with many long-established car manufacturers, the company which would become Škoda Auto started in the early 1890s by manufacturing bicycles. Škoda (then Laurin & Klement) factories were founded in 1896 as a velocipede manufacturer. In , 26-year-old Václav Klement, who was a bookseller in Mladá Boleslav, Kingdom of Bohemia (today's Czech Republic, then part of Austria-Hungary), was unable to obtain spare parts to repair his German bicycle. Klement returned his bicycle to the manufacturers, Seidel and Naumann, with a letter, in Czech, asking them to carry out repairs, only to receive a reply, in German, stating: "If you want us to answer you, we insist that you convey your message in a language we understand." Not satisfied with the reply and realising the business potential, Klement, despite having no technical experience, decided to start a bicycle repair shop, which he and Václav Laurin opened in 1896 in Mladá Boleslav. Before going into partnership with Klement, Laurin was an established bicycle manufacturer in the nearby town of Turnov. In 1898, after moving to their newly built factory, the pair bought a Werner "Motocyclette". Laurin & Klement's first motorcyclette, powered by an engine mounted on the handlebars driving the front wheels, proved dangerous and unreliable – an early accident on it cost Laurin a front tooth. To design a safer machine with its structure around the engine, the pair wrote to German ignition specialist Robert Bosch for advice on a different electromagnetic system. Their new "Slavia" motorcycle made its debut in 1899 and the company became the first motorcycle factory in Central Europe. In 1900, with a company workforce of 32, Slavia exports began and 150 machines were shipped to London for the Hewtson firm. Shortly afterwards, the press credited them as makers of the first motorcycle. By 1905, the firm was manufacturing automobiles, making it the second-oldest car manufacturer in the Czech lands after Tatra. The company, with an area of , had a workforce of 320 and used 170 special machine-tools, power-driven by of steam power. The first model, Voiturette A, was a success and the company was established both within Austria-Hungary and internationally. After World War I the Laurin & Klement company began producing trucks, but in 1924, after running into problems and being affected by a fire on their premises, the company sought a new partner. Meanwhile, ""Akciová společnost, dříve Škodovy závody"" (Limited Company, formerly the Škoda Works), an arms manufacturer and multi-sector concern in Pilsen which had become one of the largest industrial enterprises in Europe and the largest in Czechoslovakia, sought to enlarge its non-arms manufacturing base and acquired Laurin & Klement in 1925. It also started manufacturing cars in cooperation with Hispano-Suiza. Most of the later production took place under Škoda's name. An assembly line was used for production from 1930 onwards. In the same year a formal spin-off of the car manufacture into a new company, "Akciová společnost pro automobilový průmysl" or abbreviated "ASAP", took place. ASAP remained a wholly owned subsidiary of the Škoda Works and continued to sell cars under the Škoda marque. Apart from the factory in Mladá Boleslav it included also the firm's representation, sales offices and services, as well as a central workshop in Prague. At the time, the car factory in Mladá Boleslav covered an area of 215,000 square metres and employed 3,750 blue-collar and 500 white-collar workers. After a decline caused by the economic depression, Škoda introduced a new line of cars in the 1930s which significantly differed from its previous products. A new design of chassis with backbone tube and all-around independent suspension was developed under the leadership of chief engineer Vladimír Matouš and modelled on the one first introduced by Hans Ledwinka in Tatra. First used on model Škoda 420 Standard in 1933, it aimed at solving insufficient torsional stiffness of the ladder frame. The new design of chassis became the basis for models Popular (845–1,089 cc), Rapid (1,165–1,766 cc), Favorit (1,802–2,091 cc) and the Superb (2,492–3,991 cc). While in 1933 Škoda had a 14% share of the Czechoslovak car market and occupied third place behind Praga and Tatra, the new line made it a market leader by 1936, with a 39% share in 1938. During the occupation of Czechoslovakia in World War II the Škoda Works were turned into part of the "Reichswerke Hermann Göring" serving the German war effort by producing components for military terrain vehicles, military planes, other weapon components and cartridge cases. Vehicle output decreased from 7,052 in 1939 to 683 in 1944, of which only 35 were passenger cars. A total of 316 trucks were produced between January and May 1945. The UK and US air forces bombed the Škoda works repeatedly between 1940 and 1945. The final massive air raid took place on 25 April 1945 and resulted in the almost complete destruction of the Škoda armament works and about 1,000 dead or injured. When, by July 1945, the Mladá Boleslav factory had been reconstructed, production of Škoda's first post-World War II car, the 1101 series began. It was essentially an updated version of the pre-World War II Škoda Popular. In the autumn of 1948, Škoda (along with all other large manufacturers) became part of the communist planned economy, which meant it was separated from the parent company, Škoda Works. In spite of unfavourable political conditions and losing contact with technical development in non-communist countries, Škoda retained a good reputation until the 1960s, producing models such as the Škoda 440 Spartak, 445 Octavia, Felicia and Škoda 1000 MB. In late 1959, the Škoda Felicia, a compact four-cylinder convertible coupe, was imported into the United States for model year 1960. Its retail price was around US$2,700, for which one could purchase a nicely-equipped V8 domestic car that was larger, more comfortable, and had more luxury and convenience features (gasoline retailed for less than 30 cents per gallon, so fuel economy was not of primary importance in the United States at that time). Those Felicias that made it to American ownership soon had a number of reliability problems, further damaging the car's reputation. The Felicia was therefore a poor seller in the States and leftover cars ended up being hied off at a fraction of the original retail list. Since that time, Škoda automobiles have not been imported into the U.S. for retail sale. In the late 1980s, Škoda (then named "Automobilové závody, národní podnik" or abbreviated "AZNP") was still manufacturing cars that conceptually dated back to the 1960s. Rear-engined models such as the Škoda 105/120 (Estelle) and Rapid sold steadily and performed well against more modern makes in races such as the RAC Rally in the 1970s and 1980s. They won their class in the RAC rally for 17 years running. They were powered by a , engine. In spite of its dated image and becoming the subject of negative jokes – What do you call a Škoda with a sunroof? A skip! – Škodas remained a common sight on the roads of the United Kingdom and Western Europe throughout the 1970s and 1980s. Sport versions of the Estelle and earlier models were produced, using the name "Rapid". Soft-top versions were also available. The Rapid was once described as the "poor man's Porsche", and had significant sales success in the UK during the 1980s. In 1987 the Favorit was introduced, and was one of a trio of compact front-wheel drive hatchbacks from the three main Eastern Bloc manufacturers around that time, the others being VAZ's Lada Samara and Zastava's Yugo Sana. The Favorit's appearance was the work of the Italian design company Bertone. With some motor technology licensed from western Europe, but still using the Škoda-designed 1289 cc engine, Škoda engineers designed a car comparable to western production. The technological gap was still there, but began closing rapidly. The Favorit was very popular in Czechoslovakia and other Eastern Bloc countries. It also sold well in Western Europe, especially in the UK and Denmark due to its low price and was regarded as solid and reliable. However, it was perceived as having poor value compared with contemporary Western European designs. The Favorit's trim levels were improved and it continued to be sold until the introduction of the Felicia in 1994. The fall of communism with the Velvet Revolution brought great changes to Czechoslovakia and most industries were subject to privatization. In the case of Škoda Automobile, the state authorities brought in a strong foreign partner. Volkswagen was chosen by the Czech government on 9 December 1990, and, as a result, on 28 March 1991 a joint-venture partnership agreement with Volkswagen took place, marked by the transfer of a 30% share to the Volkswagen Group on 16 April 1991. Until 1990, Škoda was still making its outdated range of rear engine small family cars, although it had started production of the Favorit front-wheel drive hatchback in 1987 as an eventual replacement. In the following years, Škoda became the fourth brand of the German group, as the Volkswagen Group raised its equity share first on 19 December 1994, to 60.3%, followed on 11 December 1995, to 70%. In the competition for Škoda, Volkswagen was pitted against French car-maker Renault, which lost out because its strategic plan did not include producing high-value models in the Czech factories; Renault proposed to manufacture the Renault Twingo city car in the Škoda factories. At the time the decision was made, privatization to a major German company was somewhat controversial, since there was still anti-German sentiment lingering in the Czech Republic from WW2 and its aftermath. However, it could be argued that the subsequent fortunes of other Eastern-Bloc automobile manufacturers such as Lada, AutoVAZ, and of Škoda Works itself – once "Škoda Auto's" parent company – suggested that Volkswagen's involvement was not necessarily a result of poor judgement. Backed by Volkswagen Group expertise and investments, the design – both style and engineering – has improved greatly. The 1994 model Felicia was effectively a reskin of the Favorit, but quality and equipment improvements helped, and in the Czech Republic the car was perceived as good value for money and became popular. Sales improved across Europe, including the United Kingdom, where the Felicia was one of the best-ranking cars in customer satisfaction surveys. Volkswagen AG chairman Ferdinand Piëch personally chose Dirk van Braeckel as head of design, and the subsequent "Octavia" and "Fabia" models made their way to the demanding European Union markets. They are built on common Volkswagen Group floorpans. The Fabia, launched at the end of 1999, formed the basis for later versions of the Volkswagen Polo and SEAT Ibiza, while the Octavia, launched in 1996, has shared its floorpan with a host of cars, the most popular of which is the Volkswagen Golf Mk4. The perception of Škoda in Western Europe has completely changed since the takeover by VW, in stark comparison with the reputation of the cars throughout the 1980s described by some as "the laughing stock" of the automotive world. As technical development progressed and attractive new models were marketed, Škoda's image was initially slow to improve. In the UK, a major turnabout was achieved with the ironic "It is a Škoda, honest" campaign, which began in 2000 when the Fabia launched. In a 2003 advertisement on British television, a new employee on the production line is fitting Škoda badges on the car bonnets. When some attractive looking cars come along he stands back, not fitting the badge, since they look so good they "cannot be Škodas". This market campaign worked by confronting Škoda's image problem head-on – a tactic which marketing professionals regarded as high risk. By 2005 Škoda was selling over 30,000 cars a year in the UK, a market share of over 1%. For the first time in its UK history, a waiting list developed for deliveries from Škoda. UK owners have consistently ranked the brand at or near the top of customer satisfaction surveys since the late 1990s. 2010 was one of the most important years for Škoda Auto, in terms of both products and management. On 1 September 2010, Prof. Dr. H.C. Winfried Vahland assumed responsibility for the management of the company, becoming the CEO of Škoda Auto. In the same year, Škoda set forth plans to double the company's annual sales to at least 1.5 million by 2018 (later known as the 'Growth Strategy', ). At the 2010 Paris Motor Show in September 2010, the company unveiled the Octavia Green E Line. This e-car concept was the forerunner to the e-car test fleet that Škoda released in 2012. The final 1st-generation Octavia (Tour) was produced at the Mladá Boleslav plant in November 2010. The worldwide production of this model exceeded 1.4 million units since its release in 1996. In 2010 for the first time in history, China overtook German sales to become Škoda's largest individual market. In 2011, Škoda Auto celebrated its 20-year partnership with the Volkswagen Group. More than 75,000 visitors attended an open-house event held in Mladá Boleslav in the April. Earlier that year, the company provided details on its 2018 Growth Strategy: for at least one new or completely revised model to be released every six months. With this in mind, the company redesigned its logo and CI, which was presented at the 2011 Geneva Motor Show. Škoda's main attraction at the event was the Vision"D" design concept; a forerunner to the future 3rd generation Octavia. Škoda presented the Mission"L" design study at the IAA in Frankfurt am Main in September, which was to become the basis of the company's forthcoming compact model the European Rapid. In the same year, the company started production of the new Rapid model in Pune, India (October 2011), and launched the Škoda Citigo at Volkswagen's Bratislava plant (November 2011). In 2012 Škoda introduced two new mass production models. The European version of the Rapid premiered at the Paris Motor Show. This car was a successor to the 1st-generation Octavia in terms of its price bracket. The second model was the 3rd-generation Octavia, which premiered in December 2012. In the same month, local production of the Yeti was launched at the Nizhny Novgorod GAZ factory. In 2012 Škoda introduced an emission-free fleet of Octavia Green E Line e-cars on Czech roads to be used by external partners. Since internal tests on the fleet in late 2011, the e-fleet had driven more than 250,000 km. During the same year, Škoda celebrated several milestones, including fourteen million Škoda cars being produced since 1905 (January), three million Fabias (May), 500,000 Superbs at the Kvasiny plant (June ) and 5 years of Škoda operations in China. Massive rejuvenation of the model range was a major tune for 2013 at Škoda: The Czech car maker launched the third-generation Octavia Combi and Octavia RS (both liftback and estate) as well as facelifted Superb and Superb Combi. They were accompanied by brand new members of the Rapid family as the Rapid Spaceback, the first Škoda hatchback car in the compact segment, and the Chinese version of the Rapid. The Yeti also faced significant changes. With the facelift, two design variants of Škoda's compact SUV are now available: city-like Yeti and rugged Yeti Outdoor. Chinese customers were also given a Yeti with an extended wheelbase. In 2015, Volkswagen admitted that it had installed pollution-cheating software in many of its cars to fool regulators that its cars met emissions standards when in fact they polluted at much higher-levels than government standards. 1.2 million Škoda cars worldwide were fitted with this emissions-cheating device. Škoda stated that Volkswagen would recall and cover refitting costs for all of the cars affected by the Volkswagen emissions testing scandal. In 2015 Škoda was voted the most dependable car brand in the UK. A corporate strategy was launched in 2015 to produce a range of all-electric cars from 2019. Škoda Auto started to manufacture a large, seven-seat SUV Škoda Kodiaq in 2016, it was introduced at the Paris Motor Show in October 2016 and sales began at early 2017. In the second half of 2017 began sales of the new compact-SUV Škoda Karoq, which officially replaced the Škoda Yeti. The automaker introduced in December 2018 a new small family car Škoda Scala. In February 2019, the company introduced in Geneva new subcompact crossover Škoda Kamiq. In 2015, new Škoda chairman Bernhard Maier stated that the Volkswagen Group "is working on a modular, new electric platform and we are in the team", and that "there is no alternative to electrification." New Škoda corporate "Strategy 2025", which replaces the previous "Strategy 2018", aims to start production of a fully electric vehicle in 2020, and five electric models across different segments by 2025. The all-electric car Škoda Enyaq will be available for sale from the second half of 2020. In 2017 Auto Shanghai, Škoda displayed its Vision E concept for an all-electric 300bhp coupé-SUV, with level 3 autonomy capability and range. It is based on the VW MEB platform and Škoda Auto will also manufacture electric-vehicle batteries for the Volkswagen Group in its facility in the Czech Republic. The second development stage, the Škoda Vision iV was revealed in March 2019. A plug-in hybrid car, Škoda Superb iV is available for sale from early 2020 and a small SUV model Škoda Kamiq with a natural gas-electric hybrid powertrain and a hybrid Fabia from later the same year. By March 2018, the electrification plan was expanded to ten electrified models for 2025: six fully electric cars and four plugin-hybrids. Out of these, five models are to be available by 2020. In 2018, the brand launched its largest-ever investment plan of €2 billion over five years into its electrification. The brand's first fully electric car, a city car Škoda Citigo e iV, is sold from early 2020. Škoda has maintained sound financial stability over recent years. In 2013 the brand achieved sales revenues totalling €10.3 billion (2012: €10.4 billion). Due to the weak economic situation in many European countries and the expansion of the model range, operating profit reached a modest 522 million euros (2012: €712 million). Škoda achieved a successful start to 2014: As well as recording the highest number of deliveries to customers in a first quarter ever (247,200; up 12.1%), it recorded a significant increase in sales revenue (23.7%) to almost 3 billion euros. Operating profit increased 65.2% to 185 million Euros over the previous year. ! 1994 !! 1995 !! 1996 !! 1997 !! 1998 !! 1999 !! 2000 !! 2001 !! 2002 !! 2003 !! 2004 !! 2005 !! 2006 !! 2007 !! 2008 !! 2009 !! 2010 !! 2011 !! 2012 !! 2013 !! 2014 !! 2015 !! 2016 !! 2017 !! 2018 ! 172,000 || 210,000 || 261,000 || 336,334 || 363,500 || 385,330 || 435,403 || 460,252 || 445,525 || 449,758 || 451,675 || 492,111 || 549,667 || 630,032 || 674,530 || 684,226 || 762,600 || 879,200 || 949,412 || 920,800 || 1,037,200 || 1,055,500 || 1,127,700 || 1,200,500 || 1,253,700 As of August 2016, Škoda was being sold in 102 countries. In 2017, the top markets for Škoda by number of sales are China, Germany, Czech Republic, Great Britain, Poland, Russia, France, Turkey, Italy and Austria. In the Asia-Pacific region, Škoda is also being sold in Australia, New Zealand, Taiwan, Brunei and India. Škoda is also expanding into emerging markets such as Iran, where imports are to be started from 2018 and production of vehicles by 2020. Expansion strategy also includes Singapore. Škoda cars are now made in factories in the Czech Republic, China, Russia, India and Slovakia. A smaller number of Škoda models are additionally manufactured in Öskemen, Kazakhstan and Solomonovo, Ukraine through local partners. The following table lists the factories and their production models in 2015. The Škoda brand has been engaged in motor sport since 1901, and has gained a number of titles with various vehicles around the world. The team had competed as a manufacturer in the Intercontinental Rally Challenge (before it merged with ERC in 2013) and World Rally Championship between 1999 and 2005. Now it competes in the European Rally Championship and WRC-2. Until the final season of IRC in 2012, Škoda Motorsport was the most successful manufacturer with a total of 27 points, winning the rallying series in 2010–2012. Since 2013, When the two competing series were merged, it continued to compete in the European Rally Championship. Škoda Motorsport drivers won with the Škoda Fabia S2000 the European Rally Championships in 2012–2014. Following a long history of class victories in lower levels of motorsport, Škoda became a participant in the FIA World Rally Championship in the 1999 season, with World Rally Car models of the Škoda Octavia. Škoda's best result with the Octavia WRC was Armin Schwarz's third place at the 2001 Safari Rally. From mid 2003, the Octavia was replaced by the smaller Škoda Fabia. Škoda used the 2004 season to develop the car further, but did not achieve much success the following season. However, at the season-ending Rally Australia, 1995 world champion Colin McRae was running second before retiring. Škoda then withdrew from the series, and the 2006 season saw Škoda represented by the semi-privateer Red Bull Škoda Team. Jan Kopecký drove the Fabia WRC to fifth place at the Rally Catalunya, and as late as the 2007 Rallye Deutschland the Fabia still achieved a fifth-place result, again in the hands of Kopecký. Former works Ford and Citroen driver François Duval also drove a Fabia WRC in 2006 for the privateer First Motorsport team, achieving a sixth-place finish in Catalunya. In 2009, Škoda entered the Intercontinental Rally Challenge (IRC) for the first time, using the Fabia S2000, winning three rallies and finishing second in both the drivers and manufacturers championship. In 2010, Škoda won a total of seven IRC events winning both the manufacturers and driver championship for Juho Hänninen. These achievements were repeated in the following two seasons, with Andreas Mikkelsen as the drivers' champion. In 2013, the Intercontinental Rally Challenge was merged with the European Rally Championship (ERC) and the team gained the drivers' championship title once again for Jan Kopecký. The car was also raced by privateers in several championships, including Red Bull, Barwa, Rene Georges and Rufa in the 2010 Super 2000 World Rally Championship. Škoda Motorsport won the 2015, 2016 and 2017 WRC-2 championships with Škoda Fabia R5. In August 2011, a special Škoda Octavia vRS set a world record at the Bonneville Speedway and became the fastest production car in the world with an engine up to two litres, when it hit . The current fastest production Škoda car is the Škoda Superb III, with a top speed of and an acceleration from in 5.8 seconds. ŠKODA AUTO was founded in 1895 in Mladá Boleslav, Czech Republic. In 1923 two different trademarks were registered at the Office for Innovation and Model Registration in Plzeň. The first depicted a winged arrow pointing to the right with five feathers in a circle and the second was a winged arrow with three feathers. The famous winged arrow with three feathers still forms the Škoda logo today. The "ŠKODA" text was added to the logo in 1936. The arrow represents speed, the wings progress and freedom, the eye precision and the circle unity, completeness, world and harmony. The story goes that, on his travels through the US, Emil Škoda had once been so taken with a Native American's feathered headdress that he had returned to Pilsen with a relief image which inspired the logo.
https://en.wikipedia.org/wiki?curid=26970
San Francisco Bay San Francisco Bay is a shallow estuary in the US state of California. It is surrounded by a contiguous region known as the San Francisco Bay Area (often simply "the Bay Area"), and is dominated by the large cities of San Jose, San Francisco and Oakland. San Francisco Bay drains water from approximately 40 percent of California. Water from the Sacramento and San Joaquin rivers, and from the Sierra Nevada mountains, flow into Suisun Bay, which then travels through the Carquinez Strait to meet with the Napa River at the entrance to San Pablo Bay, which connects at its south end to San Francisco Bay. It then connects to the Pacific Ocean via the Golden Gate strait. However, this entire group of interconnected bays is often called the "San Francisco Bay". The bay was designated a Ramsar Wetland of International Importance on February 2, 2012. The bay covers somewhere between , depending on which sub-bays (such as San Pablo Bay), estuaries, wetlands, and so on are included in the measurement. The main part of the bay measures wide east-to-west and somewhere between 1 and 2 north-to-south. It is the largest Pacific estuary in the Americas. The bay was navigable as far south as San Jose until the 1850s, when hydraulic mining released massive amounts of sediment from the rivers that settled in those parts of the bay that had little or no current. Later, wetlands and inlets were deliberately filled in, reducing the Bay's size since the mid-19th century by as much as one third. Recently, large areas of wetlands have been restored, further confusing the issue of the Bay's size. Despite its value as a waterway and harbor, many thousands of acres of marshy wetlands at the edges of the bay were, for many years, considered wasted space. As a result, soil excavated for building projects or dredged from channels was often dumped onto the wetlands and other parts of the bay as landfill. From the mid-19th century through the late 20th century, more than a third of the original bay was filled and often built on. The deep, damp soil in these areas is subject to soil liquefaction during earthquakes, and most of the major damage close to the Bay in the Loma Prieta earthquake of 1989 occurred to structures on these areas. The Marina District of San Francisco, hard hit by the 1989 earthquake, was built on fill that had been placed there for the Panama-Pacific International Exposition, although liquefaction did not occur on a large scale. In the 1990s, San Francisco International Airport proposed filling in hundreds more acres to extend its overcrowded international runways in exchange for purchasing other parts of the bay and converting them back to wetlands. The idea was, and remains, controversial. (For further details, see the "Bay Fill and Depth Profile" section.) There are five large islands in San Francisco Bay. Alameda, the largest island, was created when a shipping lane was cut to form the Port of Oakland in 1901. It is now a suburban community. Angel Island was known as "Ellis Island West" because it served as the entry point for immigrants from East Asia. It is now a state park accessible by ferry. Mountainous Yerba Buena Island is pierced by a tunnel linking the east and west spans of the San Francisco–Oakland Bay Bridge. Attached to the north is the artificial and flat Treasure Island, site of the 1939 Golden Gate International Exposition. From the Second World War until the 1990s, both islands served as military bases and are now being redeveloped. Isolated in the center of the Bay is Alcatraz, the site of the famous federal penitentiary. The federal prison on Alcatraz Island no longer functions, but the complex is a popular tourist site. Despite its name, Mare Island in the northern part of the bay is a peninsula rather than an island. San Francisco Bay is thought to represent a down-warping of the Earth's crust between the San Andreas Fault to the west and the Hayward Fault to the east, though the precise nature of this remains under study. About 560,000 years ago, a tectonic shift caused the large inland Lake Corcoran to spill out the central valley and through the Carquinez Strait, carving out sediment and forming canyons in what is now the northern part of the San Francisco Bay and Golden Gate strait. Until the last ice age, the basin which is now filled by the San Francisco Bay was a large linear valley with small hills, similar to most of the valleys of the Coast Ranges. As the great ice sheets began to melt, around 11,000 years ago, the sea level started to rise. By 5000 BC the sea level rose , filling the valley with water from the Pacific. The valley become a bay, and the small hills became islands. The indigenous inhabitants of the San Francisco Bay were Ohlone. The first European to see San Francisco Bay is likely N. de Morena who was left at New Albion at Drakes Bay in Marin County, California by Sir Francis Drake in 1579 and then walked to Mexico. The first recorded European discovery of San Francisco Bay was on November 4, 1769 when Spanish explorer Gaspar de Portolà, unable to find the port of Monterey, California, continued north close to what is now Pacifica and reached the summit of the Sweeney Ridge, now marked as the place where he first sighted San Francisco Bay. Portolá and his party did not realize what they had discovered, thinking they had arrived at a large arm of what is now called Drakes Bay. At the time, Drakes Bay went by the name "Bahia de San Francisco" and thus both bodies of water became associated with the name. Eventually, the larger, more important body of water fully appropriated the name "San Francisco Bay". The first European to enter the bay is believed to have been the Spanish explorer Juan de Ayala, who passed through the Golden Gate on August 5, 1775 in his ship the "San Carlos", and moored in a bay of Angel Island now known as Ayala Cove. Ayala continued to explore the Bay area and the expedition's cartographer, José de Cañizares, gathered the information necessary to produce the first map of the San Francisco Bay area. A number of place names survive (anglicized) from that first map, including Point Reyes, Angel Island, Farallon Islands and Alcatraz Island. The United States seized the region from Mexico during the Mexican–American War (1846–1848). On February 2, 1848, the Mexican province of Alta California was annexed to the U.S. with the signing of the Treaty of Guadalupe Hidalgo. A year and a half later, California requested to join the United States on December 3, 1849 and was accepted as the 31st State of the Union on September 9, 1850. The bay became the center of American settlement and commerce in the Far West through most of the remainder of the 19th century. During the California Gold Rush (1848–1855), San Francisco Bay suddenly became one of the world's great seaports, dominating shipping in the American West until the last years of the 19th century. The bay's regional importance increased further when the First Transcontinental Railroad was connected to its western terminus at Alameda on September 6, 1869. The terminus was switched to the Oakland Long Wharf two months later on November 8, 1869. During the 20th century, the bay was subject to the Reber Plan, which would have filled in parts of the bay in order to increase industrial activity along the waterfront. In 1959, the United States Army Corps of Engineers released a report stating that if current infill trends continued, the bay would be as big as a shipping channel by 2020. This news created the Save the Bay movement in 1960, which mobilized to stop the infill of wetlands and the bay in general, which had shrunk to two-thirds of its size in the century before 1961. San Francisco Bay continues to support some of the densest industrial production and urban settlement in the United States. The San Francisco Bay Area is the American West's second-largest urban area with approximately seven million residents. Despite its urban and industrial character, San Francisco Bay and the Sacramento–San Joaquin River Delta remain perhaps California's most important ecological habitats. California's Dungeness crab, California halibut, and Pacific salmon fisheries rely on the bay as a nursery. The few remaining salt marshes now represent most of California's remaining salt marsh, supporting a number of endangered species and providing key ecosystem services such as filtering pollutants and sediments from the rivers. San Francisco Bay is recognized for protection by the California Bays and Estuaries Policy, with oversight provided by the San Francisco Estuary Partnership. Most famously, the bay is a key link in the Pacific Flyway. Millions of waterfowl annually use the bay shallows as a refuge. Two endangered species of birds are found here: the California least tern and the Ridgway's Rail. Exposed bay muds provide important feeding areas for shorebirds, but underlying layers of bay mud pose geological hazards for structures near many parts of the bay perimeter. San Francisco Bay provided the nation's first wildlife refuge, Oakland's artificial Lake Merritt, constructed in the 1860s, and America's first urban National Wildlife Refuge, the Don Edwards San Francisco Bay National Wildlife Refuge (SFBNWR) in 1972. The Bay is also plagued by non-native species. Salt produced from San Francisco Bay is produced in salt evaporation ponds and is shipped throughout the Western United States to bakeries, canneries, fisheries, cheese makers and other food industries and used to de-ice winter highways, clean kidney dialysis machines, for animal nutrition, and in many industries. Many companies have produced salt in the Bay, with the Leslie Salt Company the largest private land owner in the Bay Area in the 1940s. Low-salinity salt ponds mirror the ecosystem of the bay, with fish and fish-eating birds in abundance. Mid-salinity ponds support dense populations of brine shrimp, which provide a rich food source for millions of shorebirds. Only salt-tolerant micro-algae survive in the high salinity ponds, and impart a deep red color to these ponds from the pigment within the algae protoplasm. The salt marsh harvest mouse is an endangered species endemic to the wetlands of the San Francisco Bay with a high salt tolerance. It needs native pickleweed, which is often displaced by invasive cordgrass, for its habitat. The seasonal range of water temperature in the Bay is from January's to September's when measured at Fort Point, which is near the southern end of the Golden Gate Bridge and at the entrance to San Francisco Bay. For the first time in 65 years, Pacific Harbor Porpoise ("Phocoena phocoena") returned to the Bay in 2009. Golden Gate Cetacean Research, a non-profit organization focused on research on cetaceans, has developed a photo-identification database enabling the scientists to identify specific porpoise individuals and is trying to ascertain whether a healthier bay has brought their return. Pacific harbor porpoise range from Point Conception, California to Alaska and across to the Kamchatka Peninsula and Japan. Recent genetic studies show that there is a local stock from San Francisco to the Russian River and that eastern Pacific coastal populations rarely migrate far, unlike western Atlantic Harbor porpoise. Industrial, mining, and other uses of mercury have resulted in a widespread distribution in the bay, with uptake in the bay's phytoplankton and contamination of its sportfish. In January 1971, two Standard Oil tankers collided in the bay, creating an oil spill disaster, which spurred environmental protection of the bay. In November 2007, a ship named "COSCO Busan" collided with the San Francisco–Oakland Bay Bridge and spilled over of bunker fuel, creating the largest oil spill in the region since 1996. The Bay was once considered a hotspot for polybrominated diphenyl ether (PBDE) flame retardants used to make upholstered furniture and infant care items less flammable. PBDEs have been largely phased out and replaced with alternative phosphate flame retardants. A 2019 San Francisco Estuary Institute (SFEI) study assayed a wide range of these newer flame retardant chemicals in Bay waters, bivalve California mussels ("Mytilus californianus"), and harbor seals ("Phoca vitulina") which haul out in Corkscrew Slough on Bair Island in San Mateo County, with phosphate flame retardant contaminants such as tris(1,3-dichloro-2-propyl)phosphate (TDCPP) and triphenyl phosphate (TPhP) found at levels comparable to thresholds for aquatic toxicity. San Francisco Bay's profile changed dramatically in the late 19th century and again with the initiation of dredging by the US Army Corps of Engineers in the 20th century. Before about 1860, most bay shores (with the exception of rocky shores, such as those in Carquinez Strait; along Marin shoreline; Point Richmond; Golden Gate area) contained extensive wetlands that graded nearly invisibly from freshwater wetlands to salt marsh and then tidal mudflat. A deep channel ran through the center of the bay, following the ancient drowned river valley. In the 1860s and continuing into the early 20th century, miners dumped staggering quantities of mud and gravel from hydraulic mining operations into the upper Sacramento and San Joaquin Rivers. GK Gilbert's estimates of debris total more than eight times the amount of rock and dirt moved during construction of the Panama Canal. This material flowed down the rivers, progressively eroding into finer and finer sediment, until it reached the bay system. Here some of it settled, eventually filling in Suisun Bay, San Pablo Bay, and San Francisco Bay, in decreasing order of severity. By the end of the 19th century, these "slickens" had filled in much of the shallow bay flats, raising the entire bay profile. New marshes were created in some areas. In the decades surrounding 1900, at the behest of local political officials and following Congressional orders, the US Army Corps began dredging the Sacramento and San Joaquin Rivers and the deep channels of San Francisco Bay. This work has continued without interruption ever since an enormous federal subsidy to San Francisco Bay shipping. Some of the dredge spoils were initially dumped in the bay shallows (including helping to create Treasure Island on the former shoals to the north of Yerba Buena Island) and used to raise an island in the Sacramento-San Joaquin Delta. The net effect of dredging has been to maintain a narrow deep channel—deeper perhaps than the original bay channel—through a much shallower bay. At the same time, most of the marsh areas have been filled or blocked off from the bay by dikes. Large ships transiting the bay must follow deep underwater channels that are maintained by frequent dredging as the average depth of the bay is only as deep as a swimming pool—approximately . Between Hayward and San Mateo to San Jose it is . The deepest part of the bay is under and out of the Golden Gate Bridge, at . In the late 1990s, a 12-year harbor-deepening project for the Port of Oakland began; it was largely completed by September 2009. Previously, the bay waters and harbor facilities only allowed for ships with a draft of , but dredging activities undertaken by the United States Army Corps of Engineers in partnership with the Port of Oakland succeeded in providing access for vessels with a draft. Four dredging companies were employed in the US$432 million project, with $244 million paid for with federal funds and $188 million supplied by the Port of Oakland. Some of mud from the dredging was deposited at the western edge of Middle Harbor Shoreline Park to become a shallow-water wetlands habitat for marine and shore life. Further dredging followed in 2011, to maintain the navigation channel. This dredging enabled the arrival of the largest container ship ever to enter the San Francisco Bay, the "MSC Fabiola". Bay pilots trained for the visit on a simulator at the California Maritime Academy for over a year. The ship arrived drawing less than its full draft of because it held only three-quarters of a load after its stop in Long Beach. San Francisco Bay was traversed by watercraft before the arrival of Europeans. Indigenous peoples used canoes to fish and clam along the shoreline. Sailing ships enabled transportation between the Bay and other parts of the world—and served as ferries and freighters within the Bay and between the Bay and inland ports, such as Sacramento and Stockton. These were gradually replaced by steam-powered vessels starting in the late 19th century. Several shipyards were established around the Bay, augmented during wartime. (e.g., the Kaiser Shipyards, Richmond Shipyards) near Richmond in 1940 for World War II for construction of mass-produced, assembly line Liberty and Victory cargo ships. San Francisco Bay is spanned by nine bridges, eight of which carry cars. The Transbay Tube, an underwater rail tunnel, carries BART services between Oakland and San Francisco. Prior to the bridges and, later, the Transbay Tube, transbay transportation was dominated by fleets of ferryboats operated by the Southern Pacific Railroad and the Key System transit company. However, in recent decades, ferries have returned, primarily serving commuters from Marin County, relieving the traffic bottleneck of the Golden Gate Bridge. (See article Ferries of San Francisco Bay). The bay also continues to serve as a major seaport. The Port of Oakland is one of the largest cargo ports in the United States, while the Port of Richmond and the Port of San Francisco provide smaller services. San Francisco Bay is a mecca for sailors (boats, as well as windsurfing and kitesurfing), due to consistent strong westerly/northwesterly thermally-generated winds – Beaufort force 6 () is common on summer afternoons – and protection from large open ocean swells. Yachting and yacht racing are popular pastimes and the San Francisco Bay Area is home to many of the world's top sailors. A shoreline bicycle and pedestrian trail known as the San Francisco Bay Trail encircles the edge of the bay. The San Francisco Bay Area Water Trail, a growing network of launching and landing sites around the Bay for non-motorized small boat users (such as kayakers) is being developed. Parks and protected areas around the bay include Eden Landing Ecological Reserve, Hayward Regional Shoreline, Don Edwards San Francisco Bay National Wildlife Refuge, Hayward Shoreline Interpretive Center, Crown Memorial State Beach, Eastshore State Park, Point Isabel Regional Shoreline, Brooks Island Regional Shoreline, and César Chávez Park. The California Office of Environmental Health Hazard Assessment (OEHHA) has developed a safe eating advisory for fish caught in the San Francisco Bay based on levels of mercury or PCBs found in local species. The San Francisco Bay Area Water Trail is a planned system of designated trailheads designed to improve non-motorized small boat access to the bay. The California Coastal Conservancy approved funding in March 2011 to begin implementation of the water trail.
https://en.wikipedia.org/wiki?curid=26973
San Francisco Peninsula The San Francisco Peninsula is a peninsula in the San Francisco Bay Area that separates San Francisco Bay from the Pacific Ocean. On its northern tip is the City and County of San Francisco. Its southern base is in northern Santa Clara County, including the cities of Palo Alto, Mountain View, and Los Altos. Most of the Peninsula is occupied by San Mateo County, between San Francisco and Santa Clara counties, and including the cities and towns of Atherton, Belmont, Brisbane, Burlingame, Colma, Daly City, East Palo Alto, El Granada, Foster City, Hillsborough, Half Moon Bay, La Honda, Loma Mar, Los Altos, Menlo Park, Millbrae, Mountain View, Pacifica, Palo Alto, Pescadero, Portola Valley, Redwood City, San Bruno, San Carlos, San Mateo, South San Francisco, Sunnyvale, and Woodside. Whereas the term "peninsula" in a geographical sense technically refers to the entire San Francisco Peninsula, in local jargon, "The Peninsula" does not include the city of San Francisco. In 1795, Governor Diego de Borica gave José Darío Argüello a Spanish land grant known as Rancho de las Pulgas. This rancho was the largest grant on the peninsula consisting of . As a local geographic term, the area referred to as "The Peninsula" is distinct from that denoted by "The City", and refers to the portion south of San Francisco. The appellation may date to the period, prior to 1856, when the City of San Francisco and the County of San Francisco were separate entities, the latter then coextensive with contemporary San Mateo County and San Francisco City-County. The City-County owns several disjunct properties along the whole of the Peninsula (mostly water pumping stations connected to the Hetch Hetchy Valley on which San Francisco has a permanent leasehold); thus, most of the larger communities in San Mateo County are "de facto" suburbs of San Francisco, with the neighboring communities of Pacifica, Daly City, Broadmoor, Colma, South San Francisco, Half Moon Bay, San Bruno, and Brisbane being immediate suburbs. The remaining suburban area of the Peninsula is on the east side of the Santa Cruz Mountains, along San Francisco Bay; the west and south-central portions of the Peninsula are mostly rural, unincorporated and unorganised areas. A substantial portion of Silicon Valley is located on the peninsula. In Silicon Valley are the headquarters of some of the largest tech companies in the world, such as Google, Yahoo, Facebook and Apple. Over the last decade or so there has been an influx of immigration into the Bay Area from places like India and China to work in the technology industry. There are well over 6,600 tech startups in the Valley and new ones are created every day. The east side of the peninsula is a densely populated and largely urban and suburban area that includes portions of Silicon Valley. It forms a commuter area between San Francisco to the north and San Jose to the south. A number of major thoroughfares run north-south: El Camino Real (SR 82) and US 101 on the east side along the bay, Interstate 280 down the center, Skyline Boulevard (SR 35) along the crest of the Santa Cruz Mountains, and SR 1 on the west along the Pacific, and SR 85 which forms the southern end of the Peninsula. The Caltrain commuter rail line runs roughly parallel to the El Camino Real (State Route 82) and Highway 101 corridors. The bridges in the Peninsula include the Dumbarton Bridge, the San Francisco - Oakland Bay Bridge, the San Mateo-Hayward Bridge and the Golden Gate Bridge. Along the center line of the Peninsula is the northern half of the Santa Cruz Mountains, formed by the action of plate tectonics along the San Andreas Fault. In the middle of the Peninsula along the fault is the Crystal Springs Reservoir. Just north of the Crystal Springs reservoir is San Andreas Lake, after which the geologic fault was originally named. The San Francisco Peninsula contains a variety of habitats including estuarine, marine, oak woodland, redwood forest, coastal scrub and oak savanna. There are numerous species of wildlife present, especially along the San Francisco Bay estuarine shoreline, San Bruno Mountain, Fitzgerald Marine Reserve and the forests on the Montara Mountain block. The county is home to several endangered species including the San Francisco garter snake, the Mission blue butterfly and the San Bruno elfin butterfly, all of which are endemic to San Mateo County. The endangered California clapper rail is also found on the shores of San Francisco Bay, in the cities of Belmont and San Mateo. A number of noteworthy parks and nature preserves are found on the San Francisco Peninsula, including: There are a number of well-known structures and complexes on the San Francisco Peninsula:
https://en.wikipedia.org/wiki?curid=26974
Silicon Valley Silicon Valley is a region in the southern part of the San Francisco Bay Area in Northern California that serves as a global center for high technology, innovation, venture capital, and social media. It corresponds roughly to the geographical Santa Clara Valley. San Jose is Silicon Valley's largest city, the third-largest in California, and the tenth-largest in the United States; other major Silicon Valley cities include Sunnyvale, Santa Clara, Redwood City, Mountain View, Palo Alto, Menlo Park, and Cupertino. The San Jose Metropolitan Area has the third-highest GDP per capita in the world (after Zurich, Switzerland and Oslo, Norway), according to the Brookings Institution. The word "silicon" in the name originally referred to the large number of innovators and manufacturers in the region specializing in silicon-based MOS transistors and integrated circuit chips. The area is now home to many of the world's largest high-tech corporations, including the headquarters of more than 30 businesses in the Fortune 1000, and thousands of startup companies. Silicon Valley also accounts for one-third of all of the venture capital investment in the United States, which has helped it to become a leading hub and startup ecosystem for high-tech innovation and scientific development. It was in Silicon Valley that the silicon-based integrated circuit, the microprocessor, and the microcomputer, among other technologies, were developed. As of 2013, the region employed about a quarter of a million information technology workers. As more high-tech companies were established across San Jose and the Santa Clara Valley, and then north towards the Bay Area's two other major cities, San Francisco and Oakland, the term "Silicon Valley" has come to have two definitions: a narrower geographic one, referring to Santa Clara County and southeastern San Mateo County, and a metonymical definition referring to high-tech businesses in the entire Bay Area. The term Silicon Valley is often used as a synecdoche for the American high-technology economic sector. The name also became a global synonym for leading high-tech research and enterprises, and thus inspired similar named locations, as well as research parks and technology centers with a comparable structure all around the world. Due to the personal connection between people and computer technology, many headquarters of companies in Silicon Valley are a hotspot for tourism. The popularization of the name is credited to Don Hoefler, who first used it in the article "Silicon Valley USA", appearing in the January 11, 1971 issue of the weekly trade newspaper "Electronic News". The term gained widespread use in the early 1980s, at the time of the introduction of the IBM PC and numerous related hardware and software products to the consumer market. Silicon Valley was born through several contributing factors intersecting, including a skilled science research base housed in area universities, plentiful venture capital, and steady U.S. Department of Defense spending. Stanford University leadership was especially important in the valley's early development. Together these elements formed the basis of its growth and success. On August 23, 1899, the first ship-to-shore wireless telegraph message to be received in the US was from the San Francisco lightship outside the Golden Gate, signaling the return of the American fleet from the Philippines after their victory in the Spanish–American War in 1898. The ship had been outfitted with a wireless telegraph transmitter by a local newspaper, so that they could prepare a celebration on the return of the American sailors. Local historian Clyde Arbuckle states in "Clyde Arbuckle's History of San Jose" that "California first heard the click of a telegraph key on September 11, 1853. It marked completion of an enterprise begun by a couple of San Francisco Merchants' Exchange members named George Sweeney and Theodore E. Baugh…" He says, "In 1849, the gentleman established a wigwag telegraph station a top a high hill overlooking Portsmouth Squares for signaling arriving ships… The operator at the first station caught these signals by telescope and relayed them to the Merchant's Exchange for the waiting business community." Arbuckle points to the historic significance the Merchants Exchange Building (San Francisco) and Telegraph Hill, San Francisco when he goes on to say "The first station gave the name "Telegraph" to the hill on which it was located. It was known as the Inner Station; the second, as the Outer Station. Both used their primitive mode of communication until Messrs. Sweeney and Baugh connected the Outer Station directly with the Merchants's Exchange by electric telegraph Wire." According to Arbuckle (p. 380–381) Sweeney and Baugh's line was strictly an intra-city, San Francisco-based service; that is until California State Telegraph Company enfranchised on May 3, 1852; whereas, O.E. Allen and C. Burnham led the way to "build a line from San Francisco to Marysville via San Jose, Stockton, and Sacramento". Delays to construction occurred until September 1853; but, "…San Jose became the first station on the line when the wire arrived here on October 15. The line was completed when [James] Gamble's northbound crew met a similar crew working southward from Marysville on October 24." The Bay Area had long been a major site of United States Navy research and technology. In 1909, Charles Herrold started the first radio station in the United States with regularly scheduled programming in San Jose. Later that year, Stanford University graduate Cyril Elwell purchased the U.S. patents for Poulsen arc radio transmission technology and founded the Federal Telegraph Corporation (FTC) in Palo Alto. Over the next decade, the FTC created the world's first global radio communication system, and signed a contract with the Navy in 1912. In 1933, Air Base Sunnyvale, California, was commissioned by the United States Government for use as a Naval Air Station (NAS) to house the airship USS Macon in Hangar One. The station was renamed NAS Moffett Field, and between 1933 and 1947, U.S. Navy blimps were based there. A number of technology firms had set up shop in the area around Moffett Field to serve the Navy. When the Navy gave up its airship ambitions and moved most of its west coast operations to San Diego, the National Advisory Committee for Aeronautics (NACA, forerunner of NASA) took over portions of Moffett Field for aeronautics research. Many of the original companies stayed, while new ones moved in. The immediate area was soon filled with aerospace firms, such as Lockheed, which was Silicon Valley's largest employer from the 1950s into 1980s. The Bay Area was an early center of ham radio with about 10% of the operators in the United States. William Eitel, Jack McCullough, and Charles Litton, who together pioneered vacuum tube manufacturing in the Bay Area, were hobbyists with training in technology gained locally who participated in the development of shortwave radio by the ham radio hobby. High frequency, and especially, Very high frequency, VHF, transmission in the 10-meter band, required higher quality power tubes than were manufactured by the consortium of RCA, Western Electric, General Electric, Westinghouse which controlled vacuum tube manufacture. Litton, founder of Litton Industries, pioneered manufacturing techniques which resulted in the award of wartime contracts to manufacture transmitting tubes for radar to Eitel-McCullough, a San Bruno firm, which manufactured power-grid tubes for radio amateurs and aircraft radio equipment. A union organizing drive in 1939–40 at Eitel-McCullough by the strong Bay Area labor movement was fought off by adoption of a strategy of welfare capitalism which included pensions and other generous benefits, profit sharing, and such extras as a medical clinic and a cafeteria. An atmosphere of cooperation and collaboration was established. Successes have been few and far between for union organizing drives by UE and others in subsequent years. On October 4, 1957 the Soviet Union launched the first space satellite, Sputnik, which sparked fear that the Soviet Union was pulling ahead technologically. After President Eisenhower signed the National Aeronautics and Space Act (NASA), he turned to Fairchild Semiconductor, then the only company in the world that was able to make transistors. The president funded Fairchild's project, which was highly successful. Stanford University, its affiliates, and graduates have played a major role in the development of this area. Some examples include the work of Lee De Forest with his invention of a pioneering vacuum tube called the Audion and the oscilloscopes of Hewlett-Packard. A very powerful sense of regional solidarity accompanied the rise of Silicon Valley. From the 1890s, Stanford University's leaders saw its mission as service to the (American) West and shaped the school accordingly. At the same time, the perceived exploitation of the West at the hands of eastern interests fueled booster-like attempts to build self-sufficient local industry. Thus, regionalism helped align Stanford's interests with those of the area's high-tech firms for the first fifty years of Silicon Valley's development. After World War II, Frederick Terman, as Stanford University's dean of the school of engineering, encouraged faculty and graduates to start their own companies. In 1951, Terman spearheaded the creation of Stanford Industrial Park (now Stanford Research Park, an area surrounding Page Mill Road, south west of El Camino Real and extending beyond Foothill Expressway to Arastradero Road), whereby the University leased portions of its land to high-tech firms. He is credited with nurturing companies like Hewlett-Packard, Varian Associates, Eastman Kodak, General Electric, Lockheed Corporation, and other high-tech firms, until what would become Silicon Valley grew up around the Stanford University campus. In 1956, William Shockley, the co-inventor of the first working transistor (with John Bardeen and Walter Houser Brattain), moved from New Jersey to Mountain View, California, to start Shockley Semiconductor Laboratory to live closer to his ailing mother in Palo Alto. Shockley's work served as the basis for many electronic developments for decades. Both Fredrick Terman and William Shockley are often called "the father of Silicon Valley". During 1955–85, solid state technology research and development at Stanford University followed three waves of industrial innovation made possible by support from private corporations, mainly Bell Telephone Laboratories, Shockley Semiconductor, Fairchild Semiconductor, and Xerox PARC. In 1969, the Stanford Research Institute (now SRI International), operated one of the four original nodes that comprised ARPANET, predecessor to the Internet. After World War II, universities were experiencing enormous demand due to returning students. To address the financial demands of Stanford's growth requirements, and to provide local employment opportunities for graduating students, Frederick Terman proposed the leasing of Stanford's lands for use as an office park, named the Stanford Industrial Park (later Stanford Research Park) in the year 1951. Leases were limited to high technology companies. Its first tenant was Varian Associates, founded by Stanford alumni in the 1930s to build military radar components. However, Terman also found venture capital for civilian technology start-ups. One of the major success stories was Hewlett-Packard. Founded in Packard's garage by Stanford graduates Bill Hewlett and David Packard, Hewlett-Packard moved its offices into the Stanford Research Park shortly after 1953. In 1954, Stanford created the Honors Cooperative Program to allow full-time employees of the companies to pursue graduate degrees from the University on a part-time basis. The initial companies signed five-year agreements in which they would pay double the tuition for each student in order to cover the costs. Hewlett-Packard has become the largest personal computer manufacturer in the world, and transformed the home printing market when it released the first thermal drop-on-demand ink jet printer in 1984. Other early tenants included Eastman Kodak, General Electric, and Lockheed. Up until the late 1950s, germanium was the dominant semiconductor material for transistors and other semiconductor devices. Germanium was initially considered the more effective semiconductor material, as it was able to demonstrate better performance due to higher carrier mobility. The relative lack of performance in early silicon semiconductors was due to electrical conductivity being limited by unstable quantum surface states, preventing electricity from reliably penetrating the surface to reach the semiconducting silicon layer. In 1953, William Shockley left Bell Labs in a disagreement over the handling of the invention of the bipolar transistor. After returning to California Institute of Technology for a short while, Shockley moved to Mountain View, California, in 1956, and founded Shockley Semiconductor Laboratory. Unlike many other researchers who used germanium as the semiconductor material, Shockley believed that silicon was the better material for making transistors. Shockley intended to replace the current transistor with a new three-element design (today known as the Shockley diode), but the design was considerably more difficult to build than the "simple" transistor. In 1957, Shockley decided to end research on the silicon transistor. As a result of Shockley's abusive management style, eight engineers left the company to form Fairchild Semiconductor; Shockley referred to them as the "traitorous eight". Two of the original employees of Fairchild Semiconductor, Robert Noyce and Gordon Moore, would go on to found Intel. In 1957, Mohamed Atalla at Bell Labs developed the process of silicon surface passivation by thermal oxidation, which electrically stabilized silicon surfaces and reduced the concentration of electronic states at the surface. This enabled silicon to surpass the conductivity and performance of germanium, leading to silicon replacing germanium as the dominant semiconductor material, and paving the way for the mass-production of silicon semiconductor devices. This led to Atalla inventing the MOSFET (metal-oxide-silicon field-effect transistor), also known as the MOS transistor, with his colleague Dawon Kahng in 1959. It was the first truly compact transistor that could be miniaturised and mass-produced for a wide range of uses, and is credited with starting the silicon revolution. The MOSFET was initially overlooked and ignored by Bell Labs in favour of bipolar transistors, which led to Atalla resigning from Bell Labs and joining Hewlett-Packard in 1961. However, the MOSFET generated significant interest at RCA and Fairchild Semiconductor. In late 1960, Karl Zaininger and Charles Meuller fabricated a MOSFET at RCA, and Chih-Tang Sah built an MOS-controlled tetrode at Fairchild. MOS devices were later commercialized by General Microelectronics and Fairchild in 1964. The development of MOS technology became the focus of startup companies in California, such as Fairchild and Intel, fuelling the technological and economic growth of what would later be called Silicon Valley. On April 23, 1963, J.C.R. Licklider, the first director of the Information Processing Techniques Office (IPTO) at The Pentagon's ARPA issued an office memorandum addressed to Members and Affiliates of the Intergalactic Computer Network. It rescheduled a meeting in Palo Alto regarding his vision of a computer network, which he imagined as an electronic commons open to all, the main and essential medium of informational interaction for governments, institutions, corporations, and individuals. As head of IPTO from 1962 to 1964, "Licklider initiated three of the most important developments in information technology: the creation of computer science departments at several major universities, time-sharing, and networking." By the late 1960s, his promotion of the concept had inspired a primitive version of his vision called ARPANET, which expanded into a network of networks in the 1970s that became the Internet. The Immigration and Nationality Act of 1965 and other factors such as the mass exodus by Vietnamese boat people resulted in significant immigration, particularly by Asians, Latinos, and Portuguese, to Silicon Valley where they contributed to both the high-tech and production workforce. The Asian-American population in Santa Clara County rose from 43,000 in 1970 to 430,000 in 2000. During the same period the Latino population grew to 24% in the county and 30% in San Jose. The African-American population in the county remained steady but grew slightly to about 5%. Expansion of the H-1B visa in 1990 also played a role. Following the 1959 inventions of the monolithic integrated circuit (IC) chip by Robert Noyce at Fairchild, and the MOSFET (MOS transistor) by Mohamed Atalla and Dawon Kahng at Bell Labs, Atalla first proposed the concept of the MOS integrated circuit (MOS IC) chip in 1960, and then the first commercial MOS IC was introduced by General Microelectronics in 1964. The development of the MOS IC led to the invention of the microprocessor, incorporating the functions of a computer's central processing unit (CPU) on a single integrated circuit. The first single-chip microprocessor was the Intel 4004, designed and realized by Federico Faggin along with Ted Hoff, Masatoshi Shima and Stanley Mazor at Intel in 1971. In April 1974, Intel released the Intel 8080, a "computer on a chip", "the first truly usable microprocessor". The Homebrew Computer Club was an informal group of electronic enthusiasts and technically minded hobbyists who gathered to trade parts, circuits, and information pertaining to DIY construction of computing devices. It was started by Gordon French and Fred Moore who met at the Community Computer Center in Menlo Park. They both were interested in maintaining a regular, open forum for people to get together to work on making computers more accessible to everyone. The first meeting was held as of March 1975 at French's garage in Menlo Park, San Mateo County, California; which was on occasion of the arrival of the MITS Altair microcomputer, the first unit sent to the area for review by People's Computer Company. Steve Wozniak and Steve Jobs credit that first meeting with inspiring them to design the original Apple I and (successor) Apple II computers. As a result, the first preview of the Apple I was given at the Homebrew Computer Club. Subsequent meetings were held at an auditorium at the Stanford Linear Accelerator Center. By the early 1970s, there were many semiconductor companies in the area, computer firms using their devices, and programming and service companies serving both. Industrial space was plentiful and housing was still inexpensive. Growth during this era was fueled by the emergence of venture capital on Sand Hill Road, beginning with Kleiner Perkins and Sequoia Capital in 1972; the availability of venture capital exploded after the successful $1.3 billion IPO of Apple Computer in December 1980. Since the 1980s, Silicon Valley has been home to the largest concentration of venture capital firms in the world. In 1971 Don Hoefler traced the origins of Silicon Valley firms, including via investments from Fairchild's eight co-founders. The key investors in Kleiner Perkins and Sequoia Capital were from the same group, directly leading to "Tech Crunch" 2014 estimate of 92 public firms of 130 related listed firms then worth over US$2.1 Trillion with over 2,000 firms traced back to them. Prior to 1970, most Northern California lawyers were based in San Francisco, especially the experienced patent attorneys whom the high-tech industry needed to protect its intellectual property. During the 1970s, lawyers began to follow venture capitalists down the Peninsula to serve the booming high-tech industry in Silicon Valley. One sign of the rapid expansion of Silicon Valley legal services was that Palo Alto law firm Wilson Sonsini Goodrich & Rosati "expanded from a dozen attorneys in 1975 to more than 700 by 2000". During this era, law firms evolved from their "conventional role" as protectors of intellectual property into business advisers, intermediaries, and dealmakers, and thereby acquired "unusual prominence" in Silicon Valley. Although semiconductors are still a major component of the area's economy, Silicon Valley has been most famous in recent years for innovations in software and Internet services. Silicon Valley has significantly influenced computer operating systems, software, and user interfaces. Using money from NASA, the US Air Force, and ARPA, Doug Engelbart invented the mouse and hypertext-based collaboration tools in the mid-1960s and 1970s while at Stanford Research Institute (now SRI International), first publicly demonstrated in 1968 in what is now known as The Mother of All Demos. Engelbart's Augmentation Research Center at SRI was also involved in launching the ARPANET (precursor to the Internet) and starting the Network Information Center (now InterNIC). Xerox hired some of Engelbart's best researchers beginning in the early 1970s. In turn, in the 1970s and 1980s, Xerox's Palo Alto Research Center (PARC) played a pivotal role in object-oriented programming, graphical user interfaces (GUIs), Ethernet, PostScript, and laser printers. While Xerox marketed equipment using its technologies, for the most part its technologies flourished elsewhere. The diaspora of Xerox inventions led directly to 3Com and Adobe Systems, and indirectly to Cisco, Apple Computer, and Microsoft. Apple's Macintosh GUI was largely a result of Steve Jobs' visit to PARC and the subsequent hiring of key personnel. Cisco's impetus stemmed from the need to route a variety of protocols over Stanford University's Ethernet campus network. Commercial use of the Internet became practical and grew slowly throughout the early 1990s. In 1995, commercial use of the Internet grew substantially and the initial wave of internet startups, Amazon.com, eBay, and the predecessor to Craigslist began operations. Silicon Valley is generally considered to have been the center of the dot-com bubble, which started in the mid-1990s and collapsed after the NASDAQ stock market began to decline dramatically in April 2000. During the bubble era, real estate prices reached unprecedented levels. For a brief time, Sand Hill Road was home to the most expensive commercial real estate in the world, and the booming economy resulted in severe traffic congestion. After the dot-com crash, Silicon Valley continues to maintain its status as one of the top research and development centers in the world. A 2006 "The Wall Street Journal" story found that 12 of the 20 most inventive towns in America were in California, and 10 of those were in Silicon Valley. San Jose led the list with 3,867 utility patents filed in 2005, and number two was Sunnyvale, at 1,881 utility patents. Silicon Valley is also home to a significant number of "Unicorn" ventures, referring to startup companies whose valuation has exceeded $1 billion dollars. However, taxes and the cost of living in Silicon Valley have prompted some corporations to gradually transfer their operations to the Midwest or Sun Belt states. The San Francisco Bay Area has the largest concentration of high-tech companies in the United States, at 387,000 high-tech jobs, of which Silicon Valley accounts for 225,300 high-tech jobs. Silicon Valley has the highest concentration of high-tech workers of any metropolitan area, with 285.9 out of every 1,000 private-sector workers. Silicon Valley has the highest average high-tech salary in the United States at $144,800. Largely a result of the high technology sector, the San Jose-Sunnyvale-Santa Clara, CA Metropolitan Statistical Area has the most millionaires and the most billionaires in the United States per capita. The region is the biggest high-tech manufacturing centre in the United States. The unemployment rate of the region was 9.4% in January 2009 and has decreased to a record low of 2.7% as of August 2019. Silicon Valley received 41% of all U.S. venture investment in 2011, and 46% in 2012. More traditional industries also recognize the potential of high-tech development, and several car manufacturers have opened offices in Silicon Valley to capitalize on its entrepreneurial ecosystem. Manufacture of transistors is, or was, the core industry in Silicon Valley. The production workforce was for the most part composed of Asian and Latino immigrants who were paid low wages and worked in hazardous conditions due to the chemicals used in the manufacture of integrated circuits. Technical, engineering, design, and administrative staffs were in large part well compensated. Silicon Valley has a severe housing shortage, caused by the market imbalance between jobs created and housing units built: from 2010 to 2015, many more jobs have been created than housing units built. (400,000 jobs, 60,000 housing units) This shortage has driven home prices extremely high, far out of the range of production workers. As of 2016 a two-bedroom apartment rented for about $2,500 while the median home price was about $1 million. The Financial Post called Silicon Valley the most expensive U.S. housing region. Homelessness is a problem with housing beyond the reach of middle-income residents; there is little shelter space other than in San Jose which, as of 2015, was making an effort to develop shelters by renovating old hotels. "The Economist" also attributes the high cost of living to the success of the industries in this region. Although, this rift between high and low salaries is driving many residents out who can no longer afford to live there. In the Bay Area, the number of residents planning to leave within the next several years has had an increase of 12% since 2016, from 34% to 46%. Thousands of high technology companies are headquartered in Silicon Valley. Among those, the following are in the Fortune 1000: Additional notable companies headquartered (or with a significant presence) in Silicon Valley include (some defunct or subsumed): Depending on what geographic regions are included in the meaning of the term, the population of Silicon Valley is between 3.5 and 4 million. A 1999 study by AnnaLee Saxenian for the Public Policy Institute of California reported that a third of Silicon Valley scientists and engineers were immigrants and that nearly a quarter of Silicon Valley's high-technology firms since 1980 were run by Chinese (17 percent) or Indian descent CEOs (7 percent). There is a stratum of well-compensated technical employees and managers, including 10s of thousands of "single-digit millionaires". This income and range of assets will support a middle-class lifestyle in Silicon Valley. In November 2006, the University of California, Davis released a report analyzing business leadership by women within the state. The report showed that although 103 of the 400 largest public companies headquartered in California were located in Santa Clara County (the most of all counties), only 8.8% of Silicon Valley companies had women CEOs. This was the lowest percentage in the state. (San Francisco County had 19.2% and Marin County had 18.5%.) Silicon Valley tech leadership positions are occupied almost exclusively by men. This is also represented in the number of new companies founded by women as well as the number of women-lead startups that receive venture capital funding. Wadhwa said he believes that a contributing factor is a lack of parental encouragement to study science and engineering. He also cited a lack of women role models and noted that most famous tech leaders—like Bill Gates, Steve Jobs, and Mark Zuckerberg—are men. In 2014, tech companies Google, Yahoo!, Facebook, Apple, and others, released corporate transparency reports that offered detailed employee breakdowns. In May, Google said 17% of its tech employees worldwide were women, and, in the U.S., 1% of its tech workers were black and 2% were Hispanic. June 2014 brought reports from Yahoo! and Facebook. Yahoo! said that 15% of its tech jobs were held by women, 2% of its tech employees were black and 4% Hispanic. Facebook reported that 15% of its tech workforce was female, and 3% was Hispanic and 1% was black. In August, Apple reported that 80% of its global tech staff was male and that, in the U.S., 54% of its tech jobs were staffed by Caucasians and 23% by Asians. Soon after, "USA Today" published an article about Silicon Valley's lack of tech-industry diversity, pointing out that it is largely white or Asian, and male. "Blacks and Hispanics are largely absent," it reported, "and women are underrepresented in Silicon Valley—from giant companies to start-ups to venture capital firms." Civil rights activist Jesse Jackson said of improving diversity in the tech industry, "This is the next step in the civil rights movement" while T.J. Rodgers has argued against Jackson's assertions. As of October 2014, some high-profile Silicon Valley firms were working actively to prepare and recruit women. "Bloomberg" reported that Apple, Facebook, Google, and Microsoft attended the 20th annual Grace Hopper Celebration of Women in Computing conference to actively recruit and potentially hire female engineers and technology experts. The same month, the second annual Platform Summit was held to discuss increasing racial and gender diversity in tech. As of April 2015 experienced women were engaged in creation of venture capital firms which leveraged women's perspectives in funding of startups. After UC Davis published its "Study of California Women Business Leaders" in November 2006, some "San Jose Mercury News" readers dismissed the possibility that sexism contributed in making Silicon Valley's leadership gender gap the highest in the state. A January 2015 issue of "Newsweek" magazine featured an article detailing reports of sexism and misogyny in Silicon Valley. The article's author, Nina Burleigh, asked, "Where were all these offended people when women like Heidi Roizen published accounts of having a venture capitalist stick her hand in his pants under a table while a deal was being discussed?" Silicon Valley firms' board of directors are composed of 15.7% women compared with 20.9% in the S&P 100. The 2012 lawsuit Pao v. Kleiner Perkins was filed in San Francisco County Superior Court by executive Ellen Pao for gender discrimination against her employer, Kleiner Perkins. The case went to trial in February 2015. On March 27, 2015 the jury found in favor of Kleiner Perkins on all counts. Nevertheless, the case, which had wide press coverage, resulted in major advances in consciousness of gender discrimination on the part of venture capital and technology firms and their women employees. Two other cases have been filed against Facebook and Twitter. Funding for public schools in upscale Silicon Valley communities such as Woodside is often supplemented by grants from private foundations set up for that purpose and funded by local residents. Schools in less affluent areas such as East Palo Alto must depend on state funding. The following Santa Clara County cities are traditionally considered to be in Silicon Valley (in alphabetical order): The geographical boundaries of Silicon Valley have changed over the years, traditionally Silicon Valley is known as Santa Clara County, southern San Mateo County and southern Alameda county. However, over the years this geographical area has been expanded to include San Francisco County, Contra Costa County, and the northern parts of Alameda County and San Mateo County, this shift has occurred due to the expansion in the local economy and the development of new technologies. The United States Department of Labor's Quarterly Census of Employment and Wages (QCEW) program defined Silicon Valley as the counties of Alameda, Contra Costa, San Francisco, San Mateo, Santa Clara, and Santa Cruz. In 2015, MIT researchers developed a novel method for measuring which towns are home to startups with higher growth potential and this defines Silicon Valley to center on the municipalities of Menlo Park, Mountain View, Palo Alto, and Sunnyvale. In 1980, "Intelligent Machines Journal" changed its name to "InfoWorld", and, with offices in Palo Alto, began covering the emergence of the microcomputer industry in the valley. Local and national media cover Silicon Valley and its companies. CNN, "The Wall Street Journal", and Bloomberg News operate Silicon Valley bureaus out of Palo Alto. Public broadcaster KQED (TV) and KQED-FM, as well as the Bay Area's local ABC station KGO-TV, operate bureaus in San Jose. KNTV, NBC's local Bay Area affiliate "NBC Bay Area", is located in San Jose. Produced from this location is the nationally distributed TV Show "Tech Now" as well as the CNBC Silicon Valley bureau. San Jose-based media serving Silicon Valley include the "San Jose Mercury News" daily and the "Metro Silicon Valley" weekly. Specialty media include "El Observador" and the "San Jose / Silicon Valley Business Journal". Most of the Bay Area's other major TV stations, newspapers, and media operate in San Francisco or Oakland. Patch.com operates various web portals, providing local news, discussion and events for residents of Silicon Valley. Mountain View has a public nonprofit station, KMVT-15. KMVT-15's shows include Silicon Valley Education News (EdNews)-Edward Tico Producer. Some appearances in media, in order by release date:
https://en.wikipedia.org/wiki?curid=26976
Stanford University Stanford University, officially Leland Stanford Junior University, is a private research university in Stanford, California. The university was founded in 1885 by Leland and Jane Stanford in memory of their only child, Leland Stanford Jr., who had died of typhoid fever at age 15 the previous year. Stanford was a U.S. Senator and former Governor of California who made his fortune as a railroad tycoon. The school admitted its first students on October 1, 1891, as a coeducational and non-denominational institution. Stanford University struggled financially after the death of Leland Stanford in 1893 and again after much of the campus was damaged by the 1906 San Francisco earthquake. Following World War II, Provost Frederick Terman supported faculty and graduates' entrepreneurialism to build self-sufficient local industry in what would later be known as Silicon Valley. The university is also one of the top fundraising institutions in the country, becoming the first school to raise more than a billion dollars in a year. The university is organized around seven schools: three schools consisting of 40 academic departments at the undergraduate level as well as four professional schools that focus on graduate programs in law, medicine, education, and business. All schools are on the same campus. Students compete in 36 varsity sports, and the university is one of two private institutions in the Division I FBS Pac-12 Conference. It has gained 126 NCAA team championships, and Stanford has won the NACDA Directors' Cup for 24 consecutive years, beginning in 1994–1995. In addition, Stanford students and alumni have won 270 Olympic medals including 139 gold medals. As of March 2020, 83 Nobel laureates, 28 Turing Award laureates, and 8 Fields Medalists have been affiliated with Stanford as students, alumni, faculty or staff. In addition, Stanford University is particularly noted for its entrepreneurship and is one of the most successful universities in attracting funding for start-ups. Stanford alumni have founded numerous companies, which combined produce more than $2.7 trillion in annual revenue and have created 5.4 million jobs as of 2011, roughly equivalent to the 7th largest economy in the world (). Stanford is the alma mater of one president of the United States (Herbert Hoover), 30 living billionaires, and 17 astronauts. It is also one of the leading producers of members of the United States Congress. Stanford University was founded in 1885 by Leland and Jane Stanford, dedicated to Leland Stanford Jr, their only child. The institution opened in 1891 on Stanford's previous Palo Alto farm. Despite being impacted by earthquakes in both 1906 and 1989, the campus was rebuilt each time. In 1919, The Hoover Institution on War, Revolution and Peace was started by Herbert Hoover to preserve artifacts related to World War I. The Stanford Medical Center, completed in 1959, is a teaching hospital with over 800 beds. The SLAC National Accelerator Laboratory (originally named the Stanford Linear Accelerator Center), established in 1962, performs research in particle physics. Jane and Leland Stanford modeled their university after the great eastern universities, most specifically Cornell University. Stanford opened being called the "Cornell of the West" in 1891 due to faculty being former Cornell affiliates (either professors, alumni, or both) including its first president, David Starr Jordan. Both Cornell and Stanford were among the first to have higher education be accessible, nonsectarian, and open to women as well as to men. Cornell is credited as one of the first American universities to adopt this radical departure from traditional education, and Stanford became an early adopter as well. Most of Stanford University is on an campus, one of the largest in the United States. It is located on the San Francisco Peninsula, in the northwest part of the Santa Clara Valley (Silicon Valley) approximately southeast of San Francisco and approximately northwest of San Jose. In 2008, 60% of this land remained undeveloped. Stanford's main campus includes a census-designated place within unincorporated Santa Clara County, although some of the university land (such as the Stanford Shopping Center and the Stanford Research Park) is within the city limits of Palo Alto. The campus also includes much land in unincorporated San Mateo County (including the SLAC National Accelerator Laboratory and the Jasper Ridge Biological Preserve), as well as in the city limits of Menlo Park (Stanford Hills neighborhood), Woodside, and Portola Valley. The academic central campus is adjacent to Palo Alto, bounded by El Camino Real, Stanford Avenue, Junipero Serra Boulevard, and Sand Hill Road. The United States Postal Service has assigned it two ZIP Codes: 94305 for campus mail and 94309 for P.O. box mail. It lies within area code 650. Stanford currently operates or intends to operate in various locations outside of its central campus. On the founding grant: Off the founding grant: Locations in development: Many Stanford faculty members live in the "Faculty Ghetto", within walking or biking distance of campus. The Faculty Ghetto is composed of land owned entirely by Stanford. Similar to a condominium, the houses can be bought and sold but the land under the houses is rented on a 99-year lease. Houses in the "Ghetto" appreciate and depreciate, but not as rapidly as overall Silicon Valley values. Some of the land is managed to provide revenue for the university such as the Stanford Shopping Center and the Stanford Research Park. Stanford land is also leased for a token rent by the Palo Alto Unified School District for several schools including Palo Alto High School and Gunn High School. El Camino Park, the oldest Palo Alto city park (established 1914), is also on Stanford land. Contemporary campus landmarks include the Main Quad and Memorial Church, the Cantor Center for Visual Arts and the Bing Concert Hall, the Stanford Mausoleum with the nearby Angel of Grief, Hoover Tower, the Rodin sculpture garden, the Papua New Guinea Sculpture Garden, the Arizona Cactus Garden, the Stanford University Arboretum, Green Library and the Dish. Frank Lloyd Wright's 1937 Hanna–Honeycomb House and the 1919 Lou Henry Hoover House are both listed on the National Historic Register. White Memorial Fountain (also known as "The Claw") between the Stanford Bookstore and the Old Union is a popular place to meet and to engage in the Stanford custom of "fountain hopping"; it was installed in 1964 and designed by Aristides Demetrios after a national competition as a memorial for two brothers in the class of 1949, William N. White and John B. White II, one of whom died before graduating and one shortly after in 1952. Stanford University is a private, non-profit university that is administered as a corporate trust governed by a privately appointed board of trustees with a maximum membership of 38. Trustees serve five-year terms (not more than two consecutive terms) and meet five times annually. A new trustee is chosen by the current trustees by ballot. The Stanford trustees also oversee the Stanford Research Park, the Stanford Shopping Center, the Cantor Center for Visual Arts, Stanford University Medical Center, and many associated medical facilities (including the Lucile Packard Children's Hospital). The Board appoints a President to serve as the chief executive officer of the university and prescribe the duties of professors and course of study, manage financial and business affairs, and appoint nine vice presidents. The Provost is the chief academic and budget officer, to whom the deans of each of the seven schools report. Persis Drell became the 13th Provost in February 2017. As of 2018, the university was organized into seven academic schools. The schools of Humanities and Sciences (27 departments), Engineering (9 departments), and Earth, Energy & Environmental Sciences (4 departments) have both graduate and undergraduate programs while the Schools of Law, Medicine, Education and Business have graduate programs only. The powers and authority of the faculty are vested in the Academic Council, which is made up of tenure and non-tenure line faculty, research faculty, senior fellows in some policy centers and institutes, the president of the university, and some other academic administrators, but most matters are handled by the Faculty Senate, made up of 55 elected representatives of the faculty. The Associated Students of Stanford University (ASSU) is the student government for Stanford University and all registered students are members. Its elected leadership consists of the Undergraduate Senate elected by the undergraduate students, the Graduate Student Council elected by the graduate students, and the President and Vice President elected as a ticket by the entire student body. Stanford is the beneficiary of a special clause in the California Constitution, which explicitly exempts Stanford property from taxation so long as the property is used for educational purposes. The university's endowment, managed by the Stanford Management Company, was valued at $27.7 billion as of August 31, 2019. Payouts from the Stanford endowment covered approximately 21.8% of university expenses in the 2019 fiscal year. In the 2018 NACUBO-TIAA survey of colleges and universities in the United States and Canada, only Harvard University, the University of Texas System, and Yale University had larger endowments than Stanford. In 2006, President John L. Hennessy launched a five-year campaign called the Stanford Challenge, which reached its $4.3 billion fundraising goal in 2009, two years ahead of time, but continued fundraising for the duration of the campaign. It concluded on December 31, 2011, having raised a total of $6.23 billion and breaking the previous campaign fundraising record of $3.88 billion held by Yale. Specifically, the campaign raised $253.7 million for undergraduate financial aid, as well as $2.33 billion for its initiative in "Seeking Solutions" to global problems, $1.61 billion for "Educating Leaders" by improving K-12 education, and $2.11 billion for "Foundation of Excellence" aimed at providing academic support for Stanford students and faculty. Funds supported 366 new fellowships for graduate students, 139 new endowed chairs for faculty, and 38 new or renovated buildings. The new funding also enabled the construction of a facility for stem cell research; a new campus for the business school; an expansion of the law school; a new Engineering Quad; a new art and art history building; an on-campus concert hall; a new art museum; and a planned expansion of the medical school, among other things. In 2012, the university raised $1.035 billion, becoming the first school to raise more than a billion dollars in a year. Stanford follows a quarter system with the Autumn quarter usually starting in late September and the Spring Quarter ending in early June. The full-time, four-year undergraduate program has an arts and sciences focus with high graduate student coexistence. Stanford is accredited by the Western Association of Schools and Colleges. Full-time undergraduate tuition was $42,690 for 2013–2014. Stanford's admission process is need-blind for US citizens and permanent residents; while it is not need-blind for international students, 64% are on need-based aid, with an average aid package of $31,411. In 2012–13, the university awarded $126 million in need-based financial aid to 3,485 students, with an average aid package of $40,460. Eighty percent of students receive some form of financial aid. Stanford has a no-loan policy. For undergraduates admitted in 2015, Stanford waives tuition, room, and board for most families with incomes below $65,000, and most families with incomes below $125,000 are not required to pay tuition; those with incomes up to $150,000 may have tuition significantly reduced. 17% of students receive Pell Grants, a common measure of low-income students at a college. As of 2016 the Office of the Vice Provost and Dean of Research oversaw eighteen independent laboratories, centers, and institutes. Other Stanford-affiliated institutions include the SLAC National Accelerator Laboratory (originally the Stanford Linear Accelerator Center), the Stanford Research Institute (an independent institution which originated at the university), the Hoover Institution on War, Revolution and Peace (a major public policy think tank that attracts visiting scholars from around the world) and the Hasso Plattner Institute of Design (a multidisciplinary design school in cooperation with the Hasso Plattner Institute of University of Potsdam that integrates product design, engineering, and business management education). Stanford is home to the Martin Luther King Jr. Research and Education Institute which grew out of and still contains the Martin Luther King Jr. Papers Project, a collaboration with the King Center to publish the King papers held by the King Center. It also runs the John S. Knight Fellowship for Professional Journalists and the Center for Ocean Solutions, which brings together marine science and policy to address challenges facing the ocean. Together with UC Berkeley and UC San Francisco, Stanford is part of the Biohub, a new medical science research center founded in 2016 by a $600 million commitment from Facebook CEO and founder Mark Zuckerberg and pediatrician Priscilla Chan. As of 2014, Stanford University Libraries (SUL) held a collection of more than 9.3 million volumes, nearly 300,000 rare or special books, 1.5 million e-books, 2.5 million audiovisual materials, 77,000 serials, nearly 6 million microform holdings, and thousands of other digital resources. The main library in the SU library system is Green Library, which also contains various meeting and conference rooms, study spaces, and reading rooms. Lathrop Library (previously Meyer Library, demolished in 2015), holds various student-accessible media resources and houses one of the largest East Asia collections with 540,000 volumes. Stanford is home to the Cantor Center for Visual Arts, a museum with 24 galleries, sculpture gardens, terraces, and a courtyard first established in 1891 by Jane and Leland Stanford as a memorial to their only child. The Center's collection of works by Rodin is among the largest in the world The Thomas Welton Stanford Gallery, which was built in 1917, serves as a teaching resource for the Department of Art & Art History as well as an exhibition venue. In 2014, Stanford opened the Anderson Collection, a new museum focused on postwar American art and founded by the donation of 121 works by food service moguls Mary and Harry Anderson. There are outdoor art installations throughout the campus, primarily sculptures, but some murals as well. The Papua New Guinea Sculpture Garden near Roble Hall features includes wood carvings and "totem poles." The Stanford music department sponsors many ensembles including five choirs, the Stanford Symphony Orchestra, Stanford Taiko, and the Stanford Wind Ensemble. Extracurricular activities include theater groups such as Ram's Head Theatrical Society, the Stanford Improvisors, the Stanford Shakespeare Society, and the Stanford Savoyards, a group dedicated to performing the works of Gilbert and Sullivan. Stanford is also host to ten a cappella groups, including the Mendicants (Stanford's first), Counterpoint (the first all-female group on the West Coast), the Stanford Fleet Street Singers, Harmonics, Mixed Company, Testimony, Talisman, Everyday People, Raagapella and O-Tone. Notably, Stanford ranks high and often first in many domestic college ranking measures, leading "Slate" to dub Stanford in 2014 as "the Harvard of the 21st century," and "The New York Times" in the same year to conclude that "Stanford University has become America's 'it' school, by measures that Harvard once dominated." From polls done by The Princeton Review in 2013, 2014, 2015, 2016, 2017, 2018 and 2019, the most commonly named "dream college" for students was Stanford; separately, parents, too, most frequently named Stanford their "dream college." The inaugural 2017 Wall Street Journal/Times Higher Education College Rankings picked Stanford as the No. 1 school in the United States. Globally, the "Academic Ranking of World Universities (ARWU)" ranked Stanford second in the world most years from 2003 to 2016. In 2019, it ranked 4th among the universities around the world by SCImago Institutions Rankings. Additionally, "Times Higher Education" recognized Stanford as one of the world's "six super brands" on its "World Reputation Rankings", along with Berkeley, Cambridge, Harvard, MIT, and Oxford. Stanford was ranked fifth in the 2016, 2017, and 2018 Nature Index Annual Tables, which measure the largest contributors to papers published in 82 leading journals. Stanford is one of the most successful universities in creating companies and licensing its inventions to existing companies; it is often held up as a model for technology transfer. Stanford's Office of Technology Licensing is responsible for commercializing university research, intellectual property, and university-developed projects. The university is described as having a strong venture culture in which students are encouraged, and often funded, to launch their own companies. Companies founded by Stanford alumni generate more than $2.7 trillion in annual revenue, equivalent to the 10th-largest economy in the world. Some companies closely associated with Stanford and their connections include: Stanford enrolled 6,996 undergraduate and 10,253 graduate students as of the 2019–2020 school year. Women comprised 50.4% of undergraduates and 41.5% of graduate students. In the same academic year, the freshman retention rate was 99%. Stanford awarded 1,819 undergraduate degrees, 2,393 master's degrees, 770 doctoral degrees, and 3270 professional degrees in the 2018–2019 school year. The four-year graduation rate for the class of 2017 cohort was 72.9%, and the six-year rate was 94.4%. The relatively low four-year graduation rate is a function of the university's coterminal degree (or "coterm") program, which allows students to earn a master's degree as a 1-to-2–year extension of their undergraduate program. As of 2010, fifteen percent of undergraduates were first-generation students. As of 2013, 89% of undergraduate students lived in on-campus university housing. First-year undergraduates are required to live on campus, and all undergraduates are guaranteed housing for all four undergraduate years. Undergraduates live in 80 different houses, including dormitories, co-ops, row houses, and fraternities and sororities. At Manzanita Park, 118 mobile homes were installed as "temporary" housing from 1969 to 1991, but as of 2015 was the site of newer dorms Castano, Kimball, Lantana, and the Humanities House, completed in 2015. Most student residences are just outside the campus core, within ten minutes (on foot or bike) of most classrooms and libraries. Some are reserved for freshman, sophomores, or upperclass students and some are open to all four classes. Most residences are co-ed; seven are all-male fraternities, three are all-female sororities, and there is also one all-female non-sorority house, Roth House. In most residences, men and women live on the same floor, but a few dorms are configured for men and women to live on separate floors (single-gender floors). Several residences are considered theme houses. The Academic, Language and Culture Houses include EAST (Education And Society Themed House), Hammarskjöld (International Themed House), Haus Mitteleuropa (Central European Themed House), La Casa Italiana (Italian Language and Culture), La Maison Française (French Language and Culture House), Slavianskii Dom (Slavic/East European Themed House), Storey (Human Biology Themed House), and Yost (Spanish Language and Culture). Cross-Cultural Themed Houses include Casa Zapata (Chicano/Latino Theme in Stern Hall), Muwekma-tah-ruk (American Indian/Alaska Native, and Native Hawaiian Themed House), Okada (Asian-American Themed House in Wilbur Hall), and Ujamaa (Black/African-American Themed House in Lagunita Court). Focus Houses include Freshman-Sophomore College (Academic Focus), Branner Hall (Community Service), Kimball (Arts & Performing Arts), Crothers (Global Citizenship), and Toyon (Sophomore Priority). Theme houses predating the current "theme" classification system are Columbae (Social Change Through Nonviolence, since 1970), and Synergy (Exploring Alternatives, since 1972). Co-ops or "Self-Ops" are another housing option. These houses feature cooperative living, where residents and eating associates each contribute work to keep the house running, such as cooking meals or cleaning shared spaces. These houses have unique themes around which their community is centered. Many co-ops are hubs of music, art and philosophy. The co-ops on campus are 576 Alvarado Row (formerly Chi Theta Chi), Columbae, Enchanted Broccoli Forest (EBF), Hammarskjöld, Kairos, Terra (the unofficial LGBT house), and Synergy. Phi Sigma, at 1018 Campus Drive was formerly Phi Sigma Kappa fraternity, but in 1973 became a Self-Op. As of 2015 around 55 percent of the graduate student population lived on campus. First-year graduate students are guaranteed on-campus housing. Stanford also subsidizes off-campus apartments in nearby Palo Alto, Menlo Park, and Mountain View for graduate students who are guaranteed on-campus housing but are unable to live on campus due to a lack of space. As of 2016 Stanford had 16 male varsity sports and 20 female varsity sports, 19 club sports and about 27 intramural sports In 1930, following a unanimous vote by the Executive Committee for the Associated Students, the athletic department adopted the mascot "Indian." The Indian symbol and name were later dropped by President Richard Lyman in 1972, after objections from Native American students and a vote by the student senate. The sports teams are now officially referred to as the "Stanford Cardinal", referring to the deep red color, not the cardinal bird. Stanford is a member of the Pac-12 Conference in most sports, the Mountain Pacific Sports Federation in several other sports, and the America East Conference in field hockey with the participation in the inter-collegiate NCAA's Division I FBS. Its traditional sports rival is Berkeley, the neighbor to the north in the East Bay. The winner of the annual "Big Game" between the Cal and Cardinal football teams gains custody of the Stanford Axe. Stanford has had at least one NCAA team champion every year since the 1976–77 school year and has earned 126 NCAA national team titles since its establishment, the most among universities, and Stanford has won 522 individual national championships, the most by any university. Stanford has won the award for the top-ranked Division 1 athletic program — the NACDA Directors' Cup, formerly known as the Sears Cup – annually for the past twenty-four straight years. Stanford athletes have won medals in every Olympic Games since 1912, winning 270 Olympic medals total, 139 of them gold. In the 2008 Summer Olympics, and 2016 Summer Olympics, Stanford won more Olympic medals than any other university in the United States. Stanford athletes won 16 medals at the 2012 Summer Olympics (12 gold, 2 silver and 2 bronze), and 27 medals at the 2016 Summer Olympics. Students and staff at Stanford are of many different religions. The Stanford Office for Religious Life's mission is "to guide, nurture and enhance spiritual, religious and ethical life within the Stanford University community" by promoting enriching dialogue, meaningful ritual, and enduring friendships among people of all religious backgrounds. It is headed by a dean with the assistance of a senior associate dean and an associate dean. Stanford Memorial Church, in the center of campus, has a Sunday University Public Worship service (UPW) usually in the "Protestant Ecumenical Christian" tradition where the Memorial Church Choir sings and a sermon is preached usually by one of the Stanford deans for Religious Life. UPW sometimes has multifaith services. In addition, the church is used by the Catholic community and by some of the other Christian denominations at Stanford. Weddings happen most Saturdays and the university has for over 20 years allowed blessings of same-gender relationships and now legal weddings. In addition to the church, the Office for Religious Life has a Center for Inter-Religious Community, Learning and Experiences (CIRCLE) on the third floor of Old Union. It offers a common room, an interfaith sanctuary, a seminar room, a student lounge area, and a reading room, as well as offices housing a number of Stanford Associated Religions (SAR) member groups and the Senior Associate Dean and Associate Dean for Religious Life. Most though not all religious student groups belong to SAR. The SAR directory includes organizations that serve atheist, Baha'i, Buddhist, Christian, Hindu, Islam, Jewish, and Sikh groups, though these groups vary year by year. The Windhover Contemplation Center was dedicated in October 2014, and was intended to provide spiritual sanctuary for students and staff in the midst of their course and work schedules; the center displays the "Windhover" paintings by Nathan Oliveira, the late Stanford professor and artist. Some religions have a larger and more formal presence on campus in addition to the student groups; these include the Catholic Community at Stanford and Hillel at Stanford. Fraternities and sororities have been active on the Stanford campus since 1891, when the university first opened. In 1944, University President Donald Tresidder banned all Stanford sororities due to extreme competition. However, following Title IX, the Board of Trustees lifted the 33-year ban on sororities in 1977. Students are not permitted to join a fraternity or sorority until spring quarter of their freshman year. As of 2016 Stanford had 31 Greek organizations, including 14 sororities and 16 fraternities. Nine of the Greek organizations were housed (eight in University-owned houses and one, Sigma Chi, in their own house, although the land is owned by the University). Six chapters were members of the African American Fraternal and Sororal Association, 11 chapters were members of the Interfraternity Council, seven chapters belonged to the Intersorority Council, and six chapters belonged to the Multicultural Greek Council. As of 2020, Stanford had more than 600 student organizations. Groups are often, though not always, partially funded by the University via allocations directed by the student government organization, the ASSU. These funds include "special fees", which are decided by a Spring Quarter vote by the student body. Groups span Athletics/Recreation (see section on Athletics), Careers/Pre-professional, Community Service, Ethnic/Cultural, Fraternities/Sororities, Health/Counseling, Media/Publications, Music/Dance/Creative Arts (see section on Arts), Political/Social Awareness and Religious/Philosophical. Stanford is home to a set of student journalism publications. "The Stanford Daily" is a student-run daily newspaper and has been published since the University was founded in 1892. The student-run radio station, KZSU Stanford 90.1 FM, features freeform music programming, sports commentary, and news segments; it started in 1947 as an AM radio station. The Stanford Review is a conservative student newspaper founded in 1987. "The Fountain Hopper" ("FoHo") is a financially independent, anonymous student-run campus rag publication, notable for having broken the Brock Turner story. Stanford is also home to a large number of pre-professional student organizations, organized around missions from startup incubation to paid consulting. The Business Association of Stanford Entrepreneurial Students (BASES) is one of the largest professional organizations in Silicon Valley, with over 5,000 members. Its goal is to support the next generation of entrepreneurs. StartX is a non-profit startup accelerator for student and faculty-led startups that over 12% of the study body has applied to. It is staffed primarily by students. Stanford Women In Business (SWIB) is an on-campus business organization, aimed at helping Stanford women find paths to success in the generally male-dominated technology industry. Stanford Marketing is a student group that provides students hands-on training through research and strategy consulting projects with Fortune 500 clients, as well as workshops led by people from industry and professors in the Stanford Graduate School of Business. Stanford Finance provides mentoring and internships for students who want to enter a career in finance. Students run "SUpost.com", an online marketplace for Stanford students and alumni, in partnership with Stanford Student Enterprises (SSE) and the Stanford Pre-Business Association. The latter is intended to build connections among industry, alumni, and student communities. Other groups include: Stanford's Department of Public Safety is responsible for law enforcement and safety on the main campus. Its deputy sheriffs are peace officers by arrangement with the Santa Clara County Sheriff's Office. The department is also responsible for publishing an annual crime report covering the previous three years as required by the Clery Act. Fire protection has been provided by contract with the Palo Alto Fire Department since 1976. Murder is rare on the campus though a few of the cases have been notorious including Theodore Streleski's murder of his professor in 1978 and the unsolved 1974 murder of Arlis Perry in Stanford Memorial Church. In 2014, Stanford University was the tenth highest in the nation in "total of reports of rape" on their main campus, with 26 reports of rape. In Stanford University's 2015 Campus Climate Survey, 4.7 percent of female undergraduates reported experiencing sexual assault as defined by the university and 32.9 percent reported experiencing sexual misconduct. According to the survey, 85% of perpetrators of misconduct were Stanford students and 80% were men. Perpetrators of sexual misconduct were frequently aided by alcohol or drugs, according to the survey: "Nearly three-fourths of the students whose responses were categorized as sexual assault indicated that the act was accomplished by a person or person taking advantage of them when they were drunk or high, according to the survey. Close to 70 percent of students who reported an experience of sexual misconduct involving nonconsensual penetration and/or oral sex indicated the same." Associated Students of Stanford University and student and alumni activists with the anti-rape group Stand with Leah criticized the survey methodology for downgrading incidents involving alcohol if students did not check two separate boxes indicating they were both intoxicated and incapacity while sexually assaulted. Reporting on the Brock Turner rape case, a reporter from "The Washington Post" analyzed campus rape reports submitted by universities to the U.S. Department of Education, and found that Stanford was one of the top ten universities in campus rapes in 2014, with 26 reported that year, but when analyzed by rapes per 1000 students, Stanford was not among the top ten. On the night of January 17–18, 2015, 22-year-old Chanel Miller, who had visited campus to attend a party at the Kappa Alpha fraternity, was sexually assaulted by Brock Turner, a freshman who had a swimming scholarship. Two graduate students witnessed the attack and intervened, catching Turner when he tried to flee and holding him down on the ground until police arrived. Stanford immediately referred the case to prosecutors and offered Miller counseling, and within two weeks had barred Turner from campus after conducting an investigation. Turner was convicted on three felony charges in March 2016 and in June 2016 he received a jail sentence of six months and was declared a sex offender, requiring him to register as such for the rest of his life; prosecutors had sought a six-year prison sentence out of the maximum 14 years that was possible. The case and the relatively lenient sentence drew nationwide attention. Two years later the judge in the case, Stanford graduate Aaron Persky, was recalled by the voters. In February 2015, Elise Clougherty filed a sexual assault and harassment lawsuit against venture capitalist Joe Lonsdale. Lonsdale and Clougherty entered into a relationship in the spring of 2012 when she was a junior and he was her mentor in a Stanford entrepreneurship course. By the spring of 2013 Clougherty had broken off the relationship and filed charges at Stanford that Lonsdale had broken the Stanford policy against consensual relationships between students and faculty and that he had sexually assaulted and harassed her, which resulted in Lonsdale being banned from Stanford for 10 years. Lonsdale challenged Stanford's finding that he had had sexually assaulted and harassed her and Stanford rescinded that finding and the campus ban in the fall of 2015. Clougherty withdrew her suit that fall as well. As of late 2016, Stanford had 2,153 tenure-line faculty, senior fellows, center fellows, and medical center faculty. Stanford's current community of scholars includes: Stanford's faculty and former faculty includes 46 Nobel laureates, 5 Fields Medalists, as well as 16 winners of the Turing Award, the so-called "Nobel Prize in computer science", comprising one third of the awards given in its 44-year history. The university has 27 ACM fellows. It is also affiliated with 4 Gödel Prize winners, 4 Knuth Prize recipients, 10 IJCAI Computers and Thought Award winners, and about 15 Grace Murray Hopper Award winners for their work in the foundations of computer science. Stanford alumni have started many companies and, according to "Forbes", has produced the second highest number of billionaires of all universities. 13 Stanford alumni have won the Nobel Prize. As of 2019, 122 Stanford students or alumni have been named Rhodes Scholars.
https://en.wikipedia.org/wiki?curid=26977
Sun Microsystems Sun Microsystems, Inc. (Sun for short) was an American company that sold computers, computer components, software, and information technology services and created the Java programming language, the Solaris operating system, ZFS, the Network File System (NFS), and SPARC. Sun contributed significantly to the evolution of several key computing technologies, among them Unix, RISC processors, thin client computing, and virtualized computing. Sun was founded on February 24, 1982. At its height, the Sun headquarters were in Santa Clara, California (part of Silicon Valley), on the former west campus of the Agnews Developmental Center. On April 20, 2009, it was announced that Oracle Corporation would acquire Sun for 7.4 billion. The deal was completed on January 27, 2010. Sun products included computer servers and workstations built on its own RISC-based SPARC processor architecture, as well as on x86-based AMD Opteron and Intel Xeon processors. Sun also developed its own storage systems and a suite of software products, including the Solaris operating system, developer tools, Web infrastructure software, and identity management applications. technologies included the Java platform and NFS. In general, Sun was a proponent of open systems, particularly Unix. It was also a major contributor to open-source software, as evidenced by its $1 billion purchase, in 2008, of MySQL, an open-source relational database management system. At various times, Sun had manufacturing facilities in several locations worldwide, including Newark, California; Hillsboro, Oregon; and Linlithgow, Scotland. However, by the time the company was acquired by Oracle, it had outsourced most manufacturing responsibilities. The initial design for what became Sun's first Unix workstation, the Sun-1, was conceived by Andy Bechtolsheim when he was a graduate student at Stanford University in Palo Alto, California. Bechtolsheim originally designed the SUN workstation for the Stanford University Network communications project as a personal CAD workstation. It was designed around the Motorola 68000 processor with an advanced memory management unit (MMU) to support the Unix operating system with virtual memory support. He built the first ones from spare parts obtained from Stanford's Department of Computer Science and Silicon Valley supply houses. On February 24, 1982, Scott McNealy, Andy Bechtolsheim, and Vinod Khosla, all Stanford graduate students, founded "Sun Microsystems". Bill Joy of Berkeley, a primary developer of the Berkeley Software Distribution (BSD), joined soon after and is counted as one of the original founders. The Sun name is derived from the initials of the Stanford University Network. Sun was profitable from its first quarter in July 1982. By 1983 Sun was known for producing 68k-based systems with high-quality graphics that were the only computers other than DEC's VAX to run 4.2BSD. It licensed the computer design to other manufacturers, which typically used it to build Multibus-based systems running Unix from UniSoft. Sun's initial public offering was in 1986 under the stock symbol "SUNW", for "Sun Workstations" (later "Sun Worldwide"). The symbol was changed in 2007 to "JAVA"; Sun stated that the brand awareness associated with its Java platform better represented the company's current strategy. Sun's logo, which features four interleaved copies of the word "sun" in the form of a rotationally symmetric ambigram, was designed by professor Vaughan Pratt, also of Stanford. The initial version of the logo was orange and had the sides oriented horizontally and vertically, but it was subsequently rotated to stand on one corner and re-colored purple, and later blue. In the dot-com bubble, Sun began making much more money, and its shares rose dramatically. It also began spending much more, hiring workers and building itself out. Some of this was because of genuine demand, but much was from web start-up companies anticipating business that would never happen. In 2000, the bubble burst. Sales in Sun's important hardware division went into free-fall as customers closed shop and auctioned high-end servers. Several quarters of steep losses led to executive departures, rounds of layoffs, and other cost cutting. In December 2001, the stock fell to the 1998, pre-bubble level of about $100. It continued to fall, faster than many other technology companies. A year later, it had reached below $10 (a tenth of what it was in 1990), but it eventually bounced back to $20. In mid-2004, Sun closed their Newark, California, factory and consolidated all manufacturing to Hillsboro, Oregon and Linlithgow, Scotland. In 2006, the rest of the Newark campus was put on the market. In 2004, Sun canceled two major processor projects which emphasized high instruction-level parallelism and operating frequency. Instead, the company chose to concentrate on processors optimized for multi-threading and multiprocessing, such as the UltraSPARC T1 processor (codenamed "Niagara"). The company also announced a collaboration with Fujitsu to use the Japanese company's processor chips in mid-range and high-end Sun servers. These servers were announced on April 17, 2007, as the M-Series, part of the SPARC Enterprise series. In February 2005, Sun announced the Sun Grid, a grid computing deployment on which it offered utility computing services priced at US$1 per CPU/hour for processing and per GB/month for storage. This offering built upon an existing 3,000-CPU server farm used for internal R&D for over 10 years, which Sun marketed as being able to achieve 97% utilization. In August 2005, the first commercial use of this grid was announced for financial risk simulations which was later launched as its first software as a service product. In January 2005, Sun reported a net profit of $19 million for fiscal 2005 second quarter, for the first time in three years. This was followed by net loss of $9 million on GAAP basis for the third quarter 2005, as reported on April 14, 2005. In January 2007, Sun reported a net GAAP profit of $126 million on revenue of $3.337 billion for its fiscal second quarter. Shortly following that news, it was announced that Kohlberg Kravis Roberts (KKR) would invest $700 million in the company. Sun had engineering groups in Bangalore, Beijing, Dublin, Grenoble, Hamburg, Prague, St. Petersburg, Tel Aviv, Tokyo, Canberra and Trondheim. In 2007–2008, Sun posted revenue of $13.8 billion and had $2 billion in cash. First-quarter 2008 losses were $1.68 billion; revenue fell 7% to $12.99 billion. Sun's stock lost 80% of its value November 2007 to November 2008, reducing the company's market value to $3 billion. With falling sales to large corporate clients, Sun announced plans to lay off 5,000 to 6,000 workers, or 15–18% of its work force. It expected to save $700 million to $800 million a year as a result of the moves, while also taking up to $600 million in charges. As of May 11, 2009, the following shareholders held over 100,000 common shares of Sun and at $9.50 per share offered by Oracle, they received the amounts indicated when the acquisition closed. For the first decade of Sun's history, the company positioned its products as technical workstations, competing successfully as a low-cost vendor during the Workstation Wars of the 1980s. It then shifted its hardware product line to emphasize servers and storage. High-level telecom control systems such as Operational Support Systems service predominantly used Sun equipment. Sun originally used Motorola 68000 family central processing units for the Sun-1 through Sun-3 computer series. The Sun-1 employed a 68000 CPU, the Sun-2 series, a 68010. The Sun-3 series was based on the 68020, with the later Sun-3x using the 68030. In 1987, the company began using "SPARC", a RISC processor architecture of its own design, in its computer systems, starting with the Sun-4 line. SPARC was initially a 32-bit architecture (SPARC V7) until the introduction of the SPARC V9 architecture in 1995, which added 64-bit extensions. Sun has developed several generations of SPARC-based computer systems, including the SPARCstation, Ultra and Sun Blade series of workstations, and the SPARCserver, Netra, Enterprise and Sun Fire line of servers. In the early 1990s the company began to extend its product line to include large-scale symmetric multiprocessing servers, starting with the four-processor SPARCserver 600MP. This was followed by the 8-processor SPARCserver 1000 and 20-processor SPARCcenter 2000, which were based on work done in conjunction with Xerox PARC. In 1995 the company introduced Sun Ultra series machines that were equipped with the first 64-bit implementation of SPARC processors (UltraSPARC). In the late 1990s the transformation of product line in favor of large 64-bit SMP systems was accelerated by the acquisition of Cray Business Systems Division from Silicon Graphics. Their 32-bit, 64-processor Cray Superserver 6400, related to the SPARCcenter, led to the 64-bit Sun Enterprise 10000 high-end server (otherwise known as "Starfire"). In September 2004 Sun made available systems with UltraSPARC IV which was the first multi-core SPARC processor. It was followed by UltraSPARC IV+ in September 2005 and its revisions with higher clock speeds in 2007. These CPUs were used in the most powerful, enterprise class high-end CC-NUMA servers developed by Sun, such as Sun Fire E25K. In November 2005 Sun launched the UltraSPARC T1, notable for its ability to concurrently run 32 threads of execution on 8 processor cores. Its intent was to drive more efficient use of CPU resources, which is of particular importance in data centers, where there is an increasing need to reduce power and air conditioning demands, much of which comes from the heat generated by CPUs. The T1 was followed in 2007 by the UltraSPARC T2, which extended the number of threads per core from 4 to 8. Sun has open sourced the design specifications of both the T1 and T2 processors via the OpenSPARC project. In 2006, Sun ventured into the "blade server" (high density rack-mounted systems) market with the Sun Blade (distinct from the Sun Blade workstation). In April 2007 Sun released the SPARC Enterprise server products, jointly designed by Sun and Fujitsu and based on Fujitsu SPARC64 VI and later processors. The "M-class" SPARC Enterprise systems include high-end reliability and availability features. Later T-series servers have also been badged SPARC Enterprise rather than Sun Fire. In April 2008 Sun released servers with UltraSPARC T2 Plus, which is an SMP capable version of UltraSPARC T2, available in 2 or 4 processor configurations. It was the first CoolThreads CPU with multi-processor capability and it made possible to build standard rack-mounted servers that could simultaneously process up to massive 256 CPU threads in hardware (Sun SPARC Enterprise T5440), which is considered a record in the industry. Since 2010, all further development of Sun machines based on SPARC architecture (including new SPARC T-Series servers, SPARC T3 and T4 chips) is done as a part of Oracle Corporation hardware division. In the late 1980s, Sun also marketed an Intel 80386-based machine, the Sun386i; this was designed to be a hybrid system, running SunOS but at the same time supporting DOS applications. This only remained on the market for a brief time. A follow-up "486i" upgrade was announced but only a few prototype units were ever manufactured. Sun's brief first foray into x86 systems ended in the early 1990s, as it decided to concentrate on SPARC and retire the last Motorola systems and 386i products, a move dubbed by McNealy as "all the wood behind one arrowhead". Even so, Sun kept its hand in the x86 world, as a release of Solaris for PC compatibles began shipping in 1993. In 1997 Sun acquired Diba, Inc., followed later by the acquisition of Cobalt Networks in 2000, with the aim of building "network appliances" (single function computers meant for consumers). Sun also marketed a Network Computer (a term popularized and eventually trademarked by Oracle); the JavaStation was a diskless system designed to run Java applications. Although none of these business initiatives were particularly successful, the Cobalt purchase gave Sun a toehold for its return to the x86 hardware market. In 2002, Sun introduced its first general purpose x86 system, the LX50, based in part on previous Cobalt system expertise. This was also Sun's first system announced to support Linux as well as Solaris. In 2003, Sun announced a strategic alliance with AMD to produce x86/x64 servers based on AMD's Opteron processor; this was followed shortly by Sun's acquisition of Kealia, a startup founded by original Sun founder Andy Bechtolsheim, which had been focusing on high-performance AMD-based servers. The following year, Sun launched the Opteron-based Sun Fire V20z and V40z servers, and the Java Workstation W1100z and W2100z workstations. On September 12, 2005, Sun unveiled a new range of Opteron-based servers: the Sun Fire X2100, X4100 and X4200 servers. These were designed from scratch by a team led by Bechtolsheim to address heat and power consumption issues commonly faced in data centers. In July 2006, the Sun Fire X4500 and X4600 systems were introduced, extending a line of x64 systems that support not only Solaris, but also Linux and Microsoft Windows. On January 22, 2007, Sun announced a broad strategic alliance with Intel. Intel endorsed Solaris as a mainstream operating system and as its mission critical Unix for its Xeon processor-based systems, and contributed engineering resources to OpenSolaris. Sun began using the Intel Xeon processor in its x64 server line, starting with the Sun Blade X6250 server module introduced in June 2007. On May 5, 2008, AMD announced its Operating System Research Center (OSRC) expanded its focus to include optimization to Sun's OpenSolaris and xVM virtualization products for AMD based processors. Although Sun was initially known as a hardware company, its software history began with its founding in 1982; co-founder Bill Joy was one of the leading Unix developers of the time, having contributed the vi editor, the C shell, and significant work developing TCP/IP and the BSD Unix OS. Sun later developed software such as the Java programming language and acquired software such as StarOffice, VirtualBox and MySQL. Sun used community-based and open-source licensing of its major technologies, and for its support of its products with other open source technologies. GNOME-based desktop software called Java Desktop System (originally code-named "Madhatter") was distributed for the Solaris operating system, and at one point for Linux. Sun supported its Java Enterprise System (a middleware stack) on Linux. It released the source code for Solaris under the open-source Common Development and Distribution License, via the OpenSolaris community. Sun's positioning includes a commitment to indemnify users of some software from intellectual property disputes concerning that software. It offers support services on a variety of pricing bases, including per-employee and per-socket. A 2006 report prepared for the EU by UNU-MERIT stated that Sun was the largest corporate contributor to open source movements in the world. According to this report, Sun's open source contributions exceed the combined total of the next five largest commercial contributors. Sun is best known for its Unix systems, which have a reputation for system stability and a consistent design philosophy. Sun's first workstation shipped with UniSoft V7 Unix. Later in 1982 Sun began providing SunOS, a customized 4.1BSD Unix, as the operating system for its workstations. In the late 1980s, AT&T tapped Sun to help them develop the next release of their branded UNIX, and in 1988 announced they would purchase up to a 20% stake in Sun. UNIX System V Release 4 (SVR4) was jointly developed by AT&T and Sun; Sun used SVR4 as the foundation for Solaris 2.x, which became the successor to SunOS 4.1.x (later retrospectively named Solaris 1.x). By the mid-1990s, the ensuing Unix wars had largely subsided, AT&T had sold off their Unix interests, and the relationship between the two companies was significantly reduced. From 1992 Sun also sold Interactive Unix, an operating system it acquired when it bought Interactive Systems Corporation from Eastman Kodak Company. This was a popular Unix variant for the PC platform and a major competitor to market leader SCO UNIX. Sun's focus on Interactive Unix diminished in favor of Solaris on both SPARC and x86 systems; it was dropped as a product in 2001. Sun dropped the Solaris 2.x version numbering scheme after the Solaris 2.6 release (1997); the following version was branded Solaris 7. This was the first 64-bit release, intended for the new UltraSPARC CPUs based on the SPARC V9 architecture. Within the next four years, the successors Solaris 8 and Solaris 9 were released in 2000 and 2002 respectively. Following several years of difficult competition and loss of server market share to competitors' Linux-based systems, Sun began to include Linux as part of its strategy in 2002. Sun supported both Red Hat Enterprise Linux and SUSE Linux Enterprise Server on its x64 systems; companies such as Canonical Ltd., Wind River Systems and MontaVista also supported their versions of Linux on Sun's SPARC-based systems. In 2004, after having cultivated a reputation as one of Microsoft's most vocal antagonists, Sun entered into a joint relationship with them, resolving various legal entanglements between the two companies and receiving US$1.95 billion in settlement payments from them. Sun supported Microsoft Windows on its x64 systems, and announced other collaborative agreements with Microsoft, including plans to support each other's virtualization environments. In 2005, the company released Solaris 10. The new version included a large number of enhancements to the operating system, as well as very novel features, previously unseen in the industry. Solaris 10 update releases continued through the next 8 years, the last release from Sun Microsystems being Solaris 10 10/09. The following updates were released by Oracle under the new license agreement; the final release is Solaris 10 1/13. Previously, Sun offered a separate variant of Solaris called Trusted Solaris, which included augmented security features such as multilevel security and a least privilege access model. Solaris 10 included many of the same capabilities as Trusted Solaris at the time of its initial release; Solaris 10 11/06 included Solaris Trusted Extensions, which give it the remaining capabilities needed to make it the functional successor to Trusted Solaris. After releasing Solaris 10, its source code was opened under CDDL free software license and developed in open with contributing Opensolaris community through SXCE that used SVR4 .pkg packaging and supported Opensolaris releases that used IPS. Following acquisition of Sun by Oracle , Opensolaris continued to develop in open under illumos with illumos distributions. Oracle Corporation continued to develop OpenSolaris into next Solaris release, changing back the license to proprietary, and released it as Oracle Solaris 11 in November 2011. The Java platform was developed at Sun by James Gosling in the early 1990s with the objective of allowing programs to function regardless of the device they were used on, sparking the slogan "Write once, run anywhere" (WORA). While this objective was not entirely achieved (prompting the riposte "Write once, debug everywhere"), Java is regarded as being largely hardware- and operating system-independent. Java was initially promoted as a platform for client-side "applets" running inside web browsers. Early examples of Java applications were the HotJava web browser and the HotJava Views suite. However, since then Java has been more successful on the server side of the Internet. The platform consists of three major parts: the Java programming language, the Java Virtual Machine (JVM), and several Java Application Programming Interfaces (APIs). The design of the Java platform is controlled by the vendor and user community through the Java Community Process (JCP). Java is an object-oriented programming language. Since its introduction in late 1995, it became one of the world's most popular programming languages. Java programs are compiled to byte code, which can be executed by any JVM, regardless of the environment. The Java APIs provide an extensive set of library routines. These APIs evolved into the "Standard Edition" (Java SE), which provides basic infrastructure and GUI functionality; the "Enterprise Edition" (Java EE), aimed at large software companies implementing enterprise-class application servers; and the "Micro Edition" (Java ME), used to build software for devices with limited resources, such as mobile devices. On November 13, 2006, Sun announced it would be licensing its Java implementation under the GNU General Public License; it released its Java compiler and JVM at that time. In February 2009 Sun entered a battle with Microsoft and Adobe Systems, which promoted rival platforms to build software applications for the Internet. JavaFX was a development platform for music, video and other applications that builds on the Java programming language. In 1999, Sun acquired the German software company Star Division and with it the office suite StarOffice, which Sun later released as OpenOffice.org under both GNU LGPL and the SISSL (Sun Industry Standards Source License). OpenOffice.org supported Microsoft Office file formats (though not perfectly), was available on many platforms (primarily Linux, Microsoft Windows, Mac OS X, and Solaris) and was used in the open source community. The principal differences between StarOffice and OpenOffice.org were that StarOffice was supported by Sun, was available as either a single-user retail box kit or as per-user blocks of licensing for the enterprise, and included a wider range of fonts and document templates and a commercial quality spellchecker. StarOffice also contained commercially licensed functions and add-ons; in OpenOffice.org these were either replaced by open-source or free variants, or are not present at all. Both packages had native support for the OpenDocument format. In 2007, Sun announced the Sun xVM virtualization and datacenter automation product suite for commodity hardware. Sun also acquired VirtualBox in 2008. Earlier virtualization technologies from Sun like "Dynamic System Domains" and "Dynamic Reconfiguration" were specifically designed for high-end SPARC servers, and Logical Domains only supports the UltraSPARC T1/T2/T2 Plus server platforms. Sun marketed "Sun Ops Center" provisioning software for datacenter automation. On the client side, Sun offered virtual desktop solutions. Desktop environments and applications could be hosted in a datacenter, with users accessing these environments from a wide range of client devices, including Microsoft Windows PCs, Sun Ray virtual display clients, Apple Macintoshes, PDAs or any combination of supported devices. A variety of networks were supported, from LAN to WAN or the public Internet. Virtual desktop products included Sun Ray Server Software, Sun Secure Global Desktop and Sun Virtual Desktop Infrastructure. Sun acquired MySQL AB, the developer of the MySQL database in 2008 for US$1 billion. CEO Jonathan Schwartz mentioned in his blog that optimizing the performance of MySQL was one of the priorities of the acquisition. In February 2008, Sun began to publish results of the MySQL performance optimization work. Sun contributed to the PostgreSQL project. On the Java platform, Sun contributed to and supported Java DB. Sun offered other software products for software development and infrastructure services. Many were developed in house; others came from acquisitions, including Tarantella, Waveset Technologies, SeeBeyond, and Vaau. Sun acquired many of the Netscape non-browser software products as part a deal involving Netscape's merger with AOL. These software products were initially offered under the "iPlanet" brand; once the Sun-Netscape alliance ended, they were re-branded as "Sun ONE" (Sun Open Network Environment), and then the "Sun Java System". Sun's middleware product was branded as the "Java Enterprise System" (or JES), and marketed for web and application serving, communication, calendaring, directory, identity management and service-oriented architecture. Sun's Open ESB and other software suites were available free of charge on systems running Solaris, Red Hat Enterprise Linux, HP-UX, and Windows, with support available optionally. Sun developed data center management software products, which included the "Solaris Cluster" high availability software, and a grid management package called "Sun Grid Engine" and firewall software such as SunScreen. For Network Equipment Providers and telecommunications customers, Sun developed the Sun Netra High-Availability Suite. Sun produced compilers and development tools under the "Sun Studio" brand, for building and developing Solaris and Linux applications. Sun entered the software as a service (SaaS) market with zembly, a social cloud-based computing platform and Project Kenai, an open-source project hosting service. Sun sold its own storage systems to complement its system offerings; it has also made several storage-related acquisitions. On June 2, 2005, Sun announced it would purchase Storage Technology Corporation (StorageTek) for US$4.1 billion in cash, or $37.00 per share, a deal completed in August 2005. In 2006, Sun introduced the Sun StorageTek 5800 System, the first application-aware programmable storage solution. In 2008, Sun contributed the source code of the StorageTek 5800 System under the BSD license. Sun announced the Sun Open Storage platform in 2008 built with open source technologies. In late 2008 Sun announced the Sun Storage 7000 Unified Storage systems (codenamed Amber Road). Transparent placement of data in the systems' solid-state drives (SSD) and conventional hard drives was managed by ZFS to take advantage of the speed of SSDs and the economy of conventional hard disks. Other storage products included Sun Fire X4500 storage server and SAM-QFS filesystem and storage management software. Sun marketed the Sun Constellation System for high-performance computing (HPC). Even before the introduction of the Sun Constellation System in 2007, Sun's products were in use in many of the TOP500 systems and supercomputing centers: The "Sun HPC ClusterTools" product was a set of Message Passing Interface (MPI) libraries and tools for running parallel jobs on Solaris HPC clusters. Beginning with version 7.0, Sun switched from its own implementation of MPI to Open MPI, and donated engineering resources to the Open MPI project. Sun was a participant in the OpenMP language committee. Sun Studio compilers and tools implemented the OpenMP specification for shared memory parallelization. In 2006, Sun built the "TSUBAME supercomputer", which was until June 2008 the fastest supercomputer in Asia. Sun built "Ranger" at the Texas Advanced Computing Center (TACC) in 2007. Ranger had a peak performance of over 500 TFLOPS, and was the 6th most powerful supercomputer on the TOP500 list in November 2008. Sun announced an OpenSolaris distribution that integrated Sun's HPC products with others. Notable Sun employees included John Gilmore, Whitfield Diffie, Radia Perlman, and Marc Tremblay. Sun was an early advocate of Unix-based networked computing, promoting TCP/IP and especially NFS, as reflected in the company's motto "The Network Is The Computer", coined by John Gage. James Gosling led the team which developed the Java programming language. Jon Bosak led the creation of the XML specification at W3C. Sun staff published articles on the company's blog site. Staff were encouraged to use the site to blog on any aspect of their work or personal life, with few restrictions placed on staff, other than commercially confidential material. Jonathan I. Schwartz was one of the first CEOs of large companies to regularly blog; his postings were frequently quoted and analyzed in the press. In 2005, Sun Microsystems was one of the first Fortune 500 companies that instituted a formal Social Media program. Sun was sold to Oracle Corporation in 2009 for $5.6B. Sun's staff were asked to share anecdotes about their experiences at Sun. A web site containing videos, stories, and photographs from 27 years at Sun was made available on September 2, 2009. In October, Sun announced a second round of thousands of employees to be laid off, blamed partially on delays in approval of the merger. The transaction was completed in early 2010. In January 2011, Oracle agreed to pay $46 million to settle charges that it submitted false claims to US federal government agencies and paid "kickbacks" to systems integrators. In February 2011, Sun's former Menlo Park, California, campus of about was sold, and it was announced that it would become headquarters for Facebook. The sprawling facility built around an enclosed courtyard had been nicknamed "Sun Quentin". On September 1, 2011, Sun India legally became part of Oracle. It had been delayed due to legal issues in Indian court.
https://en.wikipedia.org/wiki?curid=26980
Saladin An-Nasir Salah ad-Din Yusuf ibn Ayyub ( / ALA-LC: "Ṣalāḥ ad-Dīn Yūsuf ibn Ayyūb"; / ALA-LC: "Selahedînê Eyûbî"), known as Salah ad-Din or Saladin (; 11374 March 1193), was the first sultan of Egypt and Syria and the founder of the Ayyubid dynasty. A Sunni Muslim of Kurdish ethnicity, Saladin led the Muslim military campaign against the Crusader states in the Levant. During his reign, Saladin has been described as the de facto Caliph of Islam and at the height of his power, his caliphate included Egypt, Syria, Upper Mesopotamia, the Hejaz, Yemen and other parts of North Africa. He was originally sent to Fatimid Egypt in 1164 alongside his uncle Shirkuh, a general of the Zengid army, on the orders of their lord Nur ad-Din to help restore Shawar as vizier of the teenage Fatimid caliph . A power struggle ensued between Shirkuh and Shawar after the latter was reinstated. Saladin, meanwhile, climbed the ranks of the Fatimid government by virtue of his military successes against Crusader assaults against its territory and his personal closeness to al-Adid. After Shawar was assassinated and Shirkuh died in 1169, al-Adid appointed Saladin vizier, a rare nomination of a Sunni Muslim to such an important position in the Isma'ili Shia caliphate. During his tenure as vizier, Saladin began to undermine the Fatimid establishment and, following al-Adid's death in 1171, he abolished the and realigned the country's allegiance with the Sunni, Baghdad-based Abbasid Caliphate. In the following years, he led forays against the Crusaders in Palestine, commissioned the successful conquest of Yemen, and staved off pro-Fatimid rebellions in Upper Egypt. Not long after Nur ad-Din's death in 1174, Saladin launched his conquest of Syria, peacefully entering Damascus at the request of its governor. By mid-1175, Saladin had conquered Hama and Homs, inviting the animosity of other Zengid lords, the official rulers of Syria's various regions. Soon after, he defeated the Zengid army at the Battle of the Horns of Hama and was thereafter proclaimed the "Sultan of Egypt and Syria" by the Abbasid caliph al-Mustadi. Saladin made further conquests in northern Syria and Jazira, escaping two attempts on his life by Assassins, before returning to Egypt in 1177 to address issues there. By 1182, Saladin had completed the conquest of Muslim Syria after capturing Aleppo, but ultimately failed to take over the Zengid stronghold of Mosul. Under Saladin's command, the Ayyubid army defeated the Crusaders at the decisive Battle of Hattin in 1187, and thereafter wrested control of Palestine—including the city of Jerusalem—from the Crusaders, who had conquered the area 88 years earlier. Although the Crusader Kingdom of Jerusalem continued to exist until the late 13th century, its defeat at Hattin marked a turning point in its conflict with the Muslim powers of the region. Saladin died in Damascus in 1193, having given away much of his personal wealth to his subjects. He is buried in a mausoleum adjacent to the Umayyad Mosque. Saladin has become a prominent figure in Muslim, Arab, Turkish and Kurdish culture, and he has often been described as being the most famous Kurd in history. Saladin was born in Tikrit in modern-day Iraq. His personal name was "Yusuf"; "Salah ad-Din" is a "laqab", an honorific epithet, meaning "Righteousness of the Faith". His family was most likely of Kurdish ancestry, and had originated from the village of Ajdanakan near the city of Dvin in central Armenia. The Rawadiya tribe he hailed from had been partially assimilated into the Arabic-speaking world by this time. In 1132, the defeated army of Imad ad-Din Zengi, the ruler of Mosul, found their retreat blocked by the Tigris River opposite the fortress of Tikrit, where Saladin's father, Najm ad-Din Ayyub served as the warden. Ayyub provided ferries for the army and gave them refuge in Tikrit. Mujahed al-Din Bihruz, a former Greek slave who had been appointed as the military governor of northern Mesopotamia for his service to the Seljuks, reprimanded Ayyub for giving Zengi refuge and in 1137 banished Ayyub from Tikrit after his brother Asad al-Din Shirkuh killed a friend of Bihruz. According to Baha ad-Din ibn Shaddad, Saladin was born on the same night that his family left Tikrit. In 1139, Ayyub and his family moved to Mosul, where Imad ad-Din Zengi acknowledged his debt and appointed Ayyub commander of his fortress in Baalbek. After the death of Zengi in 1146, his son, Nur ad-Din, became the regent of Aleppo and the leader of the Zengids. Saladin, who now lived in Damascus, was reported to have a particular fondness for the city, but information on his early childhood is scarce. About education, Saladin wrote "children are brought up in the way in which their elders were brought up". According to his biographers, Anne-Marie Eddé and al-Wahrani, Saladin was able to answer questions on Euclid, the Almagest, arithmetic, and law, but this was an academic ideal and it was study of the Qur'an and the "sciences of religion" that linked him to his contemporaries. Several sources claim that during his studies he was more interested in religious studies than joining the military. Another factor which may have affected his interest in religion was that, during the First Crusade, Jerusalem was taken by the Christians. In addition to Islam, Saladin had a knowledge of the genealogies, biographies, and histories of the Arabs, as well as the bloodlines of Arabian horses. More significantly, he knew the "Hamasah" of Abu Tammam by heart. He spoke Kurdish and Arabic. Saladin's military career began under the tutelage of his uncle Asad al-Din Shirkuh, a prominent military commander under Nur ad-Din, the Zengid emir of Damascus and Aleppo and the most influential teacher of Saladin. In 1163, the vizier to the Fatimid caliph al-Adid, Shawar, had been driven out of Egypt by his rival Dirgham, a member of the powerful Banu Ruzzaik tribe. He asked for military backing from Nur ad-Din, who complied and, in 1164, sent Shirkuh to aid Shawar in his expedition against Dirgham. Saladin, at age 26, went along with them. After Shawar was successfully reinstated as vizier, he demanded that Shirkuh withdraw his army from Egypt for a sum of 30,000 gold dinars, but he refused, insisting it was Nur ad-Din's will that he remain. Saladin's role in this expedition was minor, and it is known that he was ordered by Shirkuh to collect stores from Bilbais prior to its siege by a combined force of Crusaders and Shawar's troops. After the sacking of Bilbais, the Crusader-Egyptian force and Shirkuh's army were to engage in the Battle of al-Babein on the desert border of the Nile, just west of Giza. Saladin played a major role, commanding the right wing of the Zengid army, while a force of Kurds commanded the left, and Shirkuh was stationed in the center. Muslim sources at the time, however, put Saladin in the "baggage of the centre" with orders to lure the enemy into a trap by staging a feigned retreat. The Crusader force enjoyed early success against Shirkuh's troops, but the terrain was too steep and sandy for their horses, and commander Hugh of Caesarea was captured while attacking Saladin's unit. After scattered fighting in little valleys to the south of the main position, the Zengid central force returned to the offensive; Saladin joined in from the rear. The battle ended in a Zengid victory, and Saladin is credited with having helped Shirkuh in one of the "most remarkable victories in recorded history", according to Ibn al-Athir, although more of Shirkuh's men were killed and the battle is considered by most sources as not a total victory. Saladin and Shirkuh moved towards Alexandria where they were welcomed, given money, arms and provided a base. Faced by a superior Crusader-Egyptian force attempting to besiege the city, Shirkuh split his army. He and the bulk of his force withdrew from Alexandria, while Saladin was left with the task of guarding the city. Shirkuh was in a power struggle over Egypt with Shawar and Amalric I of the Kingdom of Jerusalem, in which Shawar requested Amalric's assistance. In 1169, Shawar was reportedly assassinated by Saladin, and Shirkuh died later that year. Nur ad-Din chose a successor for Shirkuh, but al-Adid appointed Saladin to replace Shawar as vizier. The reasoning behind the Shia caliph al-Adid's selection of Saladin, a Sunni, varies. Ibn al-Athir claims that the caliph chose him after being told by his advisers that "there is no one weaker or younger" than Saladin, and "not one of the emirs [commanders] obeyed him or served him". However, according to this version, after some bargaining, he was eventually accepted by the majority of the emirs. Al-Adid's advisers were also suspected of promoting Saladin in an attempt to split the Syria-based Zengids. Al-Wahrani wrote that Saladin was selected because of the reputation of his family in their "generosity and military prowess". Imad ad-Din wrote that after the brief mourning period for Shirkuh, during which "opinions differed", the Zengid emirs decided upon Saladin and forced the caliph to "invest him as vizier". Although positions were complicated by rival Muslim leaders, the bulk of the Syrian commanders supported Saladin because of his role in the Egyptian expedition, in which he gained a record of military qualifications. Inaugurated as vizier on 26 March, Saladin repented "wine-drinking and turned from frivolity to assume the dress of religion", according to Arabic sources of the time. Having gained more power and independence than ever before in his career, he still faced the issue of ultimate loyalty between al-Adid and Nur ad-Din. Later in the year, a group of Egyptian soldiers and emirs attempted to assassinate Saladin, but having already known of their intentions thanks to his intelligence chief Ali ibn Safyan, he had the chief conspirator, Naji, Mu'tamin al-Khilafa—the civilian controller of the Fatimid Palace—arrested and killed. The day after, 50,000 Black African soldiers from the regiments of the Fatimid army opposed to Saladin's rule, along with a number of Egyptian emirs and commoners, staged a revolt. By 23 August, Saladin had decisively quelled the uprising, and never again had to face a military challenge from Cairo. Towards the end of 1169, Saladin, with reinforcements from Nur ad-Din, defeated a massive Crusader-Byzantine force near Damietta. Afterward, in the spring of 1170, Nur ad-Din sent Saladin's father to Egypt in compliance with Saladin's request, as well as encouragement from the Baghdad-based Abbasid caliph, al-Mustanjid, who aimed to pressure Saladin in deposing his rival caliph, al-Adid. Saladin himself had been strengthening his hold on Egypt and widening his support base there. He began granting his family members high-ranking positions in the region; he ordered the construction of a college for the Maliki branch of Sunni Islam in the city, as well as one for the Shafi'i denomination to which he belonged in al-Fustat. After establishing himself in Egypt, Saladin launched a campaign against the Crusaders, besieging Darum in 1170. Amalric withdrew his Templar garrison from Gaza to assist him in defending Darum, but Saladin evaded their force and captured Gaza in 1187. In 1191 Saladin destroyed the fortifications in Gaza build by King Baldwin III for the Knights Templar. It is unclear exactly when, but during that same year, he attacked and captured the Crusader castle of Eilat, built on an island off the head of the Gulf of Aqaba. It did not pose a threat to the passage of the Muslim navy, but could harass smaller parties of Muslim ships and Saladin decided to clear it from his path. According to Imad ad-Din, Nur ad-Din wrote to Saladin in June 1171, telling him to reestablish the Abbasid caliphate in Egypt, which Saladin coordinated two months later after additional encouragement by Najm ad-Din al-Khabushani, the Shafi'i "faqih", who vehemently opposed Shia rule in the country. Several Egyptian emirs were thus killed, but al-Adid was told that they were killed for rebelling against him. He then fell ill, or was poisoned according to one account. While ill, he asked Saladin to pay him a visit to request that he take care of his young children, but Saladin refused, fearing treachery against the Abbasids, and is said to have regretted his action after realizing what al-Adid had wanted. He died on 13 September, and five days later, the Abbasid "khutba" was pronounced in Cairo and al-Fustat, proclaiming al-Mustadi as caliph. On 25 September, Saladin left Cairo to take part in a joint attack on Kerak and Montreal, the desert castles of the Kingdom of Jerusalem, with Nur ad-Din who would attack from Syria. Prior to arriving at Montreal, Saladin however withdrew back to Cairo as he received the reports that in his absence the Crusader leaders had increased their support to the traitors inside Egypt to attack Saladin from within and lessen his power especially the Fatimid who started plotting to restore their past glory. Because of this, Nur ad-Din went on alone. During the summer of 1173, a Nubian army along with a contingent of Armenian refugees were reported on the Egyptian border, preparing for a siege against Aswan. The emir of the city had requested Saladin's assistance and was given reinforcements under Turan-Shah, Saladin's brother. Consequently, the Nubians departed; but returned in 1173 and were again driven off. This time, Egyptian forces advanced from Aswan and captured the Nubian town of Ibrim. Saladin sent a gift to Nur ad-Din, who had been his friend and teacher, 60,000 dinars, "wonderful manufactured goods", some jewels, and an elephant. While transporting these goods to Damascus, Saladin took the opportunity to ravage the Crusader countryside. He did not press an attack against the desert castles, but attempted to drive out the Muslim Bedouins who lived in Crusader territory with the aim of depriving the Franks of guides. On 31 July 1173, Saladin's father Ayyub was wounded in a horse-riding accident, ultimately causing his death on 9 August. In 1174, Saladin sent Turan-Shah to conquer Yemen to allocate it and its port Aden to the territories of the Ayyubid Dynasty. In the early summer of 1174, Nur ad-Din was mustering an army, sending summons to Mosul, Diyar Bakr, and the Jazira in an apparent preparation of attack against Saladin's Egypt. The Ayyubids held a council upon the revelation of these preparations to discuss the possible threat and Saladin collected his own troops outside Cairo. On 15 May, Nur ad-Din died after falling ill the previous week and his power was handed to his eleven-year-old son as-Salih Ismail al-Malik. His death left Saladin with political independence and in a letter to as-Salih, he promised to "act as a sword" against his enemies and referred to the death of his father as an "earthquake shock". In the wake of Nur ad-Din's death, Saladin faced a difficult decision; he could move his army against the Crusaders from Egypt or wait until invited by as-Salih in Syria to come to his aid and launch a war from there. He could also take it upon himself to annex Syria before it could possibly fall into the hands of a rival, but he feared that attacking a land that formerly belonged to his master—forbidden in the Islamic principles in which he believed—could portray him as hypocritical, thus making him unsuitable for leading the war against the Crusaders. Saladin saw that in order to acquire Syria, he either needed an invitation from as-Salih, or to warn him that potential anarchy could give rise to danger from the Crusaders. When as-Salih was removed to Aleppo in August, Gumushtigin, the emir of the city and a captain of Nur ad-Din's veterans, assumed the guardianship over him. The emir prepared to unseat all his rivals in Syria and the Jazira, beginning with Damascus. In this emergency, the emir of Damascus appealed to Saif al-Din of Mosul (a cousin of Gumushtigin) for assistance against Aleppo, but he refused, forcing the Syrians to request the aid of Saladin, who complied. Saladin rode across the desert with 700 picked horsemen, passing through al-Kerak then reaching Bosra. According to his own account, was joined by "emirs, soldiers, and Bedouins—the emotions of their hearts to be seen on their faces." On 23 November, he arrived in Damascus amid general acclamation and rested at his father's old home there, until the gates of the Citadel of Damascus, whose commander Raihan initially refused to surrender, were opened to Saladin four days later, after a brief siege by his brother Tughtakin ibn Ayyub. He installed himself in the castle and received the homage and salutations of the inhabitants. Leaving his brother Tughtakin ibn Ayyub as Governor of Damascus, Saladin proceeded to reduce other cities that had belonged to Nur al-Din, but were now practically independent. His army conquered Hama with relative ease, but avoided attacking Homs because of the strength of its citadel. Saladin moved north towards Aleppo, besieging it on 30 December after Gumushtigin refused to abdicate his throne. As-Salih, fearing capture by Saladin, came out of his palace and appealed to the inhabitants not to surrender him and the city to the invading force. One of Saladin's chroniclers claimed "the people came under his spell". Gumushtigin requested Rashid ad-Din Sinan, chief "da'i" of the Assassins of Syria, who were already at odds with Saladin since he replaced the Fatimids of Egypt, to assassinate Saladin in his camp. On 11 May 1175, a group of thirteen Assassins easily gained admission into Saladin's camp, but were detected immediately before they carried out their attack by Nasih al-Din Khumartekin of Abu Qubays. One was killed by one of Saladin's generals and the others were slain while trying to escape. To deter Saladin's progress, Raymond of Tripoli gathered his forces by Nahr al-Kabir, where they were well placed for an attack on Muslim territory. Saladin later moved toward Homs instead, but retreated after being told a relief force was being sent to the city by Saif al-Din. Meanwhile, Saladin's rivals in Syria and Jazira waged a propaganda war against him, claiming he had "forgotten his own condition [servant of Nur ad-Din]" and showed no gratitude for his old master by besieging his son, rising "in rebellion against his Lord". Saladin aimed to counter this propaganda by ending the siege, claiming that he was defending Islam from the Crusaders; his army returned to Hama to engage a Crusader force there. The Crusaders withdrew beforehand and Saladin proclaimed it "a victory opening the gates of men's hearts". Soon after, Saladin entered Homs and captured its citadel in March 1175, after stubborn resistance from its defenders. Saladin's successes alarmed Saif al-Din. As head of the Zengids, including Gumushtigin, he regarded Syria and Mesopotamia as his family estate and was angered when Saladin attempted to usurp his dynasty's holdings. Saif al-Din mustered a large army and dispatched it to Aleppo, whose defenders anxiously had awaited them. The combined forces of Mosul and Aleppo marched against Saladin in Hama. Heavily outnumbered, Saladin initially attempted to make terms with the Zengids by abandoning all conquests north of the Damascus province, but they refused, insisting he return to Egypt. Seeing that confrontation was unavoidable, Saladin prepared for battle, taking up a superior position at the Horns of Hama, hills by the gorge of the Orontes River. On 13 April 1175, the Zengid troops marched to attack his forces, but soon found themselves surrounded by Saladin's Ayyubid veterans, who crushed them. The battle ended in a decisive victory for Saladin, who pursued the Zengid fugitives to the gates of Aleppo, forcing as-Salih's advisers to recognize Saladin's control of the provinces of Damascus, Homs and Hama, as well as a number of towns outside Aleppo such as Ma'arat al-Numan. After his victory against the Zengids, Saladin proclaimed himself king and suppressed the name of as-Salih in Friday prayers and Islamic coinage. From then on, he ordered prayers in all the mosques of Syria and Egypt as the sovereign king and he issued at the Cairo mint gold coins bearing his official title—"al-Malik an-Nasir Yusuf Ayyub, ala ghaya" "the King Strong to Aid, Joseph son of Job; exalted be the standard." The Abbasid caliph in Baghdad graciously welcomed Saladin's assumption of power and declared him "Sultan of Egypt and Syria". The Battle of Hama did not end the contest for power between the Ayyubids and the Zengids, with the final confrontation occurring in the spring of 1176. Saladin had gathered massive reinforcements from Egypt while Saif al-Din was levying troops among the minor states of Diyarbakir and al-Jazira. When Saladin crossed the Orontes, leaving Hama, the sun was eclipsed. He viewed this as an omen, but he continued his march north. He reached the Sultan's Mound, roughly from Aleppo, where his forces encountered Saif al-Din's army. A hand-to-hand fight ensued and the Zengids managed to plow Saladin's left wing, driving it before him, when Saladin himself charged at the head of the Zengid guard. The Zengid forces panicked and most of Saif al-Din's officers ended up being killed or captured—Saif al-Din narrowly escaped. The Zengid army's camp, horses, baggage, tents and stores were seized by the Ayyubids. The Zengid prisoners of war, however, were given gifts and freed. All of the booty from the Ayyubid victory was accorded to the army, Saladin not keeping anything himself. He continued towards Aleppo, which still closed its gates to him, halting before the city. On the way, his army took Buza'a, then captured Manbij. From there, they headed west to besiege the fortress of A'zaz on 15 May. Several days later, while Saladin was resting in one of his captain's tents, an Assassin rushed forward at him and struck at his head with a knife. The cap of his head armour was not penetrated and he managed to grip the Assassin's hand—the dagger only slashing his gambeson—and the assailant was soon killed. Saladin was unnerved at the attempt on his life, which he accused Gumushtugin and the Assassins of plotting, and so increased his efforts in the siege. A'zaz capitulated on 21 June, and Saladin then hurried his forces to Aleppo to punish Gumushtigin. His assaults were again resisted, but he managed to secure not only a truce, but a mutual alliance with Aleppo, in which Gumushtigin and as-Salih were allowed to continue their hold on the city and in return, they recognized Saladin as the sovereign over all of the dominions he conquered. The "emirs" of Mardin and Keyfa, the Muslim allies of Aleppo, also recognised Saladin as the King of Syria. When the treaty was concluded, the younger sister of as-Salih came to Saladin and requested the return of the Fortress of A'zaz; he complied and escorted her back to the gates of Aleppo with numerous presents. Saladin had by now agreed truces with his Zengid rivals and the Kingdom of Jerusalem (the latter occurred in the summer of 1175), but faced a threat from the Isma'ili sect known as the Assassins, led by Rashid ad-Din Sinan. Based in the an-Nusayriyah Mountains, they commanded nine fortresses, all built on high elevations. As soon as he dispatched the bulk of his troops to Egypt, Saladin led his army into the an-Nusayriyah range in August 1176. He retreated the same month, after laying waste to the countryside, but failing to conquer any of the forts. Most Muslim historians claim that Saladin's uncle, the governor of Hama, mediated a peace agreement between him and Sinan. Saladin had his guards supplied with link lights and had chalk and cinders strewed around his tent outside Masyaf—which he was besieging—to detect any footsteps by the Assassins. According to this version, one night Saladin's guards noticed a spark glowing down the hill of Masyaf and then vanishing among the Ayyubid tents. Presently, Saladin awoke to find a figure leaving the tent. He saw that the lamps were displaced and beside his bed laid hot scones of the shape peculiar to the Assassins with a note at the top pinned by a poisoned dagger. The note threatened that he would be killed if he did not withdraw from his assault. Saladin gave a loud cry, exclaiming that Sinan himself was the figure that had left the tent. Another version claims that Saladin hastily withdrew his troops from Masyaf because they were urgently needed to fend off a Crusader force in the vicinity of Mount Lebanon. In reality, Saladin sought to form an alliance with Sinan and his Assassins, consequently depriving the Crusaders of a potent ally against him. Viewing the expulsion of the Crusaders as a mutual benefit and priority, Saladin and Sinan maintained cooperative relations afterwards, the latter dispatching contingents of his forces to bolster Saladin's army in a number of decisive subsequent battlefronts. After leaving the an-Nusayriyah Mountains, Saladin returned to Damascus and had his Syrian soldiers return home. He left Turan Shah in command of Syria and left for Egypt with only his personal followers, reaching Cairo on 22 September. Having been absent roughly two years, he had much to organize and supervise in Egypt, namely fortifying and reconstructing Cairo. The city walls were repaired and their extensions laid out, while the construction of the Cairo Citadel was commenced. The deep Bir Yusuf ("Joseph's Well") was built on Saladin's orders. The chief public work he commissioned outside of Cairo was the large bridge at Giza, which was intended to form an outwork of defense against a potential Moorish invasion. Saladin remained in Cairo supervising its improvements, building colleges such as the Madrasa of the Sword Makers and ordering the internal administration of the country. In November 1177, he set out upon a raid into Palestine; the Crusaders had recently forayed into the territory of Damascus, so Saladin saw the truce as no longer worth preserving. The Christians sent a large portion of their army to besiege the fortress of Harim north of Aleppo, so southern Palestine bore few defenders. Saladin found the situation ripe and marched to Ascalon, which he referred to as the "Bride of Syria". William of Tyre recorded that the Ayyubid army consisted of soldiers, of which 8,000 were elite forces and were black soldiers from Sudan. This army proceeded to raid the countryside, sack Ramla and Lod, and dispersed themselves as far as the Gates of Jerusalem. The Ayyubids allowed Baldwin IV of Jerusalem to enter Ascalon with his Gaza-based Knights Templar without taking any precautions against a sudden attack. Although the Crusader force consisted of only 375 knights, Saladin hesitated to ambush them because of the presence of highly skilled generals. On 25 November, while the greater part of the Ayyubid army was absent, Saladin and his men were surprised near Ramla in the battle of Montgisard. Before they could form up, the Templar force hacked the Ayyubid army down. Initially, Saladin attempted to organize his men into battle order, but as his bodyguards were being killed, he saw that defeat was inevitable and so with a small remnant of his troops mounted a swift camel, riding all the way to the territories of Egypt. Not discouraged by his defeat at Tell Jezer, Saladin was prepared to fight the Crusaders once again. In the spring of 1178, he was encamped under the walls of Homs, and a few skirmishes occurred between his generals and the Crusader army. His forces in Hama won a victory over their enemy and brought the spoils, together with many prisoners of war, to Saladin who ordered the captives to be beheaded for "plundering and laying waste the lands of the Faithful". He spent the rest of the year in Syria without a confrontation with his enemies. Saladin's intelligence services reported to him that the Crusaders were planning a raid into Syria. He ordered one of his generals, Farrukh-Shah, to guard the Damascus frontier with a thousand of his men to watch for an attack, then to retire, avoiding battle, and to light warning beacons on the hills, after which Saladin would march out. In April 1179, the Crusaders led by King Baldwin expected no resistance and waited to launch a surprise attack on Muslim herders grazing their herds and flocks east of the Golan Heights. Baldwin advanced too rashly in pursuit of Farrukh-Shah's force, which was concentrated southeast of Quneitra and was subsequently defeated by the Ayyubids. With this victory, Saladin decided to call in more troops from Egypt; he requested al-Adil to dispatch 1,500 horsemen. In the summer of 1179, King Baldwin had set up an outpost on the road to Damascus and aimed to fortify a passage over the Jordan River, known as Jacob's Ford, that commanded the approach to the Banias plain (the plain was divided by the Muslims and the Christians). Saladin had offered 100,000 gold pieces to Baldwin to abandon the project, which was particularly offensive to the Muslims, but to no avail. He then resolved to destroy the fortress, called Chastellet and manned by the Templars, moving his headquarters to Banias. As the Crusaders hurried down to attack the Muslim forces, they fell into disorder, with the infantry falling behind. Despite early success, they pursued the Muslims far enough to become scattered, and Saladin took advantage by rallying his troops and charged at the Crusaders. The engagement ended in a decisive Ayyubid victory, and many high-ranking knights were captured. Saladin then moved to besiege the fortress, which fell on 30 August 1179. In the spring of 1180, while Saladin was in the area of Safad, anxious to commence a vigorous campaign against the Kingdom of Jerusalem, King Baldwin sent messengers to him with proposals of peace. Because droughts and bad harvests hampered his commissariat, Saladin agreed to a truce. Raymond of Tripoli denounced the truce but was compelled to accept after an Ayyubid raid on his territory in May and upon the appearance of Saladin's naval fleet off the port of Tartus. In June 1180, Saladin hosted a reception for Nur al-Din Muhammad, the Artuqid "emir" of Keyfa, at Geuk Su, in which he presented him and his brother Abu Bakr with gifts, valued at over 100,000 dinars according to Imad al-Din. This was intended to cement an alliance with the Artuqids and to impress other "emirs" in Mesopotamia and Anatolia. Previously, Saladin offered to mediate relations between Nur al-Din and Kilij Arslan II—the Seljuk sultan of Rûm—after the two came into conflict. The latter demanded that Nur al-Din return the lands given to him as a dowry for marrying his daughter when he received reports that she was being abused and used to gain Seljuk territory. Nur al-Din asked Saladin to mediate the issue, but Arslan refused. After Nur al-Din and Saladin met at Geuk Su, the top Seljuk "emir", Ikhtiyar al-Din al-Hasan, confirmed Arslan's submission, after which an agreement was drawn up. Saladin was later enraged when he received a message from Arslan accusing Nur al-Din of more abuses against his daughter. He threatened to attack the city of Malatya, saying, "it is two days march for me and I shall not dismount [my horse] until I am in the city." Alarmed at the threat, the Seljuks pushed for negotiations. Saladin felt that Arslan was correct to care for his daughter, but Nur al-Din had taken refuge with him, and therefore he could not betray his trust. It was finally agreed that Arslan's daughter would be sent away for a year and if Nur al-Din failed to comply, Saladin would move to abandon his support for him. Leaving Farrukh-Shah in charge of Syria, Saladin returned to Cairo at the beginning of 1181. According to Abu Shama, he intended to spend the fast of Ramadan in Egypt and then make the "hajj" pilgrimage to Mecca in the summer. For an unknown reason he apparently changed his plans regarding the pilgrimage and was seen inspecting the Nile River banks in June. He was again embroiled with the Bedouin; he removed two-thirds of their fiefs to use as compensation for the fief-holders at Fayyum. The Bedouin were also accused of trading with the Crusaders and, consequently, their grain was confiscated and they were forced to migrate westward. Later, Ayyubid warships were waged against Bedouin river pirates, who were plundering the shores of Lake Tanis. In the summer of 1181, Saladin's former palace administrator Qara-Qush led a force to arrest Majd al-Din—a former deputy of Turan-Shah in the Yemeni town of Zabid—while he was entertaining Imad ad-Din at his estate in Cairo. Saladin's intimates accused Majd al-Din of misappropriating the revenues of Zabid, but Saladin himself believed there was no evidence to back the allegations. He had Majd al-Din released in return for a payment of 80,000 dinars. In addition, other sums were to be paid to Saladin's brothers al-Adil and Taj al-Muluk Buri. The controversial detainment of Majd al-Din was a part of the larger discontent associated with the aftermath of Turan-Shah's departure from Yemen. Although his deputies continued to send him revenues from the province, centralized authority was lacking and internal quarrel arose between Izz al-Din Uthman of Aden and Hittan of Zabid. Saladin wrote in a letter to al-Adil: "this Yemen is a treasure house ... We conquered it, but up to this day we have had no return and no advantage from it. There have been only innumerable expenses, the sending out of troops ... and expectations which did not produce what was hoped for in the end." Saif al-Din had died earlier in June 1181 and his brother Izz al-Din inherited leadership of Mosul. On 4 December, the crown-prince of the Zengids, as-Salih, died in Aleppo. Prior to his death, he had his chief officers swear an oath of loyalty to Izz al-Din, as he was the only Zengid ruler strong enough to oppose Saladin. Izz al-Din was welcomed in Aleppo, but possessing it and Mosul put too great of a strain on his abilities. He thus, handed Aleppo to his brother Imad al-Din Zangi, in exchange for Sinjar. Saladin offered no opposition to these transactions in order to respect the treaty he previously made with the Zengids. On 11 May 1182, Saladin, along with half of the Egyptian Ayyubid army and numerous non-combatants, left Cairo for Syria. On the evening before he departed, he sat with his companions and the tutor of one of his sons quoted a line of poetry: "enjoy the scent of the ox-eye plant of Najd, for after this evening it will come no more". Saladin took this as an evil omen and he never saw Egypt again. Knowing that Crusader forces were massed upon the frontier to intercept him, he took the desert route across the Sinai Peninsula to Ailah at the head of the Gulf of Aqaba. Meeting no opposition, Saladin ravaged the countryside of Montreal, whilst Baldwin's forces watched on, refusing to intervene. He arrived in Damascus in June to learn that Farrukh-Shah had attacked the Galilee, sacking Daburiyya and capturing Habis Jaldek, a fortress of great importance to the Crusaders. In July, Saladin dispatched Farrukh-Shah to attack Kawkab al-Hawa. Later, in August, the Ayyubids launched a naval and ground assault to capture Beirut; Saladin led his army in the Bekaa Valley. The assault was leaning towards failure and Saladin abandoned the operation to focus on issues in Mesopotamia. Kukbary (Muzaffar ad-Din Gökböri), the "emir" of Harran, invited Saladin to occupy the Jazira region, making up northern Mesopotamia. He complied and the truce between him and the Zengids officially ended in September 1182. Prior to his march to Jazira, tensions had grown between the Zengid rulers of the region, primarily concerning their unwillingness to pay deference to Mosul. Before he crossed the Euphrates, Saladin besieged Aleppo for three days, signaling that the truce was over. Once he reached Bira, near the river, he was joined by Kukbary and Nur al-Din of Hisn Kayfa and the combined forces captured the cities of Jazira, one after the other. First, Edessa fell, followed by Saruj, then Raqqa, Qirqesiya and Nusaybin. Raqqa was an important crossing point and held by Qutb al-Din Inal, who had lost Manbij to Saladin in 1176. Upon seeing the large size of Saladin's army, he made little effort to resist and surrendered on the condition that he would retain his property. Saladin promptly impressed the inhabitants of the town by publishing a decree that ordered a number of taxes to be canceled and erased all mention of them from treasury records, stating "the most miserable rulers are those whose purses are fat and their people thin". From Raqqa, he moved to conquer al-Fudain, al-Husain, Maksim, Durain, 'Araban, and Khabur—all of which swore allegiance to him. Saladin proceeded to take Nusaybin which offered no resistance. A medium-sized town, Nusaybin was not of great importance, but it was located in a strategic position between Mardin and Mosul and within easy reach of Diyarbakir. In the midst of these victories, Saladin received word that the Crusaders were raiding the villages of Damascus. He replied "Let them... whilst they knock down villages, we are taking cities; when we come back, we shall have all the more strength to fight them." Meanwhile, in Aleppo, the "emir" of the city Zangi raided Saladin's cities to the north and east, such as Balis, Manbij, Saruj, Buza'a, al-Karzain. He also destroyed his own citadel at A'zaz to prevent it from being used by the Ayyubids if they were to conquer it. Saladin turned his attention from Mosul to Aleppo, sending his brother Taj al-Muluk Buri to capture Tell Khalid, 130 km northeast of the city. A siege was set, but the governor of Tell Khalid surrendered upon the arrival of Saladin himself on 17 May before a siege could take place. According to Imad ad-Din, after Tell Khalid, Saladin took a detour northwards to Ain Tab, but he gained possession of it when his army turned towards it, allowing to quickly move backward another c. 100 km towards Aleppo. On 21 May, he camped outside the city, positioning himself east of the Citadel of Aleppo, while his forces encircles the suburb of Banaqusa to the northeast and Bab Janan to the west. He stationed his men dangerously close to the city, hoping for an early success. Zangi did not offer long resistance. He was unpopular with his subjects and wished to return to his Sinjar, the city he governed previously. An exchange was negotiated where Zangi would hand over Aleppo to Saladin in return for the restoration of his control of Sinjar, Nusaybin, and Raqqa. Zangi would hold these territories as Saladin's vassals on terms of military service. On 12 June, Aleppo was formally placed in Ayyubid hands. The people of Aleppo had not known about these negotiations and were taken by surprise when Saladin's standard was hoisted over the citadel. Two "emir"s, including an old friend of Saladin, Izz al-Din Jurduk, welcomed and pledged their service to him. Saladin replaced the Hanafi courts with Shafi'i administration, despite a promise he would not interfere in the religious leadership of the city. Although he was short of money, Saladin also allowed the departing Zangi to take all the stores of the citadel that he could travel with and to sell the remainder—which Saladin purchased himself. In spite of his earlier hesitation to go through with the exchange, he had no doubts about his success, stating that Aleppo was "the key to the lands" and "this city is the eye of Syria and the citadel is its pupil". For Saladin, the capture of the city marked the end of over eight years of waiting since he told Farrukh-Shah that "we have only to do the milking and Aleppo will be ours". After spending one night in Aleppo's citadel, Saladin marched to Harim, near the Crusader-held Antioch. The city was held by Surhak, a "minor "mamluk"". Saladin offered him the city of Busra and property in Damascus in exchange for Harim, but when Surhak asked for more, his own garrison in Harim forced him out. He was arrested by Saladin's deputy Taqi al-Din on allegations that he was planning to cede Harim to Bohemond III of Antioch. When Saladin received its surrender, he proceeded to arrange the defense of Harim from the Crusaders. He reported to the caliph and his own subordinates in Yemen and Baalbek that was going to attack the Armenians. Before he could move, however, there were a number of administrative details to be settled. Saladin agreed to a truce with Bohemond in return for Muslim prisoners being held by him and then he gave A'zaz to Alam ad-Din Suleiman and Aleppo to Saif al-Din al-Yazkuj—the former was an "emir" of Aleppo who joined Saladin and the latter was a former "mamluk" of Shirkuh who helped rescue him from the assassination attempt at A'zaz. As Saladin approached Mosul, he faced the issue of taking over a large city and justifying the action. The Zengids of Mosul appealed to an-Nasir, the Abbasid caliph at Baghdad whose vizier favored them. An-Nasir sent Badr al-Badr (a high-ranking religious figure) to mediate between the two sides. Saladin arrived at the city on 10 November 1182. Izz al-Din would not accept his terms because he considered them disingenuous and extensive, and Saladin immediately laid siege to the heavily fortified city. After several minor skirmishes and a stalemate in the siege that was initiated by the caliph, Saladin intended to find a way to withdraw without damage to his reputation while still keeping up some military pressure. He decided to attack Sinjar, which was held by Izz al-Din's brother Sharaf al-Din. It fell after a 15-day siege on 30 December. Saladin's soldiers broke their discipline, plundering the city; Saladin only managed to protect the governor and his officers by sending them to Mosul. After establishing a garrison at Sinjar, he awaited a coalition assembled by Izz al-Din consisting of his forces, those from Aleppo, Mardin, and Armenia. Saladin and his army met the coalition at Harran in February 1183, but on hearing of his approach, the latter sent messengers to Saladin asking for peace. Each force returned to their cities and al-Fadil wrote: "They [Izz al-Din's coalition] advanced like men, like women they vanished." On 2 March, al-Adil from Egypt wrote to Saladin that the Crusaders had struck the "heart of Islam". Raynald de Châtillon had sent ships to the Gulf of Aqaba to raid towns and villages off the coast of the Red Sea. It was not an attempt to extend the Crusader influence into that sea or to capture its trade routes, but merely a piratical move. Nonetheless, Imad al-Din writes the raid was alarming to the Muslims because they were not accustomed to attacks on that sea, and Ibn al-Athir adds that the inhabitants had no experience with the Crusaders either as fighters or traders. Ibn Jubair was told that sixteen Muslim ships were burnt by the Crusaders, who then captured a pilgrim ship and caravan at Aidab. He also reported that they intended to attack Medina and remove Muhammad's body. Al-Maqrizi added to the rumor by claiming Muhammad's tomb was going to be relocated to Crusader territory so Muslims would make pilgrimages there. Al-Adil had his warships moved from Fustat and Alexandria to the Red Sea under the command of an Armenian mercenary Lu'lu. They broke the Crusader blockade, destroyed most of their ships, and pursued and captured those who anchored and fled into the desert. The surviving Crusaders, numbered at 170, were ordered to be killed by Saladin in various Muslim cities. From the point of view of Saladin, in terms of territory, the war against Mosul was going well, but he still failed to achieve his objectives and his army was shrinking; Taqi al-Din took his men back to Hama, while Nasir al-Din Muhammad and his forces had left. This encouraged Izz al-Din and his allies to take the offensive. The previous coalition regrouped at Harzam some 140 km from Harran. In early April, without waiting for Nasir al-Din, Saladin and Taqi al-Din commenced their advance against the coalition, marching eastward to Ras al-Ein unhindered. By late April, after three days of "actual fighting", according to Saladin, the Ayyubids had captured Amid. He handed the city to Nur al-Din Muhammad together with its stores, which consisted of 80,000 candles, a tower full of arrowheads, and 1,040,000 books. In return for a diploma granting him the city, Nur al-Din swore allegiance to Saladin, promising to follow him in every expedition in the war against the Crusaders, and repairing damage done to the city. The fall of Amid, in addition to territory, convinced Il-Ghazi of Mardin to enter the service of Saladin, weakening Izz al-Din's coalition. Saladin attempted to gain the Caliph an-Nasir's support against Izz al-Din by sending him a letter requesting a document that would give him legal justification for taking over Mosul and its territories. Saladin aimed to persuade the caliph claiming that while he conquered Egypt and Yemen under the flag of the Abbasids, the Zengids of Mosul openly supported the Seljuks (rivals of the caliphate) and only came to the caliph when in need. He also accused Izz al-Din's forces of disrupting the Muslim "Holy War" against the Crusaders, stating "they are not content not to fight, but they prevent those who can". Saladin defended his own conduct claiming that he had come to Syria to fight the Crusaders, end the heresy of the Assassins, and stop the wrong-doing of the Muslims. He also promised that if Mosul was given to him, it would lead to the capture of Jerusalem, Constantinople, Georgia, and the lands of the Almohads in the Maghreb, "until the word of God is supreme and the Abbasid caliphate has wiped the world clean, turning the churches into mosques". Saladin stressed that all this would happen by the will of God, and instead of asking for financial or military support from the caliph, he would capture and give the caliph the territories of Tikrit, Daquq, Khuzestan, Kish Island, and Oman. On 29 September 1182, Saladin crossed the Jordan River to attack Beisan, which was found to be empty. The next day his forces sacked and burned the town and moved westwards. They intercepted Crusader reinforcements from Karak and Shaubak along the Nablus road and took a number of prisoners. Meanwhile, the main Crusader force under Guy of Lusignan moved from Sepphoris to al-Fula. Saladin sent out 500 skirmishers to harass their forces, and he himself marched to Ain Jalut. When the Crusader force—reckoned to be the largest the kingdom ever produced from its own resources, but still outmatched by the Muslims—advanced, the Ayyubids unexpectedly moved down the stream of Ain Jalut. After a few Ayyubid raids—including attacks on Zir'in, Forbelet, and Mount Tabor—the Crusaders still were not tempted to attack their main force, and Saladin led his men back across the river once provisions and supplies ran low. Crusader attacks provoked further responses by Saladin. Raynald of Châtillon, in particular, harassed Muslim trading and pilgrimage routes with a fleet on the Red Sea, a water route that Saladin needed to keep open. In response, Saladin built a fleet of 30 galleys to attack Beirut in 1182. Raynald threatened to attack the holy cities of Mecca and Medina. In retaliation, Saladin twice besieged Kerak, Raynald's fortress in Oultrejordain, in 1183 and 1184. Raynald responded by looting a caravan of pilgrims on the Hajj in 1185. According to the later 13th-century "Old French Continuation of William of Tyre", Raynald captured Saladin's sister in a raid on a caravan; this claim is not attested in contemporary sources, Muslim or Frankish, however, instead stating that Raynald had attacked a preceding caravan, and Saladin set guards to ensure the safety of his sister and her son, who came to no harm. Following the failure of his Kerak sieges, Saladin temporarily turned his attention back to another long-term project and resumed attacks on the territory of ʻIzz ad-Dīn (Masʻūd ibn Mawdūd ibn Zangi), around Mosul, which he had begun with some success in 1182. However, since then, Masʻūd had allied himself with the powerful governor of Azerbaijan and Jibal, who in 1185 began moving his troops across the Zagros Mountains, causing Saladin to hesitate in his attacks. The defenders of Mosul, when they became aware that help was on the way, increased their efforts, and Saladin subsequently fell ill, so in March 1186 a peace treaty was signed. In July 1187, Saladin captured most of the Kingdom of Jerusalem. On 4 July 1187, at the Battle of Hattin, he faced the combined forces of Guy of Lusignan, King Consort of Jerusalem, and Raymond III of Tripoli. In this battle alone the Crusader force was largely annihilated by Saladin's determined army. It was a major disaster for the Crusaders and a turning point in the history of the Crusades. Saladin captured Raynald and was personally responsible for his execution in retaliation for his attacks against Muslim caravans. The members of these caravans had, in vain, besought his mercy by reciting the truce between the Muslims and the Crusaders, but Raynald ignored this and insulted the Islamic prophet, Muhammad, before murdering and torturing a number of them. Upon hearing this, Saladin swore an oath to personally execute Raynald. Guy of Lusignan was also captured. Seeing the execution of Raynald, he feared he would be next. However, his life was spared by Saladin, who said of Raynald, "[i]t is not the wont of kings, to kill kings; but that man had transgressed all bounds, and therefore did I treat him thus." Saladin had captured almost every Crusader city. Saladin preferred to take Jerusalem without bloodshed and offered generous terms, but those inside refused to leave their holy city, vowing to destroy it in a fight to the death rather than see it handed over peacefully. Jerusalem capitulated to his forces on Friday, 2 October 1187, after a siege. When the siege had started, Saladin was unwilling to promise terms of quarter to the Frankish inhabitants of Jerusalem. Balian of Ibelin threatened to kill every Muslim hostage, estimated at 5,000, and to destroy Islam's holy shrines of the Dome of the Rock and the al-Aqsa Mosque if such quarter were not provided. Saladin consulted his council and the terms were accepted. The agreement was read out through the streets of Jerusalem so that everyone might within forty days provide for himself and pay to Saladin the agreed tribute for his freedom. An unusually low ransom for the times (around US$50 today) was to be paid for each Frank in the city, whether man, woman, or child, but Saladin, against the wishes of his treasurers, allowed many families who could not afford the ransom to leave. Patriarch Heraclius of Jerusalem organised and contributed to a collection that paid the ransoms for about 18,000 of the poorer citizens, leaving another 15,000 to be enslaved. Saladin's brother al-Adil "asked Saladin for a thousand of them for his own use and then released them on the spot." Most of the foot soldiers were sold into slavery. Upon the capture of Jerusalem, Saladin summoned the Jews and permitted them to resettle in the city. In particular, the residents of Ashkelon, a large Jewish settlement, responded to his request. The subject ordered the churches repurposed as horse stables and the church towers destroyed. Tyre, on the coast of modern-day Lebanon, was the last major Crusader city that was not captured by Muslim forces. Strategically, it would have made more sense for Saladin to capture Tyre before Jerusalem; Saladin, however, chose to pursue Jerusalem first because of the importance of the city to Islam. Tyre was commanded by Conrad of Montferrat, who strengthened its defences and withstood two sieges by Saladin. In 1188, at Tortosa, Saladin released Guy of Lusignan and returned him to his wife, Queen Sibylla of Jerusalem. They went first to Tripoli, then to Antioch. In 1189, they sought to reclaim Tyre for their kingdom but were refused admission by Conrad, who did not recognize Guy as king. Guy then set about besieging Acre. Saladin was on friendly terms with Queen Tamar of Georgia. Saladin's biographer Bahā' ad-Dīn ibn Šaddād reports that, after Saladin's conquest of Jerusalem, the Georgian Queen sent envoys to the sultan to request the return of confiscated possessions of the Georgian monasteries in Jerusalem. Saladin's response is not recorded, but the queen's efforts seem to have been successful as Jacques de Vitry, the Bishop of Acre, reports the Georgians were, in contrast to the other Christian pilgrims, allowed a free passage into the city with their banners unfurled. Ibn Šaddād furthermore claims that Queen Tamar outbid the Byzantine emperor in her efforts to obtain the relics of the True Cross, offering 200,000 gold pieces to Saladin who had taken the relics as booty at the battle of Hattin, but to no avail. Hattin and the fall of Jerusalem prompted the Third Crusade (1189–1192), financed in England by a special "Saladin tithe". Richard the Lionheart, King of England led Guy's siege of Acre, conquered the city and executed 3,000 Muslim prisoners, including women and children. Bahā' ad-Dīn wrote: The armies of Saladin engaged in combat with the army of King Richard at the Battle of Arsuf on 7 September 1191, at which Saladin's forces suffered heavy losses and were forced to withdraw. After the battle of Arsuf, Richard occupied Jaffa, restoring the city's fortifications. Meanwhile, Saladin moved south, where he dismantled the fortifications of Ascalon to prevent this strategically important city, which lay at the junction between Egypt and Palestine, from falling into Crusader hands. In October 1191, Richard began restoring the inland castles on the coastal plain beyond Jaffa in preparation for an advance on Jerusalem. During this period, Richard and Saladin passed envoys back and forth, negotiating the possibility of a truce. Richard proposed that his sister, Joan of England, Queen of Sicily, should marry Saladin's brother and that Jerusalem could be their wedding gift. However, Saladin rejected this idea when Richard insisted that Saladin's brother convert to Christianity. Richard suggested that his niece Eleanor, Fair Maid of Brittany be the bride instead, an idea that Saladin also rejected. In January 1192, Richard's army occupied Beit Nuba, just twelve miles from Jerusalem, but withdrew without attacking the Holy City. Instead, Richard advanced south on Ascalon, where he restored the fortifications. In July 1192, Saladin tried to threaten Richard's command of the coast by attacking Jaffa. The city was besieged, and Saladin very nearly captured it; however, Richard arrived a few days later and defeated Saladin's army in a battle outside the city. The Battle of Jaffa (1192) proved to be the last military engagement of the Third Crusade. After Richard reoccupied Jaffa and restored its fortifications, he and Saladin again discussed terms. At last Richard agreed to demolish the fortifications of Ascalon, while Saladin agreed to recognize Crusader control of the Palestinian coast from Tyre to Jaffa. The Christians would be allowed to travel as unarmed pilgrims to Jerusalem, and Saladin's kingdom would be at peace with the Crusader states for the following three years. Saladin died of a fever on 4 March 1193, at Damascus, not long after King Richard's departure. In Saladin’s possession at the time of his death were one piece of gold and forty pieces of silver. He had given away his great wealth to his poor subjects, leaving nothing to pay for his funeral. He was buried in a mausoleum in the garden outside the Umayyad Mosque in Damascus, Syria. Originally the tomb was part of a complex which also included a school, Madrassah al-Aziziah, of which little remains except a few columns and an internal arch. Seven centuries later, Emperor Wilhelm II of Germany donated a new marble sarcophagus to the mausoleum. However, the original sarcophagus was not replaced; instead, the mausoleum, which is open to visitors, now has two sarcophagi: the marble one placed on the side and the original wooden one, which covers Saladin's tomb. (Muslims are buried in a simple shroud, so if there are any sarcophagi present, they are usually used for covering the top of the Islamic burials.) Imad ad-Din al-Isfahani compiled a list of Saladin's sons along with their dates of birth, according to information provided by Saladin late in his reign. They were: The sons who were full brothers were: The sons listed by Imad number fifteen, but elsewhere he writes that Saladin was survived by seventeen sons and one daughter. According to Abu Hamah, Imad missed two sons who were born to slave-women: Imad al-Din Shadhi and Nusrat al-Din Marwan. As for Saladin's daughter, she was Mu'nisah Khatun; she married her cousin al-Kamil Muhammad ibn Adil. Saladin also had other children who died before him, such as al-Mansur Hasan and Ahmad. Al-Zahir Dawud, whom Imad listed eighth, is recorded as being his twelfth son in a letter written by Saladin's minister. Not much is known of Saladin's wives or slave-women. He married Ismat al-Din Khatun, the widow of Nur al-Din Zengi, in 1176. She did not have children. One of his wives, Shamsah, is buried with her son al-Aziz in the tomb of al-Shafi'i. Saladin eventually achieved a great reputation in Europe as a chivalrous knight, due to his fierce struggle against the crusaders and his generosity. In "The Divine Comedy" he is mentioned as one of the virtuous non-Christians in limbo. Although Saladin faded into history after the Middle Ages, he appears in a sympathetic light in Gotthold Ephraim Lessing's play "Nathan the Wise" (1779) and in Sir Walter Scott's novel "The Talisman" (1825). The modern view of Saladin originates mainly from these texts. According to Jonathan Riley-Smith, Scott's portrayal of Saladin was that of a "modern [19th-century] liberal European gentlemen, beside whom medieval Westerners would always have made a poor showing". Despite the Crusaders' slaughter when they originally conquered Jerusalem in 1099, Saladin granted amnesty and free passage to all common Catholics and even to the defeated Christian army, as long as they were able to pay the aforementioned ransom (the Greek Orthodox Christians were treated even better, because they often opposed the western Crusaders). Notwithstanding the differences in beliefs, the Muslim Saladin was respected by Christian lords, Richard especially. Richard once praised Saladin as a great prince, saying that he was without doubt the greatest and most powerful leader in the Islamic world. Saladin in turn stated that there was not a more honorable Christian lord than Richard. After the treaty, Saladin and Richard sent each other many gifts as tokens of respect but never met face to face. In April 1191, a Frankish woman's three-month-old baby had been stolen from her camp and sold on the market. The Franks urged her to approach Saladin herself with her grievance. According to Bahā' al-Dīn, Saladin used his own money to buy the child back: In 1898, German Emperor Wilhelm II visited Saladin's tomb to pay his respects. The visit, coupled with anti-imperialist sentiments, led nationalist Arabs to reinvent the image of Saladin and portray him as a hero of the struggle against the West. The image of Saladin they used was the romantic one created by Walter Scott and other Europeans in the West at the time. It replaced Saladin's reputation as a figure who had been largely forgotten in the Muslim world, eclipsed by more successful figures, such as Baybars of Egypt. Modern Arab states have sought to commemorate Saladin through various measures, often based on the image created of him in the 19th-century west. A governorate centered around Tikrit and Samarra in modern-day Iraq, Saladin Governorate, is named after him, as is Salahaddin University in Erbil, the largest city of Iraqi Kurdistan. A suburban community of Erbil, Masif Salahaddin, is also named after him. Few structures associated with Saladin survive within modern cities. Saladin first fortified the Citadel of Cairo (1175–1183), which had been a domed pleasure pavilion with a fine view in more peaceful times. In Syria, even the smallest city is centred on a defensible citadel, and Saladin introduced this essential feature to Egypt. Although the Ayyubid dynasty that he founded would only outlive him by 57 years, the legacy of Saladin within the Arab World continues to this day. With the rise of Arab nationalism in the 20th Century, particularly with regard to the Arab–Israeli conflict, Saladin's heroism and leadership gained a new significance. Saladin's recapture of Palestine from the European Crusaders is considered an inspiration for modern-day Arabs' opposition to Zionism. Moreover, the glory and comparative unity of the Arab World under Saladin was seen as the perfect symbol for the new unity sought by Arab nationalists, such as Gamal Abdel Nasser. For this reason, the Eagle of Saladin became the symbol of revolutionary Egypt, and was subsequently adopted by several other Arab states (the United Arab Republic, Iraq, Libya, the State of Palestine, and Yemen). Among Egyptian Shias, Saladin is dubbed as "Kharab al-Din", the destroyer of religion—a derisive play on the name "Saladin."
https://en.wikipedia.org/wiki?curid=26983
Sophocles Sophocles (; "Sophoklēs", ; 497/6 – winter 406/5 BC) is one of three ancient Greek tragedians whose plays have survived. His first plays were written later than or contemporary with those of Aeschylus, and earlier than or contemporary with those of Euripides. Sophocles wrote over 120 plays during the course of his life, but only seven have survived in a complete form: "Ajax", "Antigone", "Women of Trachis", "Oedipus Rex", "Electra", "Philoctetes" and "Oedipus at Colonus". For almost 50 years, Sophocles was the most celebrated playwright in the dramatic competitions of the city-state of Athens that took place during the religious festivals of the Lenaea and the Dionysia. He competed in 30 competitions, won 24, and was never judged lower than second place. Aeschylus won 13 competitions, and was sometimes defeated by Sophocles, while Euripides won four competitions. The most famous tragedies of Sophocles feature Oedipus and also Antigone: they are generally known as the Theban plays, although each play was actually a part of a different tetralogy, the other members of which are now lost. Sophocles influenced the development of drama, most importantly by adding a third actor, thereby reducing the importance of the chorus in the presentation of the plot. He also developed his characters to a greater extent than earlier playwrights such as Aeschylus. Sophocles, the son of Sophilus, was a wealthy member of the rural "deme" (small community) of Hippeios Colonus in Attica, which was to become a setting for one of his plays, and he was probably born there. Sophocles was born a few years before the Battle of Marathon in 490 BC: the exact year is unclear, although 497/6 is the most likely. Sophocles was born into a wealthy family (his father was an armour manufacturer) and was highly educated. Sophocles' first artistic triumph was in 468 BC, when he took first prize in the Dionysia theatre competition over the reigning master of Athenian drama, Aeschylus. According to Plutarch, the victory came under unusual circumstances. Instead of following the usual custom of choosing judges by lot, the archon asked Cimon and the other "strategoi" present to decide the victor of the contest. Plutarch further contends that following this loss Aeschylus soon left for Sicily. Although Plutarch says that this was Sophocles' first production, it is now thought that his first production was probably in 470 BC. "Triptolemus" was probably one of the plays that Sophocles presented at this festival. In 480 BC Sophocles was chosen to lead the paean (a choral chant to a god), celebrating the Greek victory over the Persians at the Battle of Salamis. Early in his career, the politician Cimon might have been one of his patrons, although if he was, there was no ill will borne by Pericles, Cimon's rival, when Cimon was ostracized in 461 BC. In 443/2 he served as one of the "Hellenotamiai", or treasurers of Athena, helping to manage the finances of the city during the political ascendancy of Pericles. According to the "Vita Sophoclis", in 441 BC he was elected one of the ten generals, executive officials at Athens, as a junior colleague of Pericles, and he served in the Athenian campaign against Samos; he was supposed to have been elected to this position as the result of his production of "Antigone". In 420 BC, he welcomed and set up an altar for the image of Asclepius at his house, when the deity was introduced to Athens. For this, he was given the posthumous epithet "Dexion" (receiver) by the Athenians. He was also elected, in 413 BC, one of the commissioners ("probouloi") who responded to the catastrophic destruction of the Athenian expeditionary force in Sicily during the Peloponnesian War. Sophocles died at the age of ninety or ninety-one in the winter of 406/5 BC, having seen within his lifetime both the Greek triumph in the Persian Wars and the bloodletting of the Peloponnesian War. As with many famous men in classical antiquity, his death inspired a number of apocryphal stories. The most famous is the suggestion that he died from the strain of trying to recite a long sentence from his "Antigone" without pausing to take a breath. Another account suggests he choked while eating grapes at the Anthesteria festival in Athens. A third holds that he died of happiness after winning his final victory at the City Dionysia. A few months later, a comic poet, in a play titled "The Muses", wrote this eulogy: "Blessed is Sophocles, who had a long life, was a man both happy and talented, and the writer of many good tragedies; and he ended his life well without suffering any misfortune." According to some accounts, however, his own sons tried to have him declared incompetent near the end of his life; he is said to have refuted their charge in court by reading from his as yet unproduced "Oedipus at Colonus". One of his sons, Iophon, and a grandson, also called Sophocles, also became playwrights. An ancient source, Athenaeus’s work "Sophists at Dinner", contains references to Sophocles' homosexuality or bisexuality. In that work, a character named Myrtilus, in a lengthy banquet speech claims that Ion of Chios writes in his book "Encounters", that Sophocles loved boys as much as Euripides loved women. Myrtilus also repeats an anecdote reportedly told by Ion of Chios that involves Sophocles flirting with a serving boy at a symposium. Myrtilus also claims that in a work by Hieronymus of Rhodes entitled "Historical Notes" it is said that Sophocles once lured a boy outside to have sex, and afterwards the boy left with Sophocles' cape, while the boy's own cape was left with Sophocles. Among Sophocles' earliest innovations was the addition of a third actor, which further reduced the role of the chorus and created greater opportunity for character development and conflict between characters. Aeschylus, who dominated Athenian playwriting during Sophocles' early career, followed suit and adopted the third character into his own work towards the end of his life. Aristotle credits Sophocles with the introduction of "skenographia", or scenery-painting. It was not until after the death of the old master Aeschylus in 456 BC that Sophocles became the pre-eminent playwright in Athens. Thereafter, Sophocles emerged victorious in dramatic competitions at 18 Dionysia and 6 Lenaia festivals. In addition to innovations in dramatic structure, Sophocles' work is also known for its deeper development of characters than earlier playwrights. His reputation was such that foreign rulers invited him to attend their courts, although unlike Aeschylus who died in Sicily, or Euripides who spent time in Macedon, Sophocles never accepted any of these invitations. Aristotle used Sophocles' "Oedipus Rex" in his "Poetics" (c. 335 BC) as an example of the highest achievement in tragedy, which suggests the high esteem in which his work was held by later Greeks. Only two of the seven surviving plays can be dated securely: "Philoctetes" (409 BC) and "Oedipus at Colonus" (401 BC, staged after Sophocles' death by his grandson). Of the others, "Electra" shows stylistic similarities to these two plays, which suggests that it was probably written in the latter part of his career. "Ajax", "Antigone" and "The Trachiniae" are generally thought to be among his early works, again based on stylistic elements, with "Oedipus Rex" coming in Sophocles' middle period. Most of Sophocles' plays show an undercurrent of early fatalism and the beginnings of Socratic logic as a mainstay for the long tradition of Greek tragedy. The Theban plays consist of three plays: "Oedipus Rex" (also called "Oedipus Tyrannus" or "Oedipus the King"), "Oedipus at Colonus" and "Antigone". All three plays concern the fate of Thebes during and after the reign of King Oedipus. They have often been published under a single cover. Sophocles, however, wrote the three plays for separate festival competitions, many years apart. Not only are the Theban plays not a true trilogy (three plays presented as a continuous narrative) but they are not even an intentional series and contain some inconsistencies among them. He also wrote other plays having to do with Thebes, such as the "Epigoni", of which only fragments have survived. Each of the plays relates to the tale of the mythological Oedipus, who killed his father and married his mother without knowledge that they were his parents. His family is fated to be doomed for three generations. In "Oedipus Rex", Oedipus is the protagonist. Oedipus' infanticide is planned by his parents, Laius and Jocasta, to avert him from fulfilling a prophecy; in truth, the servant entrusted with the infanticide passes the infant on through a series of intermediaries to a childless couple, who adopt him not knowing his history. Oedipus eventually learns of the Delphic Oracle's prophecy of him, that he would kill his father and marry his mother; Oedipus attempts to flee his fate without harming those he knows as his parents (at this point, he does not know that he is adopted). Oedipus meets a man at a crossroads accompanied by servants; Oedipus and the man fight, and Oedipus kills the man (who was his father, Laius, although neither knew at the time). He becomes the ruler of Thebes after solving the riddle of the Sphinx and in the process, marries the widowed queen, his mother Jocasta. Thus the stage is set for horror. When the truth comes out, following from another true but confusing prophecy from Delphi, Jocasta commits suicide, Oedipus blinds himself and leaves Thebes. At the end of the play, order is restored. This restoration is seen when Creon, brother of Jocasta, becomes king, and also when Oedipus, before going off to exile, asks Creon to take care of his children. Oedipus's children will always bear the weight of shame and humiliation because of their father's actions. In "Oedipus at Colonus", the banished Oedipus and his daughter Antigone arrive at the town of Colonus where they encounter Theseus, King of Athens. Oedipus dies and strife begins between his sons Polyneices and Eteocles. In "Antigone", the protagonist is Oedipus' daughter, Antigone. She is faced with the choice of allowing her brother Polyneices' body to remain unburied, outside the city walls, exposed to the ravages of wild animals, or to bury him and face death. The king of the land, Creon, has forbidden the burial of Polyneices for he was a traitor to the city. Antigone decides to bury his body and face the consequences of her actions. Creon sentences her to death. Eventually, Creon is convinced to free Antigone from her punishment, but his decision comes too late and Antigone commits suicide. Her suicide triggers the suicide of two others close to King Creon: his son, Haemon, who was to wed Antigone, and his wife, Eurydice, who commits suicide after losing her only surviving son. The plays were written across thirty-six years of Sophocles' career and were not composed in chronological order, but instead were written in the order "Antigone", "Oedipus Rex", and "Oedipus at Colonus". Nor were they composed as a "trilogy" – a group of plays to be performed together, but are the remaining parts of three different groups of plays. As a result, there are some inconsistencies: notably, Creon is the undisputed king at the end of "Oedipus Rex" and, in consultation with Apollo, single-handedly makes the decision to expel Oedipus from Thebes. Creon is also instructed to look after Oedipus' daughters Antigone and Ismene at the end of "Oedipus Rex". By contrast, in the other plays there is some struggle with Oedipus' sons Eteocles and Polynices in regard to the succession. In "Oedipus at Colonus", Sophocles attempts to work these inconsistencies into a coherent whole: Ismene explains that, in light of their tainted family lineage, her brothers were at first willing to cede the throne to Creon. Nevertheless, they eventually decided to take charge of the monarchy, with each brother disputing the other's right to succeed. In addition to being in a clearly more powerful position in "Oedipus at Colonus", Eteocles and Polynices are also culpable: they consent (l. 429, Theodoridis, tr.) to their father's going to exile, which is one of his bitterest charges against them. In addition to the three Theban plays, there are four surviving plays by Sophocles: "Ajax", "Women of Trachis", "Electra", and "Philoctetes", the last of which won first prize in 409 BC in which it competed. "Ajax" focuses on the proud hero of the Trojan War, Telamonian Ajax, who is driven to treachery and eventually suicide. Ajax becomes gravely upset when Achilles’ armor is presented to Odysseus instead of himself. Despite their enmity toward him, Odysseus persuades the kings Menelaus and Agamemnon to grant Ajax a proper burial. "The Women of Trachis" (named for the Trachinian women who make up the chorus) dramatizes Deianeira's accidentally killing Heracles after he had completed his famous twelve labors. Tricked into thinking it is a love charm, Deianeira applies poison to an article of Heracles' clothing; this poisoned robe causes Heracles to die an excruciating death. Upon learning the truth, Deianeira commits suicide. "Electra" corresponds roughly to the plot of Aeschylus' "Libation Bearers". It details how Electra and Orestes avenge their father Agamemnon's murder by Clytemnestra and Aegisthus. "Philoctetes" retells the story of Philoctetes, an archer who had been abandoned on Lemnos by the rest of the Greek fleet while on the way to Troy. After learning that they cannot win the Trojan War without Philoctetes' bow, the Greeks send Odysseus and Neoptolemus to retrieve him; due to the Greeks' earlier treachery, however, Philoctetes refuses to rejoin the army. It is only Heracles' deus ex machina appearance that persuades Philoctetes to go to Troy. Although the list of over 120 titles of plays associated with Sophocles are known and presented below, little is known of the precise dating of most of them. "Philoctetes" is known to have been written in 409 BC, and "Oedipus at Colonus" is known to have only been performed in 401 BC, posthumously, at the initiation of Sophocles' grandson. The convention on writing plays for the Greek festivals was to submit them in tetralogies of three tragedies along with one satyr play. Along with the unknown dating of the vast majority of over 120 play titles, it is also largely unknown how the plays were grouped. It is, however, known that the three plays referred to in the modern era as the "Theban plays" were never performed together in Sophocles' own lifetime, and are therefore not a trilogy (which they are sometimes erroneously seen as). Fragments of "Ichneutae" ("Tracking Satyrs") were discovered in Egypt in 1907. These amount to about half of the play, making it the best preserved satyr play after Euripides' "Cyclops", which survives in its entirety. Fragments of the "Epigoni" were discovered in April 2005 by classicists at Oxford University with the help of infrared technology previously used for satellite imaging. The tragedy tells the story of the second siege of Thebes. A number of other Sophoclean works have survived only in fragments, including: There is a passage of Plutarch's tract "De Profectibus in Virtute 7 " in which Sophocles discusses his own growth as a writer. A likely source of this material for Plutarch was the "Epidemiae" of Ion of Chios, a book that recorded many conversations of Sophocles. This book is a likely candidate to have contained Sophocles' discourse on his own development because Ion was a friend of Sophocles, and the book is known to have been used by Plutarch. Though some interpretations of Plutarch's words suggest that Sophocles says that he imitated Aeschylus, the translation does not fit grammatically, nor does the interpretation that Sophocles said that he was making fun of Aeschylus' works. C. M. Bowra argues for the following translation of the line: "After practising to the full the bigness of Aeschylus, then the painful ingenuity of my own invention, now in the third stage I am changing to the kind of diction which is most expressive of character and best." Here Sophocles says that he has completed a stage of Aeschylus' work, meaning that he went through a phase of imitating Aeschylus' style but is finished with that. Sophocles' opinion of Aeschylus was mixed. He certainly respected him enough to imitate his work early on in his career, but he had reservations about Aeschylus' style, and thus did not keep his imitation up. Sophocles' first stage, in which he imitated Aeschylus, is marked by "Aeschylean pomp in the language". Sophocles' second stage was entirely his own. He introduced new ways of evoking feeling out of an audience, as in his "Ajax", when Ajax is mocked by Athene, then the stage is emptied so that he may commit suicide alone. Sophocles mentions a third stage, distinct from the other two, in his discussion of his development. The third stage pays more heed to diction. His characters spoke in a way that was more natural to them and more expressive of their individual character feelings.
https://en.wikipedia.org/wiki?curid=26984
Salinity Salinity () is the saltiness or amount of salt dissolved in a body of water, called saline water (see also soil salinity). This is usually measured in formula_1 (note that this is technically dimensionless). Salinity is an important factor in determining many aspects of the chemistry of natural waters and of biological processes within it, and is a thermodynamic state variable that, along with temperature and pressure, governs physical characteristics like the density and heat capacity of the water. A contour line of constant salinity is called an "isohaline", or sometimes "isohale". Salinity in rivers, lakes, and the ocean is conceptually simple, but technically challenging to define and measure precisely. Conceptually the salinity is the quantity of dissolved salt content of the water. Salts are compounds like sodium chloride, magnesium sulfate, potassium nitrate, and sodium bicarbonate which dissolve into ions. The concentration of dissolved chloride ions is sometimes referred to as chlorinity. Operationally, dissolved matter is defined as that which can pass through a very fine filter (historically a filter with a pore size of 0.45 μm, but nowadays usually 0.2 μm). Salinity can be expressed in the form of a mass fraction, i.e. the mass of the dissolved material in a unit mass of solution. Seawater typically has a mass salinity of around 35 g/kg, although lower values are typical near coasts where rivers enter the ocean. Rivers and lakes can have a wide range of salinities, from less than 0.01 g/kg to a few g/kg, although there are many places where higher salinities are found. The Dead Sea has a salinity of more than 200 g/kg. Rainwater before touching the ground typically has a TDS of 20 mg/L or less. Whatever pore size is used in the definition, the resulting salinity value of a given sample of natural water will not vary by more than a few percent (%). Physical oceanographers working in the abyssal ocean, however, are often concerned with precision and intercomparability of measurements by different researchers, at different times, to almost five significant digits. A bottled seawater product known as IAPSO Standard Seawater is used by oceanographers to standardize their measurements with enough precision to meet this requirement. Measurement and definition difficulties arise because natural waters contain a complex mixture of many different elements from different sources (not all from dissolved salts) in different molecular forms. The chemical properties of some of these forms depend on temperature and pressure. Many of these forms are difficult to measure with high accuracy, and in any case complete chemical analysis is not practical when analyzing multiple samples. Different practical definitions of salinity result from different attempts to account for these problems, to different levels of precision, while still remaining reasonably easy to use. For practical reasons salinity is usually related to the sum of masses of a subset of these dissolved chemical constituents (so-called "solution salinity"), rather than to the unknown mass of salts that gave rise to this composition (an exception is when artificial seawater is created). For many purposes this sum can be limited to a set of eight major ions in natural waters, although for seawater at highest precision an additional seven minor ions are also included. The major ions dominate the inorganic composition of most (but by no means all) natural waters. Exceptions include some pit lakes and waters from some hydrothermal springs. The concentrations of dissolved gases like oxygen and nitrogen are not usually included in descriptions of salinity. However, carbon dioxide gas, which when dissolved is partially converted into carbonates and bicarbonates, is often included. Silicon in the form of silicic acid, which usually appears as a neutral molecule in the pH range of most natural waters, may also be included for some purposes (e.g., when salinity/density relationships are being investigated). The term 'salinity' is, for oceanographers, usually associated with one of a set of specific measurement techniques. As the dominant techniques evolve, so do different descriptions of salinity. Salinities were largely measured using titration-based techniques before the 1980s. Titration with silver nitrate could be used to determine the concentration of halide ions (mainly chlorine and bromine) to give a chlorinity. The chlorinity was then multiplied by a factor to account for all other constituents. The resulting 'Knudsen salinities' are expressed in units of parts per thousand (ppt or ‰). The use of electrical conductivity measurements to estimate the ionic content of seawater led to the development of the scale called the "practical salinity scale 1978" (PSS-78). Salinities measured using PSS-78 do not have units. The suffix psu or PSU (denoting "practical salinity unit") is sometimes added to PSS-78 measurement values. The addition of PSU as a unit after the value is "formally incorrect and strongly discouraged". In 2010 a new standard for the properties of seawater called the "thermodynamic equation of seawater 2010" (TEOS-10) was introduced, advocating absolute salinity as a replacement for practical salinity, and conservative temperature as a replacement for potential temperature. This standard includes a new scale called the "reference composition salinity scale". Absolute salinities on this scale are expressed as a mass fraction, in grams per kilogram of solution. Salinities on this scale are determined by combining electrical conductivity measurements with other information that can account for regional changes in the composition of seawater. They can also be determined by making direct density measurements. A sample of seawater from most locations with a chlorinity of 19.37 ppt will have a Knudsen salinity of 35.00 ppt, a PSS-78 practical salinity of about 35.0, and a TEOS-10 absolute salinity of about 35.2 g/kg. The electrical conductivity of this water at a temperature of 15 °C is 42.9 mS/cm. Limnologists and chemists often define salinity in terms of mass of salt per unit volume, expressed in units of mg per litre or g per litre. It is implied, although often not stated, that this value applies accurately only at some reference temperature. Values presented in this way are typically accurate to the order of 1%. Limnologists also use electrical conductivity, or "reference conductivity", as a proxy for salinity. This measurement may be corrected for temperature effects, and is usually expressed in units of μS/cm. A river or lake water with a salinity of around 70 mg/L will typically have a specific conductivity at 25 °C of between 80 and 130 μS/cm. The actual ratio depends on the ions present. The actual conductivity usually changes by about 2% per degree Celsius, so the measured conductivity at 5 °C might only be in the range of 50–80 μS/cm. Direct density measurements are also used to estimate salinities, particularly in highly saline lakes. Sometimes density at a specific temperature is used as a proxy for salinity. At other times an empirical salinity/density relationship developed for a particular body of water is used to estimate the salinity of samples from a measured density. Marine waters are those of the ocean, another term for which is "euhaline seas". The salinity of euhaline seas is 30 to 35. "Brackish seas" or waters have salinity in the range of 0.5 to 29 and "metahaline seas" from 36 to 40. These waters are all regarded as "thalassic" because their salinity is derived from the ocean and defined as "homoiohaline" if salinity does not vary much over time (essentially constant). The table on the right, modified from Por (1972), follows the "Venice system" (1959). In contrast to homoiohaline environments are certain "poikilohaline" environments (which may also be "thalassic") in which the salinity variation is biologically significant. "Poikilohaline" water salinities may range anywhere from 0.5 to greater than 300. The important characteristic is that these waters tend to vary in salinity over some biologically meaningful range seasonally or on some other roughly comparable time scale. Put simply, these are bodies of water with quite variable salinity. Highly saline water, from which salts crystallize (or are about to), is referred to as brine. Salinity is an ecological factor of considerable importance, influencing the types of organisms that live in a body of water. As well, salinity influences the kinds of plants that will grow either in a water body, or on land fed by a water (or by a groundwater). A plant adapted to saline conditions is called a halophyte. A halophyte which is tolerant to residual sodium carbonate salinity are called glasswort or saltwort or barilla plants. Organisms (mostly bacteria) that can live in very salty conditions are classified as extremophiles, or halophiles specifically. An organism that can withstand a wide range of salinities is euryhaline. Salt is expensive to remove from water, and salt content is an important factor in water use (such as potability). Increases in salinity have been observed in lakes and rivers in the United States, due to common road salt and other salt de-icers in runoff. The degree of salinity in oceans is a driver of the world's ocean circulation, where density changes due to both salinity changes and temperature changes at the surface of the ocean produce changes in buoyancy, which cause the sinking and rising of water masses. Changes in the salinity of the oceans are thought to contribute to global changes in carbon dioxide as more saline waters are less soluble to carbon dioxide. In addition, during glacial periods, the hydrography is such that a possible cause of reduced circulation is the production of stratified oceans. In such cases, it is more difficult to subduct water through the thermohaline circulation.
https://en.wikipedia.org/wiki?curid=26985
Saxifragales The Saxifragales (saxifrages) are an order of flowering plants (Angiosperms). They are an extremely diverse group of plants which include trees, shrubs, perennial herbs, succulent and aquatic plants. The degree of diversity in terms of vegetative and floral features makes it difficult to define common features that unify the order. In the Angiosperm Phylogeny Group classification system, the Saxifragales are placed within the major division of flowering plants referred to as eudicots, specifically the core eudicots. This subgroup consist of the Dilleniaceae, superasterids and superrosids. The superrosids in turn have two components, rosids and Saxifragales. The Saxifragales order has undergone considerable revision since its original classification based purely on plant characteristics. The modern classification is based on genetic studies using molecular phylogenetics. There is an extensive fossil record from the Turonian-Campanian (late Cretaceous) time, about 90 million years ago (Myr). However, but molecular studies suggest an earlier origin in the early Cretaceous (102–108 Myr) with rapid early diversification to more modern forms. The order Saxifragales consists of 15 families, about 100 genera and 2,470 species. Of the 15 families, many are small, with 8 having only a single genus, the largest family being the Crassulaceae (stonecrops) with about 35 genera. Saxifragales are found worldwide, though rarely in the tropics, and in a wide variety of habitats from desert to aquatic. They also have a wide variety of uses, from timber to foodstuffs and ornamental plants. Apart from ornamentals, the major economically important group is the Grossulariaceae (currants and gooseberries), particularly blackcurrant. The order Saxifragales is extremely morphologically diverse (hyper-diverse). It includes trees (e.g. witch hazel, witch alder in Hamamelidaceae), fruit bearing shrubs (e.g. currants, gooseberries in Grossulariaceae), lianas, annual and perennial herbs, rock garden plants (e.g. saxifrage in Saxifragaceae), ornamental garden plants (e.g. peonies in Paeoniaceae), succulents (e.g. stonecrop in Crassulaceae) and aquatics (e.g. watermilfoil in Haloragaceae). The flowers demonstrate major variations in sepal, petal, stamen, and carpel number, as well as ovary position ("see" Biogeography and evolution). This degree of diversity makes defining synapomorphy (derived common characteristics) for the group extremely difficult, the order being defined on the basis of molecular affinity rather than morphology. However, some characteristics that are prevalent (common traits) represent potential or putative synapomorphies based on ancestral states. These include flowers that are usually radially symmetric and petals that are free. The gynoecium (female reproductive part) generally consists of two carpels (ovary, style and stigma) that are free, at least toward the apex (partially fused bicarpellate gynoecium) and possess a hypanthium (cup shaped basal floral tube). In the androecium (male reproductive part), the stamen anthers are generally basifixed (attached at its base to the filament), sometimes dorsifixed (attached at centre) ("see Figure 2"). Other commonly occurring features are fruit that is generally follicular (formed from a single carpel), seeds with abundant endosperm surrounding the embryo and leaves with glandular teeth at their margins (glandular dentate, "see image"). Within the Saxifragales, while the families of the woody clade are primarily woody, the primarily herbaceous families of Crassulaceae and Saxifragaceae exhibit woody features as a secondary transition. With 15 families, about 100 genera and about 2,470 species, Saxifragales is a relatively small angiosperm order. Saxifragales was first described in 1820 by Berchtold and Presl in 1820 as a group of plants, Saxifrageae, with five genera, including "Saxifraga", and therefore bear their names as the botanical authority (Bercht. & J.Presl). At times, that authority has also been given to Dumortier, due to a later publication (1829). Dumortier first used the word Saxifragaceae. By the time of John Lindley's "The Vegetable Kingdom" (1853), the term Saxifragales was in use, which Lindley called an Alliance, containing five families. Later, the Saxifragales were placed in the angiosperm class Dicotyledons, also called Magnoliopsida. The order Saxifragales has undergone considerable revision in both placement and composition, since the use of molecular phylogenetics, and the use of the modern Angiosperm Phylogeny Group (APG) classification. They are identified as a strongly monophyletic group. In the initial APG publication (1998), the Saxifragales were identified within the core eudicots clade but its relationship to other clades was uncertain. The core eudicots consist of the order Gunnerales and a large clade of Pentapetalae (so named for having a synapomorphy of pentamerous (5 part) perianths), the latter representing about 70% of all angiosperms, with eight major lineages. Later (2003), the order was described as "one of the major surprises of molecular phylogenetic analyses of the angiosperms", having elements previously placed in three or four separate subclasses based on morphology. This was eventually resolved in the third APG system (2009) placing Saxifragales as a sister group to the rosids (Rosidae), within the Pentapetalae clade. This large combination has subsequently been given the name superrosids (Superrosidae), representing part of an early diversification of the angiosperms. Among the rosids, they share a number of similarities with the Rosales, particularly Rosaceae, including a hypanthium, five part flowers and free floral parts. As circumscribed, Saxifragales account for 1.3% of eudicot diversity. Diversification among Saxifragales was rapid, with the extensive fossil record indicating that the order was more diverse and more widespread than an examination of the extant members suggests, with considerable phenotypic diversity occurring early. The earliest fossil evidence is found in the Turonian-Campanian (late Cretaceous), suggesting a minimum age of 89.5 Myr. However, molecular divergence time estimation suggest an earlier time of 102–108 Myr, into the early Cretaceous, for the crown and stem groups respectively. Within the order Saxifragales, the molecular data imply a very rapid initial diversification time of about 6–8 Myr, between 112–120 Myr, with major lineages appearing within 3–6 Myr. The ancestral state appears to be woody, as in Peridiscaceae and the woody clade, but is also ancestral to Grossulariaceae. A number of independent transitions to a herbaceous habit occurred in the ancestors of Crassulaceae, Saxifragaceae and the base of the Haloragaceae-Penthoraceae clade (the other two families in Haloragaceae "s.l." remaining woody}, while other taxa reverted to a woody habit, especially Crassulaceae. Most of Saxifragales have a superior ovary, but some families show frequent transition with inferior or subinferior position, particularly Saxifragaceae and to a lesser extent Hamamelidaceae. Almost all Grossulariaceae have an inferior ovary. The ancestral carpel number is two, with transition to higher numbers, such as four in Haloragaceae "s.l." and Peridiscaceae with five in Penthoraceae. The ancestral carpel number for Crassulaceae is five, decreasing to four in "Kalanchoe", where it is synapomorphic for the genus, though the most frequent transition in this family is 6–10, but only where stamen number is increased above five. Some Macaronesian taxa (Aeonieae) have 8–12, with up to 32 carpels for "Aeonium". The ancestral petal number is five, with three major transitions; 5 to 0, 5 to 4, 5 to 6–10. Increased petal number is seen in Paeoniaceae and Crassulaceae, particularly where stamen number is also increased. Cercidiphyllum + Daphniphyllum, Chrysosplenium and "Altingia" are examples of the complete loss of petals. The ancestral stamen:petal ratio is 1, with transitions characterising several clades, e.g. Paeonicaceae+woody clade >2, Crassulaceae 2 (but "Crassula" 1). Overall there has been a decrease over evolution, but independent of a decrease in petal number, so that it is the stamen number that has decreased. The ancestral habitat appears to be forests, followed by early diversification into desert and aquatic habitats, with shrubland the most recent colonization. Species diversification was rapid following a transition from a warmer, wetter Earth in the Eocene (56–40 Myr) to early Miocene (23–16 Myr), to the cooler drier conditions of the mid-Miocene (16–12 Myr). However, this appears to not have coincided with ecological and phenotypic evolution, which are themselves correlated. There is a clear lag, whereby increase in species diversification was followed later by increases in niche and phenotypic lability. The first APG classification (1998) placed 13 families with the order Saxifragales: This was subsequently revised to 15, in the fourth version (2016). The Saxifragales families have been grouped into a number of informally named suprafamilial subclades, with the exception of the basal split of Peridiscaceae, which thus forms a sister group with the rest of Saxifragales. The two major ones are (Paeoniaceae + the woody clade of primarily woody families) and the "core" Saxifragales (i.e. the primarily herbaceous families), with the latter subdivided into two further subclades, (Haloragaceae "sensu lato" + Crassulaceae) and the Saxifragaceae alliance. In the clade Haloragaceae "sensu lato" "(s.l.)" + Crassulaceae the genera constituting Haloragaceae "s.l." are all small, and APG II (2003) proposed merging them into a single larger Haloragaceae "s.l.", but transferred "Aphanopetalum" from Cunoniaceae to this group. The Saxifragaceae alliance represents Saxifragaceae together with a number of woody members of the traditional Saxifragaceae "sensu" Engler (1930). Within this, APG II (2003) proposed placing the two species of "Pterostemon" that constitute Pterostemonaceae within Iteaceae, and all subsequent versions have maintained this practice. Thus Saxifragales "sensu" APG II consisted of only 10 families. The third version (2009) added Peridiscaceae (from Malpighiales), as sister to all other families, but re-expanded Haloragaceae to provide for a narrower circumscription, Haloragaceae "sensu stricto" ("s.s."), to give a total of 14 families. APG IV (2016) added the parasitic family Cynomoriaceae to provide a total of 15 families, although its placement within the order remained unclear. Of the 15 families included in APG IV, the basal divergence Peridiscaceae underwent radical shifting and recircumscription from 2003 to 2009. Originally, it consisted of two closely related genera, "Peridiscus" and "Whittonia". The APG II system placed the family in Malpighiales, based on a DNA sequence for the "rbcL" gene from "Whittonia". This sequence turned out to be not from "Whittonia", but from other plants whose DNA had contaminated the sample. After placement in Saxifragales, it was expanded to include "Soyauxia" in 2007, and "Medusandra" in 2009. In the first of the subclades of the remaining Saxifragales, Paeoniaceae possesses many unique features and its taxonomic position was controversial for a long time, and "Paeonia" was placed in Ranunculales, close to "Glaucidium", prior to transfer to Saxifragales as sister to the woody clade. In the woody clade, the genus "Liquidamber" was included in Hamamelidaceae until molecular phylogenetic studies showed that its inclusion might make Hamamelidaceae paraphyletic, and was segregated as a separate monotypic family, Altingiaceae in 2008. Cercidiphyllaceae was for a long time associated with Hamamelidaceae and Trochodendraceae and was often thought to be closer to the latter, which is now in the basal eudicot order Trochodendrales. "Daphniphyllum" was always thought to have an anomalous combination of characters and was placed in several different orders before molecular phylogenetic analysis showed it to belong to Saxifragales. In the core Saxifragales, Crassulaceae and Tetracarpaeaceae have been associated with Saxifragaceae, while "Penthorum" has been associated both with Crassulaceae and Saxifragaceae, before being placed here. "Aphanopetalum" was often placed in Cunoniaceae, a family in Oxalidales, even though there were good reasons to put it in Saxifragales, and it was subsequently transferred. Haloragaceae was included in Myrtales, before being placed in Saxifragales. The other "core" group, the Saxifragaceae alliance comprises four families: Pterostemonaceae, Iteaceae, Grossulariaceae, and Saxifragaceae, which have long been known to be related to each other, but the circumscription of Saxifragaceae has been much reduced and Pterostemonaceae submerged as "Pterostemon" in Iteaceae. Most of the families are monogeneric. "Choristylis" is now considered a synonym of Itea, but the addition of "Pterostemon", gives Iteaceae two genera. "Liquidambar" and "Semiliquidambar" are also submerged into "Altingia", making Altingiaceae monogeneric. About 95% of the species are in five families: Crassulaceae (1400), Saxifragaceae (500), Grossulariaceae (150–200), Haloragaceae (150), and Hamamelidaceae (100). The relationships of the Saxifragales families to each other is shown in the following cladogram. The phylogeny in this cladogram still has some uncertainty as to the exact relationships, and the phylogenetic tree is subject to further revision. Cynomoriaceae, previously placed in Santales or Rosales is included in Saxifragales, but unplaced within it. Li et al. (2019) have slightly different relationships, and also place Cynomoriaceae as the first branch in the Crassulaceae+Haloragaceae "s.l." tree, i.e. as sister to those two families. The number of genera in each family is shown in parentheses: Saxifragales are found worldwide, though primarily in temperate zones and rarely in the tropics. They occupy a wide variety of habitats from arid desert (Crassulaceae) to aquatic conditions (Haloragaceae), with 6 families, including North American species, that are obligate aquatic (fully dependant on an aquatic environment), and including forests, grasslands and tundra. Saxifragales exceeds all other comparably sized clades in terms of diversity of habitats. Most of the diversity occurs in temperate (including montane and arid) conditions that expanded globally during cooling and drying trends in the last 15 My. The most common habitats are forests and cliffs, with about 300 species occupying each, but with forests being the most diverse phenotypically, where nearly all families are represented. In contrast desert and tundra, with only two families each, contain only about 10% of species. About 90% of species can be assigned to a single habitat. "Whittonia" (Peridiscaceae) is thought to be extinct. the IUCN lists 9 critically endangered, 12 endangered, 19 vulnerable and 7 near threatened species. Among the most threatened Saxifragales are "Aichryson dumosum" and "Monanthes wildpretii" (Crassulaceae), "Haloragis stokesii" and "Myriophyllum axilliflorum" (Haloragaceae), "Ribes malvifolium" and "R. sardoum" (Grossulariaceae), "Saxifraga artvinensis" (Saxifragaceae) and "Molinadendron hondurense" (Hamamelidaceae). Plants in the order Saxifragales have found a wide variety of uses, including traditional medicines, ornamental, household, aquarium, pond and garden plants, spices, foodstuffs (fruit and greens), dyestuffs, smoking, resin, timber and roof coverings (see Families). A number of Saxifragales genera are commercially cultivated. "Paeonia" are cultivated both as ornamental shrubs (generally sold as root stock) and for cut flowers, with the Netherlands representing the largest production, other more minor producers are Israel, New Zealand, Chile and the United States. "Liquidamber" is used for hardwood, with the American Sweetgum ("Liquidambar styraciflua") being among the most important sources of commercial hardwood in the Southeast United States, with one of its uses being veneer for plywood. "Hamamelis" is cultivated in New England for distilleries extracting witch-hazel, widely used in skincare, and is the largest source of this medicament in the world. Among the Crassulaceae, economic importance is limited to horticulture, with many species and cultivars important as ornamentals, including "Crassula ovata" (jade plant) and "Jovibarba" (hen and chicken). "Hylotelphium", "Phedimus", "Sedum" and "Sempervivum" are cultivated for rock gardens and for "green roofs". In particular, cultivars of the Madagascan "Kalanchoe blossfeldiana", e.g. 'Florists kalanchoe' have achieved commercial success throughout the world, being popular Christmas decorative plants. The Haloragaceae aquatic genus "Myriophyllum" and the closely related "Proserpinaca" are cultivated for the commercial aquarium trade. "Myriophyllum" is also economically important for purification of water and as feed for pigs, ducks, and fish, and polishing wood. A number of "Ribes" (Grossulariaceae) are in commercial production, concentrated in Europe and the USSR from species native to those areas. "R. nigrum" (blackcurrant) was first cultivated in monastery gardens in Russia in the 11th century, and currant cultivation more generally later in Western Europe, "R. uva-crispa" (gooseberry) production began around 1700. The first colonists in N America began cultivating currants in the late 1700s. "R. nigrum" is the most important commercial currant crop, being produced in more than 23 countries, with the major centres being Russia (more than 63 thousand hectares), Poland, Germany, Scandinavia and the UK. An important source of Vitamin C, black currants are used in the manufacture of jam, fruit jelly, compote, syrup, juice and other drinks, including the cordial Ribena and the liqueur Cassis. Other commercial crops include "R. rubrum" (red currant). World "Ribes" crop production was over 750,000 tons in 2002, of which about 150,000 tons were gooseberries, and the largest group blackcurrants.
https://en.wikipedia.org/wiki?curid=26987
CLIÉ The Sony CLIÉ is a series of personal digital assistants running the Palm Operating System developed and marketed by Sony from 2000 to 2005. The devices introduced many new features to the PDA market, such as a jog-wheel interface, high-resolution displays, and Sony technologies like Memory Stick slots and ATRAC3 audio playback. Most models were designed and manufactured in Japan. The name is an acronym for "creativity, lifestyle, innovation, emotion" though formerly "communication, link, information and entertainment". It was initially an attempt at a new coinage term, though it means "tool" in the Jèrriais language. The CLIÉ handhelds were distinguished from other Palm OS models by their emphasis on multimedia capabilities, including photo, video, and audio playback, long before any other Palm OS PDAs had such capabilities. Later models have been credited with spurring competition in the previously stagnant Palm market, closing many of the gaps that existed between Palm OS PDAs and those powered by Microsoft's Windows Mobile operating system, particularly on the multimedia front, but also with Sony's proprietary application launcher interface. In the summer of 2004, Sony announced that new CLIÉs would, from then on, be manufactured and available only in Japan, and in the spring of 2005, Sony announced the total termination of its CLIÉ line of products. The last models to be released worldwide were the PEG-TJ27, PEG-TJ37, and PEG-TH55. The last model released in Japan was the PEG-VZ90. Soon after the closure of the CLIÉ line, Sony stopped providing original installation drivers, including Sony's version of Palm Desktop for the CLIÉ, which are necessary for Hotsyncing with the PC and otherwise taking advantage of the handhelds' many features for which a PC may be required. Several CLIÉ fans took it upon themselves to offer these drivers freely for download at www.sonyclie.org. CLIÉ handhelds were released in series, usually with a few models released in each series. In later years, multiple series would be in production at the same time. Officially, the CLIÉ line did not support the Mac OS, and Sony never provided any software with the handhelds for Mac OS. However, as a Palm OS device, every CLIÉ handheld was inherently capable of HotSync operations with a Mac OS computer. This allowed for the synchronization of the basic PIM functions, as well as for the installation of new software, though this inherent capability was unusable because the Mac HotSync software would not recognize the handheld. PalmSource, however, silently added the capability to recognize older CLIÉ devices when providing new versions of its Palm Desktop software for Mac. This was necessary for those who could synchronize only via USB. The CLIÉ user community soon discovered that these "updates" were simply a matter of adding a few lines to the USB-detection property-list file. Since then, detailed instructions have been posted online for those who want to synchronize their CLIÉ handhelds. No modifications are required for Bluetooth synchronization, but Wi-Fi synchronization is impossible because the Mac OS HotSync software does not support network synchronization. Some workarounds for the multimedia features also exist. For those who desire stronger Mac OS/CLIÉ integration, the product Missing Sync made by the company Mark/Space is also available. This does make unencrypted Wi-Fi synchronization possible but a bug in the CLIÉ network stack reverses IP addresses which means that the Macintosh involved needs a palindromic IP address such as 10.0.0.10.
https://en.wikipedia.org/wiki?curid=26988
Sony Sony Corporation is the electronics business unit and the parent company of the , which is engaged in business through its four operating components: electronics (AV, IT & communication products, semiconductors, video games, network services and medical business), motion pictures (movies and TV shows), music (record labels and music publishing) and financial services (banking and insurance). These make Sony one of the most comprehensive entertainment companies in the world. The group consists of Sony Corporation, Sony Electronics, Sony Semiconductor Solutions, Sony Pictures, Sony Interactive Entertainment, Sony Music, Sony Financial Holdings, and others. Sony is among the semiconductor sales leaders and as of 2015, the fifth-largest television manufacturer in the world after Samsung Electronics, LG Electronics, TCL and Hisense, although it is the world's largest player in the premium TV market, a market for a television of at least 55-inch in size, with a price higher than $2,500. The company's current slogan is "Be Moved". Their former slogans were "The One and Only" (1979–1982), "It's a Sony" (1982–2005), "like.no.other" (2005–2009) and "make.believe" (2009–2013). Sony has a weak tie to the Sumitomo Mitsui Financial Group (SMFG) corporate group, the successor to the Mitsui group. Sony began in the wake of World War II. In 1946, Masaru Ibuka started an electronics shop in a department store building in Tokyo. The company started with a capital of ¥190,000 and a total of eight employees. On 7 May 1946, Ibuka was joined by Akio Morita to establish a company called (Tokyo Telecommunications Engineering Corporation). The company built Japan's first tape recorder, called the Type-G. In 1958, the company changed its name to "Sony". When "Tokyo Tsushin Kogyo" was looking for a romanized name to use to market themselves, they strongly considered using their initials, TTK. The primary reason they did not is that the railway company Tokyo Kyuko was known as TTK. The company occasionally used the acronym "Totsuko" in Japan, but during his visit to the United States, Morita discovered that Americans had trouble pronouncing that name. Another early name that was tried out for a while was "Tokyo Teletech" until Akio Morita discovered that there was an American company already using Teletech as a brand name. The name "Sony" was chosen for the brand as a mix of two words: one was the Latin word ""sonus"", which is the root of sonic and sound, and the other was ""sonny"", a common slang term used in 1950s America to call a young boy. In 1950s Japan, "sonny boys" was a loan word in Japanese, which connoted smart and presentable young men, which Sony founders Akio Morita and Masaru Ibuka considered themselves to be. The first Sony-branded product, the TR-55 transistor radio, appeared in 1955 but the company name did not change to Sony until January 1958. At the time of the change, it was extremely unusual for a Japanese company to use Roman letters to spell its name instead of writing it in kanji. The move was not without opposition: TTK's principal bank at the time, Mitsui, had strong feelings about the name. They pushed for a name such as Sony Electronic Industries, or Sony Teletech. Akio Morita was firm, however, as he did not want the company name tied to any particular industry. Eventually, both Ibuka and Mitsui Bank's chairman gave their approval. According to Schiffer, Sony's TR-63 radio "cracked open the U.S. market and launched the new industry of consumer microelectronics." By the mid-1950s, American teens had begun buying portable transistor radios in huge numbers, helping to propel the fledgling industry from an estimated 100,000 units in 1955 to 5 million units by the end of 1968. Sony co-founder Akio Morita founded Sony Corporation of America in 1960. In the process, he was struck by the mobility of employees between American companies, which was unheard of in Japan at that time. When he returned to Japan, he encouraged experienced, middle-aged employees of other companies to reevaluate their careers and consider joining Sony. The company filled many positions in this manner, and inspired other Japanese companies to do the same. Moreover, Sony played a major role in the development of Japan as a powerful exporter during the 1960s, 1970s and 1980s. It also helped to significantly improve American perceptions of "made in Japan" products. Known for its production quality, Sony was able to charge above-market prices for its consumer electronics and resisted lowering prices. In 1971, Masaru Ibuka handed the position of president over to his co-founder Akio Morita. Sony began a life insurance company in 1979, one of its many peripheral businesses. Amid a global recession in the early 1980s, electronics sales dropped and the company was forced to cut prices. Sony's profits fell sharply. "It's over for Sony," one analyst concluded. "The company's best days are behind it." Around that time, Norio Ohga took up the role of president. He encouraged the development of the Compact Disc in the 1970s and 1980s, and of the PlayStation in the early 1990s. Ohga went on to purchase CBS Records in 1988 and Columbia Pictures in 1989, greatly expanding Sony's media presence. Ohga would succeed Morita as chief executive officer in 1989. Under the vision of co-founder Akio Morita and his successors, the company had aggressively expanded into new businesses. Part of its motivation for doing so was the pursuit of "convergence," linking film, music and digital electronics via the Internet. This expansion proved unrewarding and unprofitable, threatening Sony's ability to charge a premium on its products as well as its brand name. In 2005, Howard Stringer replaced Nobuyuki Idei as chief executive officer, marking the first time that a foreigner had run a major Japanese electronics firm. Stringer helped to reinvigorate the company's struggling media businesses, encouraging blockbusters such as "Spider-Man" while cutting 9,000 jobs. He hoped to sell off peripheral business and focus the company again on electronics. Furthermore, he aimed to increase cooperation between business units, which he described as "silos" operating in isolation from one another. In a bid to provide a unified brand for its global operations, Sony introduced a slogan known as "make.believe" in 2009. Despite some successes, the company faced continued struggles in the mid- to late-2000s. In 2012, Kazuo Hirai was promoted to president and CEO, replacing Stringer. Shortly thereafter, Hirai outlined his company-wide initiative, named "One Sony" to revive Sony from years of financial losses and bureaucratic management structure, which proved difficult for former CEO Stringer to accomplish, partly due to differences in business culture and native languages between Stringer and some of Sony's Japanese divisions and subsidiaries. Hirai outlined three major areas of focus for Sony's electronics business, which include imaging technology, gaming and mobile technology, as well as a focus on reducing the major losses from the television business. In February 2014, Sony announced the sale of its Vaio PC division to a new corporation owned by investment fund Japan Industrial Partners and spinning its TV division into its own corporation as to make it more nimble to turn the unit around from past losses totaling $7.8 billion over a decade. Later that month, they announced that they would be closing 20 stores. In April, the company announced that they would be selling 9.5 million shares in Square Enix (roughly 8.2 percent of the game company's total shares) in a deal worth approximately $48 million. In May 2014 the company announced it was forming two joint ventures with Shanghai Oriental Pearl Group to manufacture and market Sony's PlayStation games consoles and associated software in China. It was reported in December 2016 by multiple news outlets that Sony was considering restructuring its U.S. operations by merging its TV & film business, Sony Pictures Entertainment, with its gaming business, Sony Interactive Entertainment. According to the reports, such a restructuring would have placed Sony Pictures under Sony Interactive's CEO, Andrew House, though House wouldn't have taken over day-to-day operations of the film studio. According to one report, Sony was set to make a final decision on the possibility of the merger of the TV, film, & gaming businesses by the end of its fiscal year in March of the following year (2017). In 2017, Sony sold its battery business to Murata Manufacturing. In 2019, Sony merged its Mobile, TV and Camera businesses. On April 1, 2020, Sony Electronics Corporation was established as an intermediate holding company to own and oversee its electronics and IT solutions businesses. On May 19, 2020, the company announced that it will rename Sony Group Corporation as of April 1, 2021. Subsequently, Sony Electronics Corporation will be renamed to Sony Corporation. On the same day the company announced that it will turn Sony Financial Holdings, of which Sony already owns 65.06% of shares, to a wholly owned subsidiary through a takeover bid. Sony has historically been notable for creating its own in-house standards for new recording and storage technologies, instead of adopting those of other manufacturers and standards bodies. Sony (either alone or with partners) has introduced several of the most popular recording formats, including the floppy disk, Compact Disc and Blu-ray Disc. The company launched the Betamax videocassette recording format in 1975. Sony was involved in the videotape format war of the early 1980s, when they were marketing the Betamax system for video cassette recorders against the VHS format developed by JVC. In the end, VHS gained critical mass in the marketbase and became the worldwide standard for consumer VCRs. Betamax is, for all practical purposes, an obsolete format. Sony's professional-oriented component video format called Betacam, which was derived from Betamax, was used until 2016 when Sony announced it was stopping production of all remaining 1/2-inch video tape recorders and players, including the Digital Betacam format. In 1985, Sony launched their Handycam products and the Video8 format. Video8 and the follow-on hi-band Hi8 format became popular in the consumer camcorder market. In 1987 Sony launched the 4 mm DAT or Digital Audio Tape as a new digital audio tape standard. In 1979, the Walkman brand was introduced, in the form of the world's first portable music player using the compact cassette format. Sony introduced the MiniDisc format in 1992 as an alternative to Philips DCC or Digital Compact Cassette and as a successor to the compact cassette. Since the introduction of MiniDisc, Sony has attempted to promote its own audio compression technologies under the ATRAC brand, against the more widely used MP3. Until late 2004, Sony's Network Walkman line of digital portable music players did not support the MP3 standard natively. In 2004, Sony built upon the MiniDisc format by releasing Hi-MD. Hi-MD allows the playback and recording of audio on newly introduced 1 GB Hi-MD discs in addition to playback and recording on regular MiniDiscs. In addition to saving audio on the discs, Hi-MD allows the storage of computer files such as documents, videos and photos. In 1993, Sony challenged the industry standard Dolby Digital 5.1 surround sound format with a newer and more advanced proprietary motion picture digital audio format called SDDS (Sony Dynamic Digital Sound). This format employed eight channels (7.1) of audio opposed to just six used in Dolby Digital 5.1 at the time. Ultimately, SDDS has been vastly overshadowed by the preferred DTS (Digital Theatre System) and Dolby Digital standards in the motion picture industry. SDDS was solely developed for use in the theatre circuit; Sony never intended to develop a home theatre version of SDDS. Sony and Philips jointly developed the Sony-Philips digital interface format (S/PDIF) and the high-fidelity audio system SACD. The latter became entrenched in a format war with DVD-Audio. Neither gained a major foothold with the general public. CDs were preferred by consumers because of the ubiquitous presence of CD drives in consumer devices until the early 2000s when the iPod and streaming services became available. In 1983, Sony followed their counterpart Philips to the compact disc (CD). In addition to developing consumer-based recording media, after the launch of the CD Sony began development of commercially based recording media. In 1986 they launched Write-Once optical discs (WO) and in 1988 launched Magneto-optical discs which were around 125MB size for the specific use of archival data storage. In 1984, Sony launched the Discman series which extended their Walkman brand to portable CD products. In the early 1990s, two high-density optical storage standards were being developed: one was the MultiMedia Compact Disc (MMCD), backed by Philips and Sony, and the other was the Super Density disc (SD), supported by Toshiba and many others. Philips and Sony abandoned their MMCD format and agreed upon Toshiba's SD format with only one modification. The unified disc format was called DVD and was introduced in 1997. Sony was one of the leading developers of the Blu-ray optical disc format, the newest standard for disc-based content delivery. The first Blu-ray players became commercially available in 2006. The format emerged as the standard for HD media over the competing format, Toshiba's HD DVD, after a two-year-long high-definition optical disc format war. In 1983, Sony introduced 90 mm micro diskettes (better known as floppy disks), which it had developed at a time when there were 4" floppy disks, and a lot of variations from different companies, to replace the then on-going 5.25" floppy disks. Sony had great success and the format became dominant. 3.5" floppy disks gradually became obsolete as they were replaced by current media formats. Sony launched in 1998, their Memory Stick format, flash memory cards for use in Sony lines of digital cameras and portable music players. It has seen little support outside of Sony's own products, with Secure Digital cards (SD) commanding considerably greater popularity. Sony has made updates to the Memory Stick format with Memory Stick Duo and Memory Stick Micro. Sony offers products in a variety of product lines around the world. Sony has developed a music playing robot called Rolly, dog-shaped robots called AIBO and a humanoid robot called QRIO. As of 1 April 2016, Sony is organized into the following business segments: Game & Network Services (G&NS), Music, Pictures, Electronics Products & Solutions (EP&S), Imaging & Sensing Solutions (I&SS), Financial Services, and Others The network and medical businesses are included in the G&NS and EP&S, respectively. Sony Corporation is the electronics business unit and the parent company of the Sony Group. It primarily conducts strategic business planning of the group, research and development (R&D), planning, designing and marketing for electronics products. Its subsidiaries such as Sony Global Manufacturing & Operations Corporation (SGMO; 4 plants in Japan), Sony Semiconductor Manufacturing Corporation (7 plants in Japan), Sony Storage Media and Devices Corporation, Sony Energy Devices Corporation and its subsidiaries outside Japan (Brazil, China, UK (Wales), India, Malaysia, Singapore, South Korea, Thailand, Ireland and United States) are responsible for manufacturing as well as product engineering (SGMO is also responsible for customer service operations). In 2012, Sony rolled most of its consumer content services (including video, music and gaming) into the Sony Entertainment Network. Sony produced the world's first portable music player, the Walkman in 1979. This line fostered a fundamental change in music listening habits by allowing people to carry music with them and listen to music through lightweight headphones. Walkman originally referred to portable audio cassette players. The company now uses the Walkman brand to market its portable audio and video players as well as a line of former Sony Ericsson mobile phones. Sony utilized a related brand, Discman, to refer to its CD players. It dropped this name in the late 1990s. Sony produced computers (MSX home computers and NEWS workstations) during the 1980s. The company withdrew from the computer business around 1990. Sony entered again into the global computer market under the new VAIO brand, began in 1996. Short for "Video Audio Integrated Operation", the line was the first computer brand to highlight visual-audio features. Sony faced considerable controversy when some of its laptop batteries exploded and caught fire in 2006, resulting in the largest computer-related recall to that point in history. In a bid to join the tablet computer market, the company launched its Sony Tablet line of Android tablets in 2011. Since 2012, Sony's Android products have been marketed under the Xperia brand used for its smartphones. On 4 February 2014, Sony announced that it would sell its VAIO PC business due to poor sales and Japanese company Japan Industrial Partners (JIP) will purchase the VAIO brand, with the deal finalized by the end of March 2014. Sony maintains a minority stake in the new, independent company. Sony offers a wide range of digital cameras. Point-and-shoot models adopt the Cyber-shot name, while digital single-lens reflex models are branded using Alpha. The first Cyber-shot was introduced in 1996. At the time, digital cameras were a relative novelty. Sony's market share of the digital camera market fell from a high of 20% to 9% by 2005. Sony entered the market for digital single-lens reflex cameras in 2006 when it acquired the camera business of Konica Minolta. Sony rebranded the company's line of cameras as its Alpha line. Sony is the world's third largest manufacturer of the cameras, behind Canon and Nikon respectively. There are also a variety of Camcorders which are manufactured by Sony. In 1968, Sony introduced the Trinitron brand name for its lines of aperture grille cathode ray tube televisions and (later) computer monitors. Sony stopped production of Trinitron for most markets, but continued producing sets for markets such as Pakistan, Bangladesh and China. Sony discontinued its series of Trinitron computer monitors in 2005. The company discontinued the last Trinitron-based television set in the US in early 2007. The end of Trinitron marked the end of Sony's analog television sets and monitors. Sony used the LCD WEGA name for its LCD TVs until summer 2005. The company then introduced the BRAVIA name. BRAVIA is an in house brand owned by Sony which produces high-definition LCD televisions, projection TVs and front projectors, home cinemas and the BRAVIA home theatre range. All Sony high-definition flat-panel LCD televisions in North America have carried the logo for BRAVIA since 2005. Sony is the third-largest maker of televisions in the world. , Sony's television business has been unprofitable for eight years. In December 2011, Sony agreed to sell all stake in an LCD joint venture with Samsung Electronics (S-LCD) for about $940 million. On 28 March 2012, Sony Corporation and Sharp Corporation announced that they have agreed to further amend the joint venture agreement originally executed by the parties in July 2009, as amended in April 2011, for the establishment and operation of Sharp Display Products Corporation ("SDP"), a joint venture to produce and sell large-sized LCD panels and modules. On 9 November 2015, Sony announced that they are going to stop producing Betamax Tapes in March 2016. Sony also sells a range of DVD players. It has shifted its focus in recent years to promoting the Blu-ray format, including discs and players. Sony produces a wide range of semiconductors and electronic components including image sensors (Exmor), image processor (BIONZ), laser diodes, system LSIs, mixed-signal LSIs, OLED panels, etc. The company has a strong presence in the image sensor market. Sony-manufactured CMOS image sensors are widely used in digital cameras, tablet computers and smartphones. In April 2018, Sony announced to join the market for satellite communications and develop laser communication products for small satellites. Sony wants to use its heritage with optical disc technology used in products like CD players and plans to start initial tests in collaboration with JAXA in 2018. Sony has targeted medical, healthcare and biotechnology business as a growth sector in the future. The company acquired iCyt Mission Technology, Inc. (renamed Sony Biotechnology Inc. in 2012), a manufacturer of flow cytometers, in 2010 and Micronics, Inc., a developer of microfluidics-based diagnostic tools, in 2011. In 2012, Sony announced that it will acquire all shares of So-net Entertainment Corporation, which is the majority shareholder of M3, Inc., an operator of portal sites (m3.com, MR-kun, MDLinx and MEDI:GATE) for healthcare professionals. On 28 September 2012, Olympus and Sony announced that the two companies will establish a joint venture to develop new surgical endoscopes with 4K resolution (or higher) and 3D capability. Sony Olympus Medical Solutions Inc. (Sony 51%, Olympus 49%) was established on 16 April 2013. On 28 February 2014, Sony, M3 and Illumina established a joint venture called P5, Inc. to provide a genome analysis service for research institutions and enterprises in Japan. Sony Mobile Communications Inc. (formerly Sony Ericsson) is a multinational mobile phone manufacturing company headquartered in Tokyo, Japan and a wholly owned subsidiary of Sony Corporation. In 2001, Sony entered into a joint venture with Swedish telecommunications company Ericsson, forming Sony Ericsson. Initial sales were rocky, and the company posted losses in 2001 and 2002. However, SMC reached a profit in 2003. Sony Ericsson distinguished itself with multimedia-capable mobile phones, which included features such as cameras. These were unusual for the time. Despite their innovations, SMC faced intense competition from Apple's iPhone which released in 2007. From 2008 to 2010, amid a global recession, SMC slashed its workforce by several thousand. Sony acquired Ericsson's share of the venture in 2012 for over US$1 billion. In 2009, SMC was the fourth-largest mobile phone manufacturer in the world (after Nokia, Samsung and LG). By 2010, its market share had fallen to sixth place. Sony Mobile Communications now focuses exclusively on the smartphone market under the Xperia name. In 2015, Sony released Xperia Z5 Premium in Canada following US and Europe. In the year 2013, Sony contributed to around two percent of the mobile phone market with 37 million mobile phones sold. Sony Interactive Entertainment (formerly Sony Computer Entertainment) is best known for producing the popular line of PlayStation consoles. The line grew out of a failed partnership with Nintendo. Originally, Nintendo requested Sony to develop an add-on for its console that would play Compact Discs. In 1991 Sony announced the add-on, as well as a dedicated console known as the "Play Station". However, a disagreement over software licensing for the console caused the partnership to fall through. Sony then continued the project independently. Launched in 1994, the first PlayStation gained 61% of global console sales and broke Nintendo's long-standing lead in the market. Sony followed up with the PlayStation 2 in 2000, which was even more successful. The console has become the most successful of all time, selling over 150 million units . Sony released the PlayStation 3, a high-definition console, in 2006. It was the first console to use the Blu-ray format, and was considerably more expensive than competitors Xbox 360 and Wii due to a Cell processor. Early on, poor sales performance resulted in significant losses for the company, pushing it to sell the console at a loss. The PlayStation 3 sold generally more poorly than its competitors in the early years of its release but managed to overtake the Xbox 360 in global sales later on. It later introduced the PlayStation Move, an accessory that allows players to control video games using motion gestures. Sony extended the brand to the portable games market in 2004 with the PlayStation Portable (PSP). The console has sold reasonably, but has taken a second place to a rival handheld, the Nintendo DS. Sony developed the Universal Media Disc (UMD) optical disc medium for use on the PlayStation Portable. Early on, the format was used for movies, but it has since lost major studio support. Sony released a disc-less version of its PlayStation Portable, the PSP Go, in 2009. The company went on to release its second portable video game system, PlayStation Vita, in 2011 and 2012. Sony launched its fourth console, the PlayStation 4, on 15 November 2013, which as of 31 December 2017 has sold 73.6 million units globally. On 18 March 2014, at GDC, president of Sony Computer Entertainment Worldwide Studios Shuhei Yoshida announced their new virtual reality technology dubbed Project Morpheus, and later named PlayStation VR, for PlayStation 4. The headset brought VR gaming and non-gaming software to the company's console. According to a report released by Houston-based patent consulting firm LexInnova in May 2015, Sony is leading the virtual reality patent race. According to the firm's analysis of nearly 12,000 patents or patent applications, Sony has 366 virtual reality patents or patent applications. PlayStation VR was released worldwide on 13 October 2016. In 2014, Sony participated within NRG Energy eVgo Ready for Electric Vehicle (REV) program, for EV charging parking lots. Sony is in the business of electric vehicle lithium-ion batteries; on 28 July 2016, Sony announced that the company will sell its battery business to Murata Manufacturing. IT giants such as Google (driverless car) and Apple (iCar/Project Titan) are working on electric vehicles and self driving cars, competing with Tesla; Sony is entering into this field by investing $842,000 in the ZMP company. On 6 January 2020 at the Consumer Electronics Show in Las Vegas, Sony announced a concept Electric Vehicle called the Vision-S. In January 2020 Sony unveiled a concept electric car at the Consumer Electronics Show, named Vision-S, designed in collaboration with components manufacturer Magna International. At the occasion Sony also stated its goal of developing technology for the automotive sector, especially concerning autonomous driving and entertainment. Sony Entertainment has two divisions: Sony Pictures Entertainment, Sony Music Group (Sony Music Entertainment, Sony/ATV Music Publishing). Sony USA also previously owned and operated Sony Trans Com: a technology business that provided in-flight entertainment programming as well as video and audio playback equipment for the airline industry. Sony had purchased the business from Sundstrand Corp. in 1989 and subsequently sold it to Rockwell Collins in 2000. Sony Pictures Entertainment Inc. (SPE) is the television and film production/distribution unit of Sony. With 12.5% box office market share in 2011, the company was ranked third among movie studios. Its group sales in 2010 were US$7.2 billion. The company has produced many notable movie franchises, including "Spider-Man", "The Karate Kid" and "Men in Black". It has also produced the popular television game shows "Jeopardy!" and "Wheel of Fortune". Sony entered the television and film production market when it acquired Columbia Pictures Entertainment in 1989 for $3.4 billion. Columbia lives on in the Sony Pictures Motion Picture Group, a division of SPE which in turn owns Columbia Pictures and TriStar Pictures among other film production and distribution companies such as Screen Gems, Sony Pictures Classics, Sony Pictures Home Entertainment. SPE's television division is known as Sony Pictures Television. For the first several years of its existence, Sony Pictures Entertainment performed poorly, leading many to suspect the company would sell off the division. Sony Pictures Entertainment encountered controversy in the early 2000s. In July 2000, a marketing executive working for Sony Corporation created a fictitious film critic, David Manning, who gave consistently good reviews for releases from Sony subsidiary Columbia Pictures that generally received poor reviews amongst real critics. Sony later pulled the ads, suspended Manning's creator and his supervisor and paid fines to the state of Connecticut and to fans who saw the reviewed films in the US. In 2006 Sony started using ARccOS Protection on some of their film DVDs, but later issued a recall. In late 2014, Sony Pictures became the target of a hack attack from a clandestine group called Guardians of Peace, weeks before releasing the anti-North Korean comedy film "The Interview". Sony Music Entertainment (also known as SME or Sony Music) is the second-largest global recorded music company of the "big three" record companies and is controlled by Sony Corporation of America, the United States subsidiary of Japan's "Sony". In one of its largest-ever acquisitions, Sony purchased CBS Record Group in 1988 for US$2 billion. In the process, Sony partnered and gained the rights to the ATV catalogue of Michael Jackson, considered by the "Guinness Book of World Records" to be the most successful entertainer of all time. The acquisition of CBS Records provided the foundation for the formation of Sony Music Entertainment, which Sony established in 1991. In 2004, Sony entered into a joint venture with Bertelsmann AG, merging Sony Music Entertainment with Bertelsmann Music Group to create Sony BMG. In 2005, Sony BMG faced a copy protection scandal, because its music CDs had installed malware on users' computers that was posing a security risk to affected customers. In 2007, the company acquired Famous Music for US$370 million, gaining the rights to the catalogues of Eminem and Akon, among others. Sony bought out Bertelsmann's share in the company and formed a new Sony Music Entertainment in 2008. Since then, the company has undergone management changes. In January 1988, Sony acquired CBS Records and the 50% of CBS/Sony Group. In March 1988, four wholly owned subsidiaries were folded into CBS/Sony Group and the company was renamed as Sony Music Entertainment Japan Sony purchased digital music recognition company Gracenote for US$260 million in 2008. Tribune Media Company acquired Gracenote from Sony in 2014 for $170 million. Besides its record label, Sony operates other music businesses. In 1995, Sony merged its publisher with Michael Jackson's ATV Music Publishing, forming Sony/ATV Music Publishing. At the time, the publishing company was the second largest of its kind in the world. The company owns the publishing rights to over 4 million compositions, including The Beatles' Lennon-McCartney catalogue, Bob Dylan, Eminem, Lady Gaga, Sam Smith, Ed Sheeran, and Taylor Swift. In 2012, Sony/ATV then acquired a majority stake in EMI Music Publishing, making them the world's largest music publishing company. As of 2016, Sony owns all of Sony/ATV. Sony Financial Holdings is a holding company for Sony's financial services business. It owns and oversees the operation of Sony Life (in Japan and the Philippines), Sony Assurance, Sony Bank and Sony Bank Securities. The company is headquartered in Tokyo, Japan. Sony Financial accounts for half of Sony's global earnings. The unit proved the most profitable of Sony's businesses in fiscal year 2006, earning $1.7 billion in profit. Sony Financial's low fees have aided the unit's popularity while threatening Sony's premium brand name. Sony wants to contend with Apple and Samsung on mobile payments in Asia. Sony plans to use its contact-less payment technology to make ground in the public transportation industry across Asia. The system, known as FeliCa, relies on two forms of technologies to make it viable, either chips embedded in smartphones or plastic cards with chips embedded in them. Sony plans to implement this technology in train systems in Indonesia as early as Spring 2016. Sony is a "kabushiki gaisha" registered to the Tokyo Stock Exchange in Japan and the New York Stock Exchange for overseas trading. As of 30 September 2017, there are 484,812 shareholders and 1,264,649,260 shares issued. Most of these shares are held by foreign institutions and investors. Sony is one of Japan's largest corporations by revenue. It had revenues of ¥6.493 trillion in 2012. It also maintains large reserves of cash, with ¥895 billion on hand as of 2012. In May 2012, Sony's Market Capitalization were valued at about $15 billion. The company was immensely profitable throughout the 1990s and early 2000s, in part because of the success of its new PlayStation line. The company encountered financial difficulty in the mid- to late-2000s due to a number of factors: the global financial crisis, increased competition for PlayStation, and the devastating Japanese earthquake of 2011. The company faced three consecutive years of losses leading up to 2011. While noting the negative effects of intervening circumstances such as natural disasters and fluctuating currency exchange rates, the "Financial Times" criticized the company for its "lack of resilience" and "inability to gauge the economy." The newspaper voiced skepticism about Sony's revitalization efforts, given a lack of tangible results. In September 2000 Sony had a market capitalization of $100 billion; but by December 2011 it had plunged to $18 billion, reflecting falling prospects for Sony but also reflecting grossly inflated share prices of the 'Dot-com bubble' years. Net worth, as measured by stockholder equity, has steadily grown from $17.9 billion in March 2002 to $35.6 billion through December 2011. Earnings yield (inverse of the price to earnings ratio) has never been more than 5% and usually much less; thus Sony has always traded in over-priced ranges with the exception of the 2009 market bottom. On 9 December 2008, Sony Corporation announced that it would be cutting 8,000 jobs, dropping 8,000 contractors and reducing its global manufacturing sites by 10% to save $1.1 billion per year. In April 2012, Sony announced that it would reduce its workforce by 10,000 (6% of its employee base) as part of CEO Kaz Hirai's effort to get the company back into the black. This came after a loss of 520 billion yen (roughly US$6.36 billion) for fiscal 2012, the worst since the company was founded. Accumulation loss for the past four years was 919.32 billion-yen. Sony planned to increase its marketing expenses by 30% in 2012. 1,000 of the jobs cut come from the company's mobile phone unit's workforce. 700 jobs will be cut in the 2012–2013 fiscal year and the remaining 300 in the following fiscal year. In January 2013, Sony announced it was selling its US headquarters building for $1.1 billion to a consortium led by real estate developer The Chetrit Group. On 28 January 2014, Moody's Investors Services dropped Sony's credit rating to Ba1—"judged to have speculative elements and a significant credit risk"—saying that the company's "profitability is likely to remain weak and volatile." On 6 February 2014, Sony announced it would trim as many as 5,000 jobs as it attempts to sell its PC business and focus on mobile and tablets. In 2014, Sony South Africa closed its TV, Hi-Fi and camera divisions with the purpose of reconsidering its local distribution model and, in 2017, it returned facilitated by Premium Brand Distributors (Pty) Ltd. In November 2011, Sony was ranked 9th (jointly with Panasonic) in Greenpeace's Guide to Greener Electronics. This chart grades major electronics companies on their environmental work. The company scored 3.6/10, incurring a penalty point for comments it has made in opposition to energy efficiency standards in California. It also risks a further penalty point in future editions for being a member of trade associations that have commented against energy efficiency standards. Together with Philips, Sony receives the highest score for energy policy advocacy after calling on the EU to adopt an unconditional 30% reduction target for greenhouse gas emissions by 2020. Meanwhile, it receives full marks for the efficiency of its products. In June 2007, Sony ranked 14th on the Greenpeace guide. Sony fell from its earlier 11th-place ranking due to Greenpeace's claims that Sony had double standards in their waste policies. Since 1976, Sony has had an Environmental Conference. Sony's policies address their effects on global warming, the environment, and resources. They are taking steps to reduce the amount of greenhouse gases that they put out as well as regulating the products they get from their suppliers in a process that they call "green procurement". Sony has said that they have signed on to have about 75 percent of their Sony Building running on geothermal power. The "Sony Take Back Recycling Program" allow consumers to recycle the electronics products that they buy from Sony by taking them to eCycle (Recycling) drop-off points around the U.S. The company has also developed a biobattery that runs on sugars and carbohydrates that works similarly to the way living creatures work. This is the most powerful small biobattery to date. In 2000, Sony faced criticism for a document entitled "NGO Strategy" that was leaked to the press. The document involved the company's surveillance of environmental activists in an attempt to plan how to counter their movements. It specifically mentioned environmental groups that were trying to pass laws that held electronics-producing companies responsible for the cleanup of the toxic chemicals contained in their merchandise. Sony Corporation is actively involved in the EYE SEE project conducted by UNICEF. EYE SEE digital photography workshops have been run for children in Argentina, Tunisia, Mali, South Africa, Ethiopia, Madagascar, Rwanda, Liberia and Pakistan. Sony assists The South Africa Primary Education Support Initiative (SAPESI) through financial donations and children book donations to the South Africa Mobile Library Project. The Sony Canada Charitable Foundation (SCCF) is a non-profit organization which supports three key charities; the Make-A-Wish Canada, the United Way of Canada and the EarthDay and ECOKIDS program. After the 2011 Queensland floods and Victorian bushfires, Sony Music released benefit albums with money raised going to the Sony Foundation. You Can is the youth cancer program of Sony Foundation. Sony launched its Open Planet Ideas Crowdsourcing Project, in partnership with the World Wildlife Fund and the design group, IDEO. On the occasion of the 2014 World Cup in Brazil, Sony partnered with and launched the Street Football Stadium Project to support football-based educational programmes in local communities across Latin America and Brazil. More than 25 Street Stadiums were developed since the project's inception.
https://en.wikipedia.org/wiki?curid=26989
Social psychology In the field of psychology, social psychology is the scientific study of how the thoughts, feelings, and behaviors of individuals are influenced by the actual, imagined, and implied presence of others. In this definition, "scientific" refers to the empirical investigation using the scientific method, while the terms "thoughts", "feelings", and "behaviors" refer to the psychological variables that can be measured in humans. Moreover, the notion that the presence of others may be "imagined" or "implied" suggests that humans are to social influences even when alone, such as when watching videos, quietly appreciating art, or even sitting on the toilet. In such situations, people can be influenced to follow internalized cultural norms. Social psychologists typically explain human behavior as a result of the relation between mental state and social situation, studying the factors/conditions under which certain behavior, actions, and feelings occur. Social psychology, thus, is concerned with the way these feelings, thoughts, beliefs, intentions, and goals, are cognitively constructed and how these mental representations, in turn, influence our interactions with others. Traditionally, the emergence of this discipline bridged the gap between psychology and sociology. During the years immediately following World War II, there was frequent collaboration between psychologists and sociologists. The two disciplines, however, have become increasingly specialized and isolated from each other in recent years, with sociologists generally focusing on more macro features (e.g., social structure)—as the prefix of denotes a "societal" influence—whereas psychologists may be more concerned with more micro features. Nevertheless, sociological approaches to psychology remain an important counterpart to psychological research in this area. In addition to the split between psychology and sociology, there has been a somewhat less pronounced difference in emphasis between American and European social psychologists, as, the former traditionally have focused more on the individual, whereas the latter have generally paid more attention to group-level phenomena. Although older writings regarding social psychology have existed—such as those by Islamic philosopher Al-Farabi (aka Alpharabius)—the discipline of social psychology, as its modern-day definition, began in the United States at the beginning of the 20th century. By this time, however, the discipline itself had already developed a significant foundation. Following the 18th century, those in the emerging field of social psychology were concerned with developing concrete explanations for different aspects of human nature. They would attempt to discover concrete cause-and-effect relationships that explain the social interactions in the world around them. In order to do so, they believed that the scientific method, an empirically based scientific measure, could be applied to human behavior. The first published study in this field was an experiment in 1898 by Norman Triplett, on the phenomenon of social facilitation. During the 1930s, many Gestalt psychologists, most notably Kurt Lewin, fled to the United States from Nazi Germany. They would be instrumental in developing the field as an area separate from that of the dominant behavioral and psychoanalytic schools during that time. Social psychology would continue to maintain the legacy of the foundational interests in perception and cognition. As such, attitudes and small group phenomena were the most commonly studied topics in this era. During World War II, social psychologists were primarily engaged with studies of persuasion and propaganda for the U.S. military (see also psychological warfare). Following the war, researchers would become interested in a variety of social problems, including issues of gender and racial prejudice. Most notable, revealing, and of these were the shock experiments on obedience to authority conducted by Stanley Milgram. In the 1960s, there would be growing interest in topics such as cognitive dissonance, bystander intervention, and aggression. By the 1970s, however, social psychology in America had reached a crisis, as heated debates would emerge over: ethical concerns about laboratory experimentation; whether attitude could actually predict behavior; and how much science could really be done in a cultural context. This was also the time when a radical situationist approach came to challenge the relevance of self and personality in psychology. Throughout the 1980s and 1990s, social psychology reached a more mature level, especially in regard to theory and methodology. Now, careful ethical standards regulate research, and pluralistic and multicultural perspectives have emerged. Modern researchers are interested in many phenomena, though attribution, social cognition, and the self-concept are perhaps the greatest areas of growth in recent years. Social psychologists have also maintained their applied interests with contributions in the social psychology of health, education, law, and the workplace. In social psychology, "attitude" is defined as learned, global evaluations of a person, object, place, or issue that influence thought and action. In simpler terms, attitudes are basic expressions of approval and disapproval, favorability and unfavorability, or, as Bem (1970) suggests, likes and dislikes (e.g. enjoying chocolate ice cream, or endorsing the values of a particular political party.) In regard to attitudes, social psychologists have studied attitude formation; the structure of attitudes; attitude change; the function of attitudes; and the relationship between attitude and behavior. Because people are influenced by situation, general attitudes are not always good predictors of specific behavior, e.g. a person may value the environment, but, for a variety of reasons, not recycle a plastic bottle on a particular day. In recent times, research on attitudes has examined the distinction between traditional self-reported attitude measures and "implicit" or unconscious attitudes. Experiments using the "implicit-association test", for instance, have found that people often demonstrate implicit bias against other races, even when their explicit responses reveal equal mindedness. Likewise, one study found that explicit attitudes correlate with verbal behavior in interracial interactions, whereas implicit attitudes correlate with nonverbal behavior. One hypothesis on how attitudes are formed, first advanced in 1983 by Abraham Tesser, is that strong likes and dislikes are ingrained in our genetic make-up. Tesser speculates that individuals are disposed to hold certain strong attitudes as a result of inborn physical, sensory, and cognitive skills, temperament, and personality traits. Whatever disposition nature elects to give us, our most treasured attitudes are often formed as a result of exposure to attitude objects; our history of rewards and punishments; the attitude that our parents, friends, and enemies express; the social and cultural context in which we live; and other types of experiences we have. Obviously, attitudes are formed through the basic process of learning. Numerous studies have shown that people can form strong positive and negative attitudes toward neutral objects that are in some way linked to emotionally charged stimuli. Attitudes are also involved in several other areas of the discipline, such as conformity, interpersonal attraction, social perception, and prejudice. The topic of persuasion has received a great deal of attention in recent years. Persuasion is an active method of influence that attempts to guide people toward the adoption of an attitude, idea, or behavior by rational or emotive means. Persuasion relies on "appeals" rather than strong pressure or coercion. The process of persuasion has been founded to be influenced by numerous variables ("who" said "what" to "whom" and "how"), which generally fall into one of five major categories: Dual-process theories of persuasion (such as the elaboration likelihood model) maintain that the persuasive process is mediated by two separate routes; central and peripheral. The central route of persuasion is more fact-based and results in longer lasting change, but requires motivation to process. The peripheral route is more superficial and results in shorter lasting change, but does not require as much motivation to process. An example of a peripheral route of persuasion might be a politician using a flag lapel pin, smiling, and wearing a crisp, clean shirt. Notice that this does not require motivation to be persuasive, but should not last as long as persuasion based on the central route. If that politician were to outline exactly what they believed, and their previous voting record, this would be using the central route, and would result in longer lasting change, but would require a good deal of motivation to process. Social cognition is a growing area of social psychology that studies how people perceive, think about, and remember information about others. Much research rests on the assertion that people think about (other) people differently from non-social targets. This assertion is supported by the social cognitive deficits exhibited by people with Williams syndrome and autism. Person perception is the study of how people form impressions of others. The study of how people form beliefs about each other while interacting is known as interpersonal perception. A major research topic in social cognition is attribution. Attributions are the explanations we make for people's behavior, either our own behavior or the behavior of others. One element of attribution ascribes the locus of a behavior to either internal or external factors. An "internal", or dispositional, attribution assigns behavior to causes related to inner traits such as personality, disposition, character or ability. An "external", or situational, attribution involves situational elements, such as the weather. A second element of attribution ascribes the cause of behavior to either stable or unstable factors (whether the behavior will be repeated or changed under similar circumstances). Finally, we also attribute causes of behavior to either controllable or uncontrollable factors: how much control one has over the situation at hand. Numerous biases in the attribution process have been discovered. For instance, the fundamental attribution error is the tendency to make dispositional attributions for behavior, overestimating the influence of personality and underestimating the influence of situations. The actor-observer difference is a refinement of this bias, the tendency to make dispositional attributions for other people's behavior and situational attributions for our own. The self-serving bias is the tendency to attribute dispositional causes for successes, and situational causes for failure, particularly when self-esteem is threatened. This leads to assuming one's successes are from innate traits, and one's failures are due to situations, including other people. Other ways people protect their self-esteem are by believing in a just world, blaming victims for their suffering, and making defensive attributions, which explain our behavior in ways which defend us from feelings of vulnerability and mortality. Researchers have found that mildly depressed individuals often lack this bias and actually have more realistic perceptions of reality (as measured by the opinions of others). Heuristics are cognitive shortcuts. Instead of weighing all the evidence when making a decision, people rely on heuristics to save time and energy. The availability heuristic occurs when people estimate the probability of an outcome based on how easy that outcome is to imagine. As such, vivid or highly memorable possibilities will be perceived as more likely than those that are harder to picture or are difficult to understand, resulting in a corresponding cognitive bias. The representativeness heuristic is a shortcut people use to categorize something based on how similar it is to a prototype they know of. Numerous other biases have been found by social cognition researchers. The hindsight bias is a false memory of having predicted events, or an exaggeration of actual predictions, after becoming aware of the outcome. The confirmation bias is a type of bias leading to the tendency to search for, or interpret information in a way that confirms one's preconceptions. Another key concept in social cognition is the assumption that reality is too complex to easily discern. As a result, we tend to see the world according to simplified schemas or images of reality. Schemas are generalized mental representations that organize knowledge and guide information processing. Schemas often operate automatically and unintentionally, and can lead to biases in perception and memory. Expectations from schemas may lead us to see something that is not there. One experiment found that people are more likely to misperceive a weapon in the hands of a black man than a white man. This type of schema is actually a stereotype, a generalized set of beliefs about a particular group of people (when incorrect, an ultimate attribution error). Stereotypes are often related to negative or preferential attitudes (prejudice) and behavior (discrimination). Schemas for behaviors (e.g., going to a restaurant, doing laundry) are known as "scripts". Self-concept is a term referring to the whole sum of beliefs that people have about themselves. However, what specifically does self-concept consist of? According to Hazel Markus (1977), the self-concept is made up of cognitive molecules called self-schemas—beliefs that people have about themselves that guide the processing of self-reliant information. For example, an athlete at a university would have multiple selves that would process different information pertinent to each self: the student would be one "self," who would process information pertinent to a student (taking notes in class, completing a homework assignment, etc.); the athlete would be the "self" who processes information about things related to being an athlete (recognizing an incoming pass, aiming a shot, etc.). These "selves" are part of one's identity and the self-reliant information is the information that relies on the proper "self" to process and react on it. If a "self" is not part of one's identity, then it is much more difficult for one to react. For example, a civilian may not know how to handle a hostile threat as a trained Marine would. The Marine contains a "self" that would enable him/her to process the information about the hostile threat and react accordingly, whereas a civilian may not contain that self, disabling them from properly processing the information from the hostile threat and, furthermore, debilitating them from acting accordingly. Self-schemas are to an individual's total self–concept as a hypothesis is to a theory, or a book is to a library. A good example is the body weight self-schema; people who regard themselves as over or underweight, or for those whom body image is a significant self-concept aspect, are considered "schematics" with respect to weight. For these people a range of otherwise mundane events – grocery shopping, new clothes, eating out, or going to the beach – can trigger thoughts about the self. In contrast, people who do not regard their weight as an important part of their lives are "a-schematic" on that attribute. It is rather clear that the self is a special object of our attention. Whether one is mentally focused on a memory, a conversation, a foul smell, the song that is stuck in one's head, or this sentence, consciousness is like a spotlight. This spotlight can shine on only one object at a time, but it can switch rapidly from one object to another and process the information out of awareness. In this spotlight the self is front and center: things relating to the self have the spotlight more often. The "ABCs" of self are: "Affective forecasting" is the process of predicting how one would feel in response to future emotional events. Studies done in 2003 by Timothy Wilson and Daniel Gilbert have shown that people overestimate the strength of reaction to anticipated positive and negative life events that they actually feel when the event does occur. There are many theories on the perception of our own behavior. Daryl Bem's (1972) "self-perception theory" claims that when internal cues are difficult to interpret, people gain self-insight by observing their own behavior. Leon Festinger's (1954) "social comparison theory" is that people evaluate their own abilities and opinions by comparing themselves to others when they are uncertain of their own ability or opinions. There is also the "facial feedback hypothesis": changes in facial expression can lead to corresponding changes in emotion. The fields of social psychology and personality have merged over the years, and social psychologists have developed an interest in self-related phenomena. In contrast with traditional personality theory, however, social psychologists place a greater emphasis on cognitions than on traits. Much research focuses on the self-concept, which is a person's understanding of their self. The self-concept is often divided into a cognitive component, known as the "self-schema", and an evaluative component, the "self-esteem". The need to maintain a healthy self-esteem is recognized as a central human motivation in the field of social psychology. Self-efficacy beliefs are associated with the self-schema. These are expectations that performance on some task will be effective and successful. Social psychologists also study such self-related processes as self-control and self-presentation. People develop their self-concepts by varied means, including introspection, feedback from others, self-perception, and social comparison. By comparing themselves to relevant others, people gain information about themselves, and they make inferences that are relevant to self-esteem. Social comparisons can be either "upward" or "downward," that is, comparisons to people who are either higher in status or ability, or lower in status or ability. Downward comparisons are often made in order to elevate self-esteem. Self-perception is a specialized form of attribution that involves making inferences about oneself after observing one's own behavior. Psychologists have found that too many extrinsic rewards (e.g. money) tend to reduce intrinsic motivation through the self-perception process, a phenomenon known as overjustification. People's attention is directed to the reward and they lose interest in the task when the reward is no longer offered. This is an important exception to reinforcement theory. Social influence is an overarching term given to describe the persuasive effects people have on each other. It is seen as a fundamental value in social psychology and overlaps considerably with research on attitudes and persuasion. The three main areas of social influence include: conformity, compliance, and obedience. Social influence is also closely related to the study of group dynamics, as most principles of influence are strongest when they take place in social groups. The first major area of social influence is conformity. Conformity is defined as the tendency to act or think like other members of a group. The identity of members within a group, i.e. status, similarity, expertise, as well as cohesion, prior commitment, and accountability to the group help to determine the level of conformity of an individual. Individual variation among group members plays a key role in the dynamic of how willing people will be to conform. Conformity is usually viewed as a negative tendency in American culture, but a certain amount of conformity is adaptive in some situations, as is nonconformity in other situations. The second major area of social influence research is compliance. Compliance refers to any change in behavior that is due to a request or suggestion from another person. The foot-in-the-door technique is a compliance method in which the persuader requests a small favor and then follows up with requesting a larger favor, e.g., asking for the time and then asking for ten dollars. A related trick is the bait and switch. The third major form of social influence is obedience; this is a change in behavior that is the result of a direct order or command from another person. Obedience as a form of compliance was dramatically highlighted by the Milgram study, wherein people were ready to administer shocks to a person in distress on a researcher's command. An unusual kind of social influence is the self-fulfilling prophecy. This is a prediction that, in being made, actually causes itself to become true. For example, in the stock market, if it is widely believed that a crash is imminent, investors may lose confidence, sell most of their stock, and thus actually cause the crash. Similarly, people may expect hostility in others and actually induce this hostility by their own behavior. Psychologist have spent decades studying the power of social influence, and the way in which it manipulates people's opinions and behavior. Specifically, social influence refers to the way in which individuals change their ideas and actions to meet the demands of a social group, received authority, social role or a minority within a group wielding influence over the majority. No matter if you are student, teacher, doctor, lawyer or entrepreneur, you will encounter some type of social influence. A group can be defined as two or more individuals that are connected to each another by social relationships. Groups tend to interact, influence each other, and share a common identity. They have a number of emergent qualities that distinguish them from aggregates: Temporary groups and aggregates share few or none of these features, and do not qualify as true social groups. People waiting in line to get on a bus, for example, do not constitute a group. Groups are important not only because they offer social support, resources, and a feeling of belonging, but because they supplement an individual's self-concept. To a large extent, humans define themselves by the group memberships which form their social identity. The shared social identity of individuals within a group influences intergroup behavior, the way in which groups behave towards and perceive each other. These perceptions and behaviors in turn define the social identity of individuals within the interacting groups. The tendency to define oneself by membership in a group may lead to intergroup discrimination, which involves favorable perceptions and behaviors directed towards the in-group, but negative perceptions and behaviors directed towards the out-group. On the other hand, such discrimination and segregation may sometimes exist partly to facilitate a diversity which strengthens society. Intergroup discrimination leads to prejudice and stereotyping, while the processes of social facilitation and group polarization encourage extreme behaviors towards the out-group. Groups often moderate and improve decision making, and are frequently relied upon for these benefits, such as in committees and juries. A number of group biases, however, can interfere with effective decision making. For example, group polarization, formerly known as the "risky shift," occurs when people polarize their views in a more extreme direction after group discussion. More problematic is the phenomenon of groupthink. This is a collective thinking defect that is characterized by a premature consensus or an incorrect assumption of consensus, caused by members of a group failing to promote views which are not consistent with the views of other members. Groupthink occurs in a variety of situations, including isolation of a group and the presence of a highly directive leader. Janis offered the 1961 Bay of Pigs Invasion as a historical case of groupthink. Groups also affect performance and productivity. Social facilitation, for example, is a tendency to work harder and faster in the presence of others. Social facilitation increases the "dominant response"s likelihood, which tends to improve performance on simple tasks and reduce it on complex tasks. In contrast, social loafing is the tendency of individuals to slack off when working in a group. Social loafing is common when the task is considered unimportant and individual contributions are not easy to see. Social psychologists study group-related (collective) phenomena such as the behavior of crowds. An important concept in this area is deindividuation, a reduced state of self-awareness that can be caused by feelings of anonymity. Deindividuation is associated with uninhibited and sometimes dangerous behavior. It is common in crowds and mobs, but it can also be caused by a disguise, a uniform, alcohol, dark environments, or online anonymity. A major area in the study of people's relations to each other is interpersonal attraction. This refers to all forces that lead people to like each other, establish relationships, and (in some cases) fall in love. Several general principles of attraction have been discovered by social psychologists, but many still continue to experiment and do research to find out more. One of the most important factors in interpersonal attraction is how similar two particular people are. The more similar two people are in general attitudes, backgrounds, environments, worldviews, and other traits, the more probable an attraction is possible. Physical attractiveness is an important element of romantic relationships, particularly in the early stages characterized by high levels of passion. Later on, similarity and other compatibility factors become more important, and the type of love people experience shifts from "passionate" to "companionate". Robert Sternberg (1986) has suggested that there are actually three components of love: intimacy, passion, and commitment. When two (or more) people experience all three, they are said to be in a state of consummate love. According to social exchange theory, relationships are based on rational choice and cost-benefit analysis. If one partner's costs begin to outweigh their benefits, that person may leave the relationship, especially if there are good alternatives available. This theory is similar to the minimax principle proposed by mathematicians and economists (despite the fact that human relationships are not zero-sum games). With time, long-term relationships tend to become communal rather than simply based on exchange. Social psychology is an empirical science that attempts to answer questions about human behavior by testing hypotheses, both in the laboratory and in the field. Careful attention to sampling, research design, and statistical analysis is important; results are published in peer reviewed journals such as the "Journal of Experimental Social Psychology", "Personality and Social Psychology Bulletin" and the "Journal of Personality and Social Psychology". Social psychology studies also appear in general science journals such as "Psychological Science" and "Science". Experimental methods involve the researcher altering a variable in the environment and measuring the effect on another variable. An example would be allowing two groups of children to play violent or nonviolent videogames, and then observing their subsequent level of aggression during free-play period. A valid experiment is controlled and uses random assignment. Correlational methods examine the statistical association between two naturally occurring variables. For example, one could correlate the amount of violent television children watch at home with the number of violent incidents the children participate in at school. Note that this study would "not" prove that violent TV causes aggression in children: it is quite possible that aggressive children choose to watch more violent TV. Observational methods are purely descriptive and include naturalistic observation, "contrived" observation, participant observation, and archival analysis. These are less common in social psychology but are sometimes used when first investigating a phenomenon. An example would be to unobtrusively observe children on a playground (with a videocamera, perhaps) and record the number and types of aggressive actions displayed. Whenever possible, social psychologists rely on controlled experimentation. Controlled experiments require the manipulation of one or more independent variables in order to examine the effect on a dependent variable. Experiments are useful in social psychology because they are high in internal validity, meaning that they are free from the influence of confounding or extraneous variables, and so are more likely to accurately indicate a causal relationship. However, the small samples used in controlled experiments are typically low in external validity, or the degree to which the results can be generalized to the larger population. There is usually a trade-off between experimental control (internal validity) and being able to generalize to the population (external validity). Because it is usually impossible to test everyone, research tends to be conducted on a sample of persons from the wider population. Social psychologists frequently use survey research when they are interested in results that are high in external validity. Surveys use various forms of random sampling to obtain a sample of respondents that are representative of a population. This type of research is usually descriptive or correlational because there is no experimental control over variables. Some psychologists have raised concerns about social psychological research for relying too heavily on studies conducted on university undergraduates in academic settings, or participants from crowdsourcing labor markets such as Amazon Mechanical Turk. In a study by David O. Sears (1986), over 70% of experiments used North American undergraduates as subjects, a subset of the population that are unrepresentative of the population as a whole. Regardless of which method has been chosen to be used, the results are of high importance. Results need to be used to evaluate the hypothesis of the research that is done. These results should either confirm or reject the original hypothesis that was predicted. There are two different types of testing social psychologists use in order to test their results. Statistics and probability testing define a significant finding that can be as low as 5% or less, likely to be due to chance. Replications are important, to ensure that the result is valid and not due to chance, or some feature of a particular sample. False positive conclusions, often resulting from the pressure to publish or the author's own confirmation bias, are a hazard in the field. The Asch conformity experiments demonstrated the power of conformity in small groups with a line length estimation task that was designed to be extremely easy. In well over a third of the trials, participants conformed to the majority, who had been instructed to provide incorrect answers, even though the majority judgment was clearly wrong. Seventy-five percent of the participants conformed at least once during the experiment. Additional manipulations to the experiment showed participant conformity decreased when at least one other individual failed to conform, but increased when the individual began conforming or withdrew from the experiment. Also, participant conformity increased substantially as the number of incorrect individuals increased from one to three, and remained high as the incorrect majority grew. Participants with three incorrect opponents made mistakes 31.8% of the time, while those with one or two incorrect opponents made mistakes only 3.6% and 13.6% of the time, respectively. In Leon Festinger's cognitive dissonance experiment, participants were asked to perform a boring task. They were divided into 2 groups and given two different pay scales. At the study's end, some participants were paid $1 to say that they enjoyed the task and another group of participants was paid $20 to say the same lie. The first group ($1) later reported liking the task better than the second group ($20). Festinger's explanation was that for people in the first group being paid only $1 is not sufficient incentive for lying and those who were paid $1 experienced dissonance. They could only overcome that dissonance by justifying their lies by changing their previously unfavorable attitudes about the task. Being paid $20 provides a reason for doing the boring task, therefore no dissonance. One of the most notable experiments in social psychology was the Milgram experiment, which studied how far people would go to obey an authority figure. Following the events of The Holocaust in World War II, the experiment showed that (most) normal American citizens were capable of following orders from an authority even when they believed they were causing an innocent person to suffer. In the Stanford prison study, by Philip Zimbardo, a simulated exercise between student prisoners and guards showed how far people would follow an adopted role. In just a few days, the "guards" became brutal and cruel, and the prisoners became miserable and compliant. This was initially argued to be an important demonstration of the power of the immediate social situation and its capacity to overwhelm normal personality traits. However, to this day, it remains a matter of contention what conclusions may be drawn from this study. For example, it has been pointed out that participant self-selection may have affected the participants' behaviour, and that the participants' personality influenced their reactions in a variety of ways, including how long they chose to remain in the study. One of the most concerted empirical revisitations of the themes raised by Zimbardo came with the 2002 BBC prison study. Muzafer Sherif's "robbers' cave" study divided boys into two competing groups to explore how much hostility and aggression would emerge. Sherif's explanation of the results became known as realistic group conflict theory, because the intergroup conflict was induced through competition over resources. Inducing cooperation and superordinate goals later reversed this effect. Albert Bandura's Bobo doll experiment demonstrated how aggression is learned by imitation. This set of studies fueled debates regarding media violence which continue to be waged among scholars. The goal of social psychology is to understand cognition and behavior as they naturally occur in a social context, but the very act of observing people can influence and alter their behavior. For this reason, many social psychology experiments utilize deception to conceal or distort certain aspects of the study. Deception may include false cover stories, false participants (known as confederates or stooges), false feedback given to the participants, and so on. The practice of deception has been challenged by some psychologists who maintain that deception under any circumstances is unethical, and that other research strategies (e.g., role-playing) should be used instead. Unfortunately, research has shown that role-playing studies do not produce the same results as deception studies and this has cast doubt on their validity. In addition to deception, experimenters have at times put people into potentially uncomfortable or embarrassing situations (e.g., the Milgram experiment and Stanford prison experiment), and this has also been criticized for ethical reasons. To protect the rights and well-being of research participants, and at the same time discover meaningful results and insights into human behavior, virtually all social psychology research must pass an ethical review process. At most colleges and universities, this is conducted by an ethics committee or Institutional Review Board. This group examines the proposed research to make sure that no harm is likely to be done to the participants, and that the study's benefits outweigh any possible risks or discomforts to people taking part in the study. Furthermore, a process of informed consent is often used to make sure that volunteers know what will happen in the experiment and understand that they are allowed to quit the experiment at any time. A debriefing is typically done at the experiment's conclusion in order to reveal any deceptions used and generally make sure that the participants are unharmed by the procedures. Today, most research in social psychology involves no more risk of harm than can be expected from routine psychological testing or normal daily activities. Social Psychology plays a key role in a child's development. During this time, teens are faced with many issues and decisions that can impact a teen's social development. They are faced with self esteem issues, peer pressure, drugs, alcohol, tobacco, sex, social media and more. Psychologists today are not fully aware of the effect of social media. Social media is worldwide, so one can be influenced by something they will never encounter in real life. In 2019, social media had become the single most important activity in adolescents and even some older adults lives. Social psychology has recently found itself at the center of a "replication crisis" due to some research findings proving difficult to replicate. Replication failures are not unique to social psychology and are found in all fields of science. However, several factors have combined to put social psychology at the center of the current controversy. Firstly, questionable research practices (QRP) have been identified as common in the field. Such practices, while not necessarily intentionally fraudulent, involve converting undesired statistical outcomes into desired outcomes via the manipulation of statistical analyses, sample size or data management, typically to convert non-significant findings into significant ones. Some studies have suggested that at least mild versions of QRP are highly prevalent. One of the critics of Daryl Bem in the feeling the future controversy has suggested that the evidence for precognition in this study could (at least in part) be attributed to QRP. Secondly, social psychology has found itself at the center of several recent scandals involving outright fraudulent research. Most notably the admitted data fabrication by Diederik Stapel as well as allegations against others. However, most scholars acknowledge that fraud is, perhaps, the lesser contribution to replication crises. Third, several effects in social psychology have been found to be difficult to replicate even before the current replication crisis. For example, the scientific journal "Judgment and Decision Making" has published several studies over the years that fail to provide support for the unconscious thought theory. Replications appear particularly difficult when research trials are pre-registered and conducted by research groups not highly invested in the theory under questioning. These three elements together have resulted in renewed attention for replication supported by Daniel Kahneman. Scrutiny of many effects have shown that several core beliefs are hard to replicate. A 2014 special edition of "Social Psychology" focused on replication studies and a number of previously held beliefs were found to be difficult to replicate. Likewise, a 2012 special edition of "Perspectives on Psychological Science" also focused on issues ranging from publication bias to null-aversion that contribute to the replication crises in psychology. It is important to note that this replication crisis does not mean that social psychology is unscientific. Rather this process is a healthy if sometimes acrimonious part of the scientific process in which old ideas or those that cannot withstand careful scrutiny are pruned. The consequence is that some areas of social psychology once considered solid, such as social priming, have come under increased scrutiny due to failed replications.
https://en.wikipedia.org/wiki?curid=26990
Suleiman the Magnificent Suleiman I ( '; or '; 6 November 1494 – 6 September 1566), commonly known as Suleiman the Magnificent in the West and Kanunî Sultan Süleyman (; "Suleiman the Lawgiver") in his realm, was the tenth and longest-reigning Sultan of the Ottoman Empire from 1520 until his death in 1566. Under his administration, the Ottoman state ruled over at least 25 million people. Suleiman succeeded his father as sultan in September 1520 and began his reign with campaigns against the Christian powers in central Europe and the Mediterranean. Belgrade fell to him in 1521 and Rhodes, long under the rule of the Knights of St. John, in 1522–23. At Mohács, in August 1526, Suleiman broke the military strength of Hungary, with the Hungarian king Louis II losing his life in the battle. Suleiman became a prominent monarch of 16th-century Europe, presiding over the apex of the Ottoman Empire's economic, military and political power. Suleiman personally led Ottoman armies in conquering the Christian strongholds of Belgrade and Rhodes as well as most of Hungary before his conquests were checked at the Siege of Vienna in 1529. He annexed much of the Middle East in his conflict with the Safavids and large areas of North Africa as far west as Algeria. Under his rule, the Ottoman fleet dominated the seas from the Mediterranean to the Red Sea and through the Persian Gulf. At the helm of an expanding empire, Suleiman personally instituted major legislative changes relating to society, education, taxation and criminal law. His reforms, carried out in conjunction with the empire's chief judicial official Ebussuud Efendi, harmonized the relationship between the two forms of Ottoman law: sultanic (Kanun) and religious (Sharia). He was a distinguished poet and goldsmith; he also became a great patron of culture, overseeing the "Golden" age of the Ottoman Empire in its artistic, literary and architectural development. Breaking with Ottoman tradition, Suleiman married Hürrem Sultan, a woman from his harem, an Orthodox Christian of Ruthenian origin who converted to Islam, and who became famous in the West by the name Roxelana, purportedly due to her red hair. Their son Selim II succeeded Suleiman following his death in 1566 after 46 years of rule. Suleiman's other potential heirs, Mehmed and Mustafa, had died; the former had died from smallpox, and the latter had been strangled to death 13 years earlier at the sultan's order. His other son Bayezid was executed in 1561 on Suleiman's orders, along with Bayezid's four sons, after a rebellion. Although scholars no longer believe that the empire declined after his death, the end of Suleiman's reign is still frequently characterized as a watershed in Ottoman history. In the decades after Suleiman, the empire began to experience significant political, institutional, and economic changes, a phenomenon often referred to as the Transformation of the Ottoman Empire. Suleiman the Magnificent ( "Muḥteşem Süleymān"), as he was known in the West, was also called Suleiman the First ( "Sulṭān Süleymān-ı Evvel"), and Suleiman the Lawgiver ( "Ḳānūnī Sulṭān Süleymān") for his reform of the Ottoman legal system. It is unclear when exactly the term "Kanunî" (the Lawgiver) first came to be used as an epithet for Suleiman. It is entirely absent from sixteenth and seventeenth-century Ottoman sources, and may date from the early 18th century. Suleiman was born in Trabzon along the east coast of the Black Sea to Şehzade Selim (later Selim I), probably on 6 November 1494, although this date is not known with absolute certainty. His mother was Hafsa Sultan, a convert to Islam of unknown origins, who died in 1534. At the age of seven, Suleiman was sent to study science, history, literature, theology and military tactics in the schools of the imperial Topkapı Palace in Constantinople (modern Istanbul). As a young man, he befriended Pargalı Ibrahim, a slave who later became one of his most trusted advisers (but who was later executed on Suleiman's orders). At age seventeen, he was appointed as the governor of first Kaffa (Theodosia), then Manisa, with a brief tenure at Edirne. Upon the death of his father, Selim I (r. 1512–1520), Suleiman entered Constantinople and ascended to the throne as the tenth Ottoman Sultan. An early description of Suleiman, a few weeks following his accession, was provided by the Venetian envoy Bartolomeo Contarini: Upon succeeding his father, Suleiman began a series of military conquests, eventually suppressing a revolt led by the Ottoman-appointed governor of Damascus in 1521. Suleiman soon made preparations for the conquest of Belgrade from the Kingdom of Hungary—something his great-grandfather Mehmed II had failed to achieve because of John Hunyadi's strong defense in the region. Its capture was vital in removing the Hungarians and Croats who, following the defeats of the Albanians, Bosniaks, Bulgarians, Byzantines and the Serbs, remained the only formidable force who could block further Ottoman gains in Europe. Suleiman encircled Belgrade and began a series of heavy bombardments from an island in the Danube. Belgrade, with a garrison of only 700 men, and receiving no aid from Hungary, fell in August 1521. The road to Hungary and Austria lay open, but Suleiman turned his attention instead to the Eastern Mediterranean island of Rhodes, the home base of the Knights Hospitaller. Suleiman built a large fortification, Marmaris Castle, that served as a base for the Ottoman Navy. Following the five-month Siege of Rhodes (1522), Rhodes capitulated and Suleiman allowed the Knights of Rhodes to depart. The conquest of the island cost the Ottomans 50,000 to 60,000 dead from battle and sickness (Christian claims went as high as 64,000 Ottoman battle deaths and 50,000 disease deaths). As relations between Hungary and the Ottoman Empire deteriorated, Suleiman resumed his campaign in Central Europe, and on 29 August 1526 he defeated Louis II of Hungary (1506–1526) at the Battle of Mohács. Upon encountering the lifeless body of King Louis, Suleiman is said to have lamented: "I came indeed in arms against him; but it was not my wish that he should be thus cut off before he scarcely tasted the sweets of life and royalty." While Suleiman was campaigning in Hungary, Turkmen tribes in central Anatolia (in Cilicia) revolted under the leadership of Kalender Çelebi. Some Hungarian nobles proposed that Ferdinand, who was the ruler of neighboring Austria and tied to Louis II's family by marriage, be King of Hungary, citing previous agreements that the Habsburgs would take the Hungarian throne if Louis died without heirs. However, other nobles turned to the nobleman Ioan Zápolya, who was being supported by Suleiman. Under Charles V and his brother Ferdinand I, the Habsburgs reoccupied Buda and took possession of Hungary. Reacting in 1529, Suleiman marched through the valley of the Danube and regained control of Buda; in the following autumn, his forces laid siege to Vienna. This was to be the Ottoman Empire's most ambitious expedition and the apogee of its drive to the West. With a reinforced garrison of 16,000  men, the Austrians inflicted the first defeat on Suleiman, sowing the seeds of a bitter Ottoman–Habsburg rivalry that lasted until the 20th century. His second attempt to conquer Vienna failed in 1532, as Ottoman forces were delayed by the siege of Güns and failed to reach Vienna. In both cases, the Ottoman army was plagued by bad weather, forcing them to leave behind essential siege equipment, and was hobbled by overstretched supply lines. By the 1540s a renewal of the conflict in Hungary presented Suleiman with the opportunity to avenge the defeat suffered at Vienna. In 1541, the Habsburgs attempted to lay siege to Buda but were repulsed, and more Habsburg fortresses were captured by the Ottomans in two consecutive campaigns in 1541 and 1544 as a result, Ferdinand and Charles were forced to conclude a humiliating five-year treaty with Suleiman. Ferdinand renounced his claim to the Kingdom of Hungary and was forced to pay a fixed yearly sum to the Sultan for the Hungarian lands he continued to control. Of more symbolic importance, the treaty referred to Charles V not as 'Emperor' but as the 'King of Spain', leading Suleiman to identify as the true 'Caesar'. As Suleiman stabilized his European frontiers, he now turned his attention to the ever-present threat posed by the Shi'a Safavid dynasty of Persia. Two events in particular were to precipitate a recurrence of tensions. First, Shah Tahmasp had the Baghdad governor loyal to Suleiman killed and replaced with an adherent of the Shah, and second, the governor of Bitlis had defected and sworn allegiance to the Safavids. As a result, in 1533, Suleiman ordered his Grand Vizier Pargalı Ibrahim Pasha to lead an army into eastern Asia Minor where he retook Bitlis and occupied Tabriz without resistance. Having joined Ibrahim in 1534, Suleiman made a push towards Persia, only to find the Shah sacrificing territory instead of facing a pitched battle, resorting to harassment of the Ottoman army as it proceeded along the harsh interior. When in the following year Suleiman made a grand entrance into Baghdad, he greatly enhanced his prestige by restoring the tomb of Abu Hanifa, the founder of the Hanafi school of Islamic law to which the Ottomans adhered. Attempting to defeat the Shah once and for all, Suleiman embarked upon a second campaign in 1548–1549. As in the previous attempt, Tahmasp avoided confrontation with the Ottoman army and instead chose to retreat, using scorched earth tactics in the process and exposing the Ottoman army to the harsh winter of the Caucasus. Suleiman abandoned the campaign with temporary Ottoman gains in Tabriz and the Urmia region, a lasting presence in the province of Van, control of the western half of Azerbaijan and some forts in Georgia. In 1553 Suleiman began his third and final campaign against the Shah. Having initially lost territories in Erzurum to the Shah's son, Suleiman retaliated by recapturing Erzurum, crossing the Upper Euphrates and laying waste to parts of Persia. The Shah's army continued its strategy of avoiding the Ottomans, leading to a stalemate from which neither army made any significant gain. In 1555, a settlement known as the Peace of Amasya was signed which defined the borders of the two empires. By this treaty, Armenia and Georgia were divided equally between the two, with Western Armenia, western Kurdistan, and western Georgia (incl. western Samtskhe) falling in Ottoman hands while Eastern Armenia, eastern Kurdistan, and eastern Georgia (incl. eastern Samtskhe) stayed in Safavid hands. The Ottoman Empire obtained most of Iraq, including Baghdad, which gave them access to the Persian Gulf, while the Persians retained their former capital Tabriz and all their other northwestern territories in the Caucasus and as they were prior to the wars, such as Dagestan and all of what is now Azerbaijan. Ottoman ships had been sailing in the Indian Ocean since the year 1518. Ottoman admirals such as Hadim Suleiman Pasha, Seydi Ali Reis and Kurtoğlu Hızır Reis are known to have voyaged to the Mughal imperial ports of Thatta, Surat and Janjira. The Mughal Emperor Akbar the Great himself is known to have exchanged six documents with Suleiman the Magnificent. Suleiman led several naval campaigns against the Portuguese in an attempt to remove them and reestablish trade with the Mughal Empire. Aden in Yemen was captured by the Ottomans in 1538, in order to provide an Ottoman base for raids against Portuguese possessions on the western coast of the Mughal Empire. Sailing on, the Ottomans failed against the Portuguese at the Siege of Diu in September 1538, but then returned to Aden, where they fortified the city with 100 pieces of artillery. From this base, Sulayman Pasha managed to take control of the whole country of Yemen, also taking Sana'a. With its strong control of the Red Sea, Suleiman successfully managed to dispute control of the trade routes to the Portuguese and maintained a significant level of trade with the Mughal Empire throughout the 16th century. From 1526 till 1543, Suleiman stationed over 900 Turkish soldiers to fight alongside the Somali Adal Sultanate led by Ahmad ibn Ibrahim al-Ghazi during the Conquest of Abyssinia. After the first Ajuran-Portuguese war, the Ottoman Empire would in 1559 absorb the weakened Adal Sultanate into its domain. This expansion fathered Ottoman rule in Somalia and the Horn of Africa. This also increased its influence in the Indian Ocean to compete with the Portuguese Empire with its close ally, the Ajuran Empire. In 1564, Suleiman received an embassy from Aceh (a sultanate on Sumatra, in modern Indonesia), requesting Ottoman support against the Portuguese. As a result, an Ottoman expedition to Aceh was launched, which was able to provide extensive military support to the Acehnese. The discovery of new maritime trade routes by Western European states allowed them to avoid the Ottoman trade monopoly. The Portuguese discovery of the Cape of Good Hope in 1488 initiated a series of Ottoman-Portuguese naval wars in the Ocean throughout the 16th century. The Ajuran Sultanate allied with the Ottomans defied the Portuguese economic monopoly in the Indian Ocean by employing a new coinage which followed the Ottoman pattern, thus proclaiming an attitude of economic independence in regard to the Portuguese. Having consolidated his conquests on land, Suleiman was greeted with the news that the fortress of Koroni in Morea (the modern Peloponnese, peninsular Greece) had been lost to Charles V's admiral, Andrea Doria. The presence of the Spanish in the Eastern Mediterranean concerned Suleiman, who saw it as an early indication of Charles V's intention to rival Ottoman dominance in the region. Recognizing the need to reassert naval preeminence in the Mediterranean, Suleiman appointed an exceptional naval commander in the form of Khair ad Din, known to Europeans as Barbarossa. Once appointed admiral-in-chief, Barbarossa was charged with rebuilding the Ottoman fleet. In 1535, Charles V led a Holy League of 27,000 soldiers (10,000 Spaniards, 8,000 Italians, 8,000 Germans, and 700 Knights of St. John) to victory against the Ottomans at Tunis, which together with the war against Venice the following year, led Suleiman to accept proposals from Francis I of France to form an alliance against Charles. Huge Muslim territories in North Africa were annexed. The piracy carried on thereafter by the Barbary pirates of North Africa can be seen in the context of the wars against Spain. In 1542, facing a common Habsburg enemy, Francis I sought to renew the Franco-Ottoman alliance. In early 1542, Polin successfully negotiated the details of the alliance, with the Ottoman Empire promising to send 60,000 troops against the territories of the German king Ferdinand, as well as 150 galleys against Charles, while France promised to attack Flanders, harass the coasts of Spain with a naval force, and send 40 galleys to assist the Turks for operations in the Levant. Elsewhere in the Mediterranean, when the Knights Hospitallers were re-established as the Knights of Malta in 1530, their actions against Muslim navies quickly drew the ire of the Ottomans, who assembled another massive army in order to dislodge the Knights from Malta. The Ottomans invaded Malta in 1565, undertaking the Great Siege of Malta, which began on 18 May and lasted until 8 September, and is portrayed vividly in the frescoes of Matteo Perez d'Aleccio in the Hall of St. Michael and St. George. At first it seemed that this would be a repeat of the battle on Rhodes, with most of Malta's cities destroyed and half the Knights killed in battle; but a relief force from Spain entered the battle, resulting in the loss of 10,000 Ottoman troops and the victory of the local Maltese citizenry. While Sultan Suleiman was known as "the Magnificent" in the West, he was always "Kanuni" Suleiman or "The Lawgiver" () to his Ottoman subjects. The overriding law of the empire was the Shari'ah, or Sacred Law, which as the divine law of Islam was outside of the Sultan's powers to change. Yet an area of distinct law known as the "Kanuns" (, canonical legislation) was dependent on Suleiman's will alone, covering areas such as criminal law, land tenure and taxation. He collected all the judgments that had been issued by the nine Ottoman Sultans who preceded him. After eliminating duplications and choosing between contradictory statements, he issued a single legal code, all the while being careful not to violate the basic laws of Islam. It was within this framework that Suleiman, supported by his Grand Mufti Ebussuud, sought to reform the legislation to adapt to a rapidly changing empire. When the Kanun laws attained their final form, the code of laws became known as the "kanun‐i Osmani" (), or the "Ottoman laws". Suleiman's legal code was to last more than three hundred years. The Sultan also played a role in protecting the Jewish subjects of his empire for centuries to come. In late 1553 or 1554, on the suggestion of his favorite doctor and dentist, the Spanish Jew Moses Hamon, the Sultan issued a "firman" () formally denouncing blood libels against the Jews. Furthermore, Suleiman enacted new criminal and police legislation, prescribing a set of fines for specific offenses, as well as reducing the instances requiring death or mutilation. In the area of taxation, taxes were levied on various goods and produce, including animals, mines, profits of trade, and import-export duties. Higher "medreses" provided education of university status, whose graduates became "imams" () or teachers. Educational centers were often one of many buildings surrounding the courtyards of mosques, others included libraries, baths, soup kitchens, residences and hospitals for the benefit of the public. Under Suleiman's patronage, the Ottoman Empire entered the golden age of its cultural development. Hundreds of imperial artistic societies (called the "Ehl-i Hiref", "Community of the Craftsmen") were administered at the Imperial seat, the Topkapı Palace. After an apprenticeship, artists and craftsmen could advance in rank within their field and were paid commensurate wages in quarterly annual installments. Payroll registers that survive testify to the breadth of Suleiman's patronage of the arts, the earliest of the documents dating from 1526 list 40 societies with over 600 members. The "Ehl-i Hiref" attracted the empire's most talented artisans to the Sultan's court, both from the Islamic world and from the recently conquered territories in Europe, resulting in a blend of Arabic, Turkish and European cultures. Artisans in service of the court included painters, book binders, furriers, jewellers and goldsmiths. Whereas previous rulers had been influenced by Persian culture (Suleiman's father, Selim I, wrote poetry in Persian), Suleiman's patronage of the arts saw the Ottoman Empire assert its own artistic legacy. Suleiman himself was an accomplished poet, writing in Persian and Turkish under the takhallus (nom de plume) "Muhibbi" (, "Lover"). Some of Suleiman's verses have become Turkish proverbs, such as the well-known "Everyone aims at the same meaning, but many are the versions of the story". When his young son Mehmed died in 1543, he composed a moving chronogram to commemorate the year: "Peerless among princes, my Sultan Mehmed". In Turkish the chronogram reads ("Şehzadeler güzidesi Sultan Muhammed'üm"), in which the Arabic Abjad numerals total 955, the equivalent in the Islamic calendar of 1543 AD. In addition to Suleiman's own work, many great talents enlivened the literary world during Suleiman's rule, including Fuzûlî and Bâkî. The literary historian Elias John Wilkinson Gibb observed that "at no time, even in Turkey, was greater encouragement given to poetry than during the reign of this Sultan". Suleiman's most famous verse is: The people think of wealth and power as the greatest fate, But in this world a spell of health is the best state. What men call sovereignty is a worldly strife and constant war; Worship of God is the highest throne, the happiest of all estates. Suleiman also became renowned for sponsoring a series of monumental architectural developments within his empire. The Sultan sought to turn Constantinople into the center of Islamic civilization by a series of projects, including bridges, mosques, palaces and various charitable and social establishments. The greatest of these were built by the Sultan's chief architect, Mimar Sinan, under whom Ottoman architecture reached its zenith. Sinan became responsible for over three hundred monuments throughout the empire, including his two masterpieces, the Süleymaniye and Selimiye mosques—the latter built in Adrianople (now Edirne) in the reign of Suleiman's son Selim II. Suleiman also restored the Dome of the Rock in Jerusalem and the Walls of Jerusalem (which are the current walls of the Old City of Jerusalem), renovated the Kaaba in Mecca, and constructed a complex in Damascus. Suleiman had two known consorts, though in total there were 17 women in his harem. Suleiman had several children with his consorts, including: Suleiman was infatuated with Hürrem Sultan, a harem girl from Ruthenia, then part of Poland. Western diplomats, taking notice of the palace gossip about her, called her "Russelazie" or "Roxelana", referring to her Ruthenian origins. The daughter of an Orthodox priest, she was captured by Tatars from Crimea, sold as a slave in Constantinople, and eventually rose through the ranks of the Harem to become Suleiman's favorite. Hürrem, a former concubine, became the legal wife of the Sultan, much to the astonishment of the observers in the palace and the city. He also allowed Hürrem Sultan to remain with him at court for the rest of her life, breaking another tradition—that when imperial heirs came of age, they would be sent along with the imperial concubine who bore them to govern remote provinces of the Empire, never to return unless their progeny succeeded to the throne. Under his pen name, Muhibbi, Sultan Suleiman composed this poem for Hürrem Sultan: Throne of my lonely niche, my wealth, my love, my moonlight. My most sincere friend, my confidant, my very existence, my Sultan, my one and only love. The most beautiful among the beautiful ... My springtime, my merry faced love, my daytime, my sweetheart, laughing leaf ... My plants, my sweet, my rose, the one only who does not distress me in this room ... My Istanbul, my karaman, the earth of my Anatolia My Badakhshan, my Baghdad and Khorasan My woman of the beautiful hair, my love of the slanted brow, my love of eyes full of misery ... I'll sing your praises always I, lover of the tormented heart, Muhibbi of the eyes full of tears, I am happy. Suleiman could speak Ottoman Turkish, Arabic, Chagatai, Persian and Serbian. Pargalı Ibrahim Pasha was a friend of Suleiman from before his accession. Ibrahim was originally a Christian from Parga (in Epirus), who was captured in a raid during the 1499–1503 Ottoman–Venetian War, and was given as a slave to Suleiman most likely in 1514. Ibrahim converted to Islam and Suleiman made him the royal falconer, then promoted him to first officer of the Royal Bedchamber. Ibrahim Pasha rose to Grand Vizier in 1523 and commander-in-chief of all the armies. Suleiman also conferred upon Ibrahim Pasha the honor of "beylerbey" of Rumelia (first-ranking military governor-general), granting Ibrahim authority over all Ottoman territories in Europe, as well as command of troops residing within them in times of war. During his thirteen years as Grand Vizier, his rapid rise to power and vast accumulation of wealth had made Ibrahim many enemies at the Sultan's court. Suleiman's suspicion of Ibrahim was worsened by a quarrel between the latter and the finance secretary ("defterdar") İskender Çelebi. The dispute ended in the disgrace of Çelebi on charges of intrigue, with Ibrahim convincing Suleiman to sentence the "defterdar" to death. Ibrahim also supported Şehzade Mustafa as the successor of Suleiman. This caused disputes between him and Hürrem Sultan, who wanted her sons to succeed to the throne. Ibrahim eventually fell from grace with the Sultan. Suleiman consulted his Qadi, who suggested that Ibrahim be put to death. The Sultan recruited assassins and ordered them to strangle Ibrahim in his sleep. Sultan Suleiman's two known consorts (Hürrem and Mahidevran) had borne him six sons, four of whom survived past the 1550s. They were Mustafa, Selim, Bayezid, and Cihangir. Of these, the eldest was not Hürrem's son, but rather Mahidevran's. Hürrem is usually held at least partly responsible for the intrigues in nominating a successor, though there is no evidence to support this. Although she was Suleiman's wife, she exercised no official public role. This did not, however, prevent Hürrem from wielding powerful political influence. Since the Empire lacked, until the reign of Ahmed I, any formal means of nominating a successor, successions usually involved the death of competing princes in order to avert civil unrest and rebellions. By 1552, when the campaign against Persia had begun with Rüstem appointed commander-in-chief of the expedition, intrigues against Mustafa began. Rüstem sent one of Suleiman's most trusted men to report that since Suleiman was not at the head of the army, the soldiers thought the time had come to put a younger prince on the throne; at the same time he spread rumors that Mustafa had proved receptive to the idea. Angered by what he came to believe were Mustafa's plans to claim the throne, the following summer upon return from his campaign in Persia, Suleiman summoned him to his tent in the Ereğli valley. When Mustafa entered his father's tent to meet with him, Suleiman's eunuchs attacked Mustafa, and after a long struggle the mutes killed him using a bow-string. Cihangir is said to have died of grief a few months after the news of his half-brother's murder. The two surviving brothers, Selim and Bayezid, were given command in different parts of the empire. Within a few years, however, civil war broke out between the brothers, each supported by his loyal forces. With the aid of his father's army, Selim defeated Bayezid in Konya in 1559, leading the latter to seek refuge with the Safavids along with his four sons. Following diplomatic exchanges, the Sultan demanded from the Safavid Shah that Bayezid be either extradited or executed. In return for large amounts of gold, the Shah allowed a Turkish executioner to strangle Bayezid and his four sons in 1561, clearing the path for Selim's succession to the throne five years later. On 6 September 1566, Suleiman, who had set out from Constantinople to command an expedition to Hungary, died before an Ottoman victory at the Battle of Szigetvár in Hungary and the Grand Vizier kept his death secret during the retreat for the enthronement of Selim II. Just the night before the sickly sultan died in his tent, two months before he would have turned 72. The sultan's body was taken back to Istanbul to be buried, while his heart, liver, and some other organs were buried in Turbék, outside Szigetvár. A mausoleum constructed above the burial site came to be regarded as a holy place and pilgrimage site. Within a decade a mosque and Sufi hospice were built near it, and the site was protected by a salaried garrison of several dozen men. The formation of Suleiman's legacy began even before his death. Throughout his reign literary works were commissioned praising Suleiman and constructing an image of him as an ideal ruler, most significantly by Celalzade Mustafa, chancellor of the empire from 1534–1557. Later Ottoman writers applied this idealized image of Suleiman to the Near Eastern literary genre of advice literature named "naṣīḥatnāme", urging sultans to conform to his model of rulership and to maintain the empire's institutions in their sixteenth-century form. Such writers were pushing back against the political and institutional transformation of the empire after the middle of the sixteenth century, and portrayed deviation from the norm as it had existed under Suleiman as evidence of the decline of the empire. Western historians, failing to recognize that these 'decline writers' were working within an established literary genre and often had deeply personal reasons for criticizing the empire, long took their claims at face value and consequently adopted the idea that the empire entered a period of decline after the death of Suleiman. Since the 1980s this view has been thoroughly reexamined, and modern scholars have come to overwhelmingly reject the idea of decline, labeling it an "untrue myth". Suleiman's conquests had brought under the control of the Empire major Muslim cities (such as Baghdad), many Balkan provinces (reaching present day Croatia and Hungary), and most of North Africa. His expansion into Europe had given the Ottoman Turks a powerful presence in the European balance of power. Indeed, such was the perceived threat of the Ottoman Empire under the reign of Suleiman that Austria's ambassador Busbecq warned of Europe's imminent conquest: "On [the Turks'] side are the resources of a mighty empire, strength unimpaired, habituation to victory, endurance of toil, unity, discipline, frugality and watchfulness ... Can we doubt what the result will be? ... When the Turks have settled with Persia, they will fly at our throats supported by the might of the whole East; how unprepared we are I dare not say." Suleiman's legacy was not, however, merely in the military field. The French traveler Jean de Thévenot bears witness a century later to the "strong agricultural base of the country, the well being of the peasantry, the abundance of staple foods and the pre-eminence of organization in Suleiman's government". Even thirty years after his death, "Sultan Solyman" was quoted by the English playwright William Shakespeare as a military prodigy in "The Merchant of Venice", where the Prince of Morocco boasts about his prowess by saying that he defeated Suleiman in three battles (Act 2, Scene 1). Through the distribution of court patronage, Suleiman also presided over a Golden Age in Ottoman arts, witnessing immense achievement in the realms of architecture, literature, art, theology and philosophy. Today the skyline of the Bosphorus and of many cities in modern Turkey and the former Ottoman provinces, are still adorned with the architectural works of Mimar Sinan. One of these, the Süleymaniye Mosque, is the final resting place of Suleiman: he is buried in a domed mausoleum attached to the mosque. Nevertheless, assessments of Suleiman's reign have frequently fallen into the trap of the Great Man theory of history. The administrative, cultural, and military achievements of the age were a product not of Suleiman alone, but also of the many talented figures who served him, such as grand viziers Ibrahim Pasha and Rüstem Pasha, the Grand Mufti Ebussuud Efendi, who played a major role in legal reform, and chancellor and chronicler Celalzade Mustafa, who played a major role in bureaucratic expansion and in constructing Suleiman's legacy. He has also been portrayed as a main playable character in the hit game Age of Empires III which is developed by Ensemble Studios. Printed sources Additional on-line sources
https://en.wikipedia.org/wiki?curid=26992
Scotland Scotland (, ) is a country that is part of the United Kingdom. Covering the northern third of the island of Great Britain, mainland Scotland has a 96 mile (154 km) border with England to the southeast and is otherwise surrounded by the Atlantic Ocean to the north and west, the North Sea to the northeast and the Irish Sea to the south. In addition, Scotland includes more than 790 islands; principally within the Northern Isles and the Hebrides archipelagos. The Kingdom of Scotland emerged as an independent sovereign state in the European Early Middle Ages and continued to exist until 1707. By inheritance in 1603, James VI of Scotland became king of England and Ireland, thus forming a personal union of the three kingdoms. Scotland subsequently entered into a political union with the Kingdom of England on 1 May 1707 to create the new Kingdom of Great Britain. The union also created a new Parliament of Great Britain, which succeeded both the Parliament of Scotland and the Parliament of England. In 1801, the Kingdom of Great Britain entered into a political union with the Kingdom of Ireland to create the United Kingdom of Great Britain and Ireland (in 1922, the Irish Free State seceded from the United Kingdom, leading to the latter being officially renamed the United Kingdom of Great Britain and Northern Ireland in 1927). Within Scotland, the monarchy of the United Kingdom has continued to use a variety of styles, titles and other royal symbols of statehood specific to the pre-union Kingdom of Scotland. The legal system within Scotland has also remained separate from those of England and Wales and Northern Ireland; Scotland constitutes a distinct jurisdiction in both public and private law. The continued existence of legal, educational, religious and other institutions distinct from those in the remainder of the UK have all contributed to the continuation of Scottish culture and national identity since the 1707 union with England. In 1999, a Scottish Parliament was re-established, in the form of a devolved unicameral legislature comprising 129 members, having authority over many areas of domestic policy. The head of the Scottish Government is the first minister of Scotland, who is supported by the deputy first minister of Scotland. Scotland is represented in the United Kingdom Parliament by 59 MPs. Scotland is also a member of the British–Irish Council, sending five members of the Scottish Parliament to the British–Irish Parliamentary Assembly. Scotland is divided into 32 administrative subdivisions or local authorities, known as council areas. Glasgow City is the largest council area in terms of population, with Highland being the largest in terms of area. Limited self-governing power, covering matters such as education, social services and roads and transportation, is devolved from the Scottish Government to each subdivision. "Scotland" comes from "Scoti", the Latin name for the Gaels. Philip Freeman has speculated on the likelihood of a group of raiders adopting a name from an Indo-European root, *"skot", citing the parallel in Greek "skotos" (σκότος), meaning "darkness, gloom". The Late Latin word "Scotia" ("land of the Gaels") was initially used to refer to Ireland. By the 11th century at the latest, "Scotia" was being used to refer to (Gaelic-speaking) Scotland north of the River Forth, alongside "Albania" or "Albany", both derived from the Gaelic "Alba". The use of the words "Scots" and "Scotland" to encompass all of what is now Scotland became common in the Late Middle Ages. Repeated glaciations, which covered the entire land mass of modern Scotland, destroyed any traces of human habitation that may have existed before the Mesolithic period. It is believed the first post-glacial groups of hunter-gatherers arrived in Scotland around 12,800 years ago, as the ice sheet retreated after the last glaciation. At the time, Scotland was covered in forests, had more bog-land, and the main form of transport was by water. These settlers began building the first known permanent houses on Scottish soil around 9,500 years ago, and the first villages around 6,000 years ago. The well-preserved village of Skara Brae on the mainland of Orkney dates from this period. Neolithic habitation, burial, and ritual sites are particularly common and well preserved in the Northern Isles and Western Isles, where a lack of trees led to most structures being built of local stone. Evidence of sophisticated pre-Christian belief systems is demonstrated by sites such as the Callanish Stones on Lewis and the Maes Howe on Orkney, which were built in the third millennium BCE. The first written reference to Scotland was in 320 BC by Greek sailor Pytheas, who called the northern tip of Britain "Orcas", the source of the name of the Orkney islands. During the first millennium BCE, the society changed dramatically to a chiefdom model, as consolidation of settlement led to the concentration of wealth and underground stores of surplus food. The first Roman incursion into Scotland occurred in 79 AD, when Agricola invaded Scotland; he defeated a Caledonian army at the Battle of Mons Graupius in 83 AD. After the Roman victory, Roman forts were briefly set along the Gask Ridge close to the Highland line, but by three years after the battle, the Roman armies had withdrawn to the Southern Uplands. The Romans erected Hadrian's Wall in northern England and the "Limes Britannicus" became the northern border of the Roman Empire. The Roman influence on the southern part of the country was considerable, and they introduced Christianity to Scotland. Beginning in the sixth century, the area that is now Scotland was divided into three areas: Pictland, a patchwork of small lordships in central Scotland; the Anglo-Saxon Kingdom of Northumbria, which had conquered southeastern Scotland; and Dál Riata, founded by settlers from Ireland, bringing Gaelic language and culture with them. These societies were based on the family unit and had sharp divisions in wealth, although the vast majority were poor and worked full-time in subsistence agriculture. The Picts kept slaves (mostly captured in war) through the ninth century. Gaelic influence over Pictland and Northumbria was facilitated by the large number of Gaelic-speaking clerics working as missionaries. Operating in the sixth century on the island of Iona, Saint Columba was one of the earliest and best-known missionaries. The Vikings began to raid Scotland in the eighth century. Although the raiders sought slaves and luxury items, their main motivation was to acquire land. The oldest Norse settlements were in northwest Scotland, but they eventually conquered many areas along the coast. Old Norse entirely displaced Gaelic in the Northern Isles. In the ninth century, the Norse threat allowed a Gael named Cináed mac Ailpín (Kenneth I) to seize power over Pictland, establishing a royal dynasty to which the modern monarchs trace their lineage, and marking the beginning of the end of Pictish culture. The kingdom of Cináed and his descendants, called Alba, was Gaelic in character but existed on the same area as Pictland. By the end of the tenth century, the Pictish language went extinct as its speakers shifted to Gaelic. From a base in eastern Scotland north of the River Forth and south of the River Spey, the kingdom expanded first southwards, into the former Northumbrian lands, and northwards into Moray. Around the turn of the millennium, there was a centralization in agricultural lands and the first towns began to be established. In the twelfth and thirteenth centuries, with much of Scotland under the control of a single ruler and united by the Gaelic language, a modern nation-state first emerged, as did Scottish national consciousness. The domination of Gaelic was diminished during the reign of David I (1124–53), during which many English-speaking colonists settled in Scotland. David I and his successors centralized royal power and united mainland Scotland, capturing regions such as Moray, Galloway, and Caithness, although he did not succeed at extending his power over the Hebrides, which had been ruled by various Scottish clans following the death of Somerled in 1164. The system of feudalism was consolidated, with both Anglo-Norman incomers and native Gaelic chieftains being granted land in exchange for serving the king. The Scottish kings rejected English demands to subjugate themselves; in fact, England invaded Scotland several times to prevent Scotland's expansion into northern England. The death of Alexander III in March 1286 broke the succession line of Scotland's kings. Edward I of England arbitrated between various claimants for the Scottish crown. In return for surrendering Scotland's nominal independence, John Balliol was pronounced king in 1292. In 1294, Balliol and other Scottish lords refused Edward's demands to serve in his army against the French. Scotland and France sealed a treaty on 23 October 1295, known as the Auld Alliance. War ensued, and John was deposed by Edward who took personal control of Scotland. Andrew Moray and William Wallace initially emerged as the principal leaders of the resistance to English rule in the Wars of Scottish Independence, until Robert the Bruce was crowned king of Scotland in 1306. Victory at the Battle of Bannockburn in 1314 proved the Scots had regained control of their kingdom. In 1320 the world's first documented declaration of independence, the Declaration of Arbroath, won the support of Pope John XXII, leading to the legal recognition of Scottish sovereignty by the English Crown. A civil war between the Bruce dynasty and their long-term Comyn-Balliol rivals lasted until the middle of the 14th century. Although the Bruce faction was successful, David II's lack of an heir allowed his half-nephew Robert II to come to the throne and establish the House of Stewart. The Stewarts ruled Scotland for the remainder of the Middle Ages. The country they ruled experienced greater prosperity from the end of the 14th century through the Scottish Renaissance to the Reformation, despite the effects of the Black Death in 1349 and increasing division between Highlands and Lowlands. Multiple truces reduced warfare on the southern border. The Treaty of Perpetual Peace was signed in 1502 by James IV of Scotland and Henry VII of England. James married Henry's daughter, Margaret Tudor. James invaded England in support of France under the terms of the Auld Alliance and became the last British monarch to die in battle, at Flodden in 1513. In 1560, the Treaty of Edinburgh brought an end to the Anglo-French conflict and recognized the Protestant Elizabeth I as Queen of England. The Parliament of Scotland met and immediately adopted the Scots Confession, which signaled the Scottish Reformation's sharp break from papal authority and Catholic teaching. The Catholic Mary, Queen of Scots was forced to abdicate in 1567. In 1603, James VI, King of Scots inherited the thrones of the Kingdom of England and the Kingdom of Ireland in the Union of the Crowns, and moved to London. The military was strengthened, allowing the imposition of royal authority on the western Highland clans. The 1609 Statutes of Iona compelled the cultural integration of Hebridean clan leaders. With the exception of a short period under the Protectorate, Scotland remained a separate state, but there was considerable conflict between the crown and the Covenanters over the form of church government. The Glorious Revolution of 1688–89 saw the overthrow of King James VII of Scotland and II of England by the English Parliament in favour of William III and Mary II. The Battle of Altimarlach in 1680 was the last significant clan battle fought between highland clans. In common with countries such as France, Norway, Sweden and Finland, Scotland experienced famines during the 1690s. Mortality, reduced childbirths and increased emigration reduced the population of parts of the country about 10–15%. In 1698, the Company of Scotland attempted a project to secure a trading colony on the Isthmus of Panama. Almost every Scottish landowner who had money to spare is said to have invested in the Darien scheme. Its failure bankrupted these landowners, but not the burghs. Nevertheless, the nobles' bankruptcy, along with the threat of an English invasion, played a leading role in convincing the Scots elite to back a union with England. On 22 July 1706, the Treaty of Union was agreed between representatives of the Scots Parliament and the Parliament of England. The following year, twin Acts of Union were passed by both parliaments to create the united Kingdom of Great Britain with effect from 1 May 1707 with popular opposition and anti-union riots in Edinburgh, Glasgow, and elsewhere. With trade tariffs with England abolished, trade blossomed, especially with Colonial America. The clippers belonging to the Glasgow Tobacco Lords were the fastest ships on the route to Virginia. Until the American War of Independence in 1776, Glasgow was the world's premier tobacco port, dominating world trade. The disparity between the wealth of the merchant classes of the Scottish Lowlands and the ancient clans of the Scottish Highlands grew, amplifying centuries of division. The deposed Jacobite Stuart claimants had remained popular in the Highlands and north-east, particularly amongst non-Presbyterians, including Roman Catholics and Episcopalian Protestants. However, two major Jacobite risings launched in 1715 and 1745 failed to remove the House of Hanover from the British throne. The threat of the Jacobite movement to the United Kingdom and its monarchs effectively ended at the Battle of Culloden, Great Britain's last pitched battle. The Scottish Enlightenment and the Industrial Revolution turned Scotland into an intellectual, commercial and industrial powerhouse – so much so Voltaire said "We look to Scotland for all our ideas of civilisation." With the demise of Jacobitism and the advent of the Union, thousands of Scots, mainly Lowlanders, took up numerous positions of power in politics, civil service, the army and navy, trade, economics, colonial enterprises and other areas across the nascent British Empire. Historian Neil Davidson notes "after 1746 there was an entirely new level of participation by Scots in political life, particularly outside Scotland." Davidson also states "far from being 'peripheral' to the British economy, Scotland – or more precisely, the Lowlands – lay at its core." In the Highlands, clan chiefs gradually started to think of themselves more as commercial landlords than leaders of their people. These social and economic changes included the first phase of the Highland Clearances and, ultimately, the demise of clanship. The Scottish Reform Act 1832 increased the number of Scottish MPs and widened the franchise to include more of the middle classes. From the mid-century, there were increasing calls for Home Rule for Scotland and the post of Secretary of State for Scotland was revived. Towards the end of the century Prime Ministers of Scottish descent included William Gladstone, and the Earl of Rosebery. In the late 19th century the growing importance of the working classes was marked by Keir Hardie's success in the Mid Lanarkshire by-election, 1888, leading to the foundation of the Scottish Labour Party, which was absorbed into the Independent Labour Party in 1895, with Hardie as its first leader. Glasgow became one of the largest cities in the world and known as "the Second City of the Empire" after London. After 1860 the Clydeside shipyards specialised in steamships made of iron (after 1870, made of steel), which rapidly replaced the wooden sailing vessels of both the merchant fleets and the battle fleets of the world. It became the world's pre-eminent shipbuilding centre. The industrial developments, while they brought work and wealth, were so rapid that housing, town-planning, and provision for public health did not keep pace with them, and for a time living conditions in some of the towns and cities were notoriously bad, with overcrowding, high infant mortality, and growing rates of tuberculosis. While the Scottish Enlightenment is traditionally considered to have concluded toward the end of the 18th century, disproportionately large Scottish contributions to British science and letters continued for another 50 years or more, thanks to such figures as the physicists James Clerk Maxwell and Lord Kelvin, and the engineers and inventors James Watt and William Murdoch, whose work was critical to the technological developments of the Industrial Revolution throughout Britain. In literature, the most successful figure of the mid-19th century was Walter Scott. His first prose work, "Waverley" in 1814, is often called the first historical novel. It launched a highly successful career that probably more than any other helped define and popularise Scottish cultural identity. In the late 19th century, a number of Scottish-born authors achieved international reputations, such as Robert Louis Stevenson, Arthur Conan Doyle, J. M. Barrie and George MacDonald. Scotland also played a major part in the development of art and architecture. The Glasgow School, which developed in the late 19th century, and flourished in the early 20th century, produced a distinctive blend of influences including the Celtic Revival the Arts and Crafts movement, and Japonism, which found favour throughout the modern art world of continental Europe and helped define the Art Nouveau style. Proponents included architect and artist Charles Rennie Mackintosh. This period saw a process of rehabilitation for Highland culture. In the 1820s, as part of the Romantic revival, tartan and the kilt were adopted by members of the social elite, not just in Scotland, but across Europe, prompted by the popularity of Macpherson's Ossian cycle and then Walter Scott's Waverley novels. However, the Highlands remained poor, the only part of mainland Britain to continue to experience recurrent famine, with a limited range of products exported out of the region, negligible industrial production, but a continued population growth that tested the subsistence agriculture. These problems, and the desire to improve agriculture and profits were the driving forces of the ongoing Highland Clearances, in which many of the population of the Highlands suffered eviction as lands were enclosed, principally so that they could be used for sheep farming. The first phase of the clearances followed patterns of agricultural change throughout Britain. The second phase was driven by overpopulation, the Highland Potato Famine and the collapse of industries that had relied on the wartime economy of the Napoleonic Wars. The population of Scotland grew steadily in the 19th century, from 1,608,000 in the census of 1801 to 2,889,000 in 1851 and 4,472,000 in 1901. Even with the development of industry, there were not enough good jobs. As a result, during the period 1841–1931, about 2 million Scots migrated to North America and Australia, and another 750,000 Scots relocated to England. After prolonged years of struggle in the Kirk, in 1834 the Evangelicals gained control of the General Assembly and passed the Veto Act, which allowed congregations to reject unwanted "intrusive" presentations to livings by patrons. The following "Ten Years' Conflict" of legal and political wrangling ended in defeat for the non-intrusionists in the civil courts. The result was a schism from the church by some of the non-intrusionists led by Dr Thomas Chalmers, known as the Great Disruption of 1843. Roughly a third of the clergy, mainly from the North and Highlands, formed the separate Free Church of Scotland. In the late 19th century growing divisions between fundamentalist Calvinists and theological liberals resulted in a further split in the Free Church as the rigid Calvinists broke away to form the Free Presbyterian Church in 1893. Catholic emancipation in 1829 and the influx of large numbers of Irish immigrants, particularly after the famine years of the late 1840s, mainly to the growing lowland centres like Glasgow, led to a transformation in the fortunes of Catholicism. In 1878, despite opposition, a Roman Catholic ecclesiastical hierarchy was restored to the country, and Catholicism became a significant denomination within Scotland. Industrialisation, urbanisation and the Disruption of 1843 all undermined the tradition of parish schools. From 1830 the state began to fund buildings with grants; then from 1846 it was funding schools by direct sponsorship; and in 1872 Scotland moved to a system like that in England of state-sponsored largely free schools, run by local school boards. The historic University of Glasgow became a leader in British higher education by providing the educational needs of youth from the urban and commercial classes, as opposed to the upper class. The University of St Andrews pioneered the admission of women to Scottish universities. From 1892 Scottish universities could admit and graduate women and the numbers of women at Scottish universities steadily increased until the early 20th century. Caused by the advent of refrigeration and imports of lamb, mutton and wool from overseas, the 1870s brought with them a collapse of sheep prices and an abrupt halt in the previous sheep farming boom. Land prices subsequently plummeted, too, and accelerated the process of the so-called "Balmoralisation" of Scotland, an era in the second half of the 19th century that saw an increase in tourism and the establishment of large estates dedicated to field sports like deer stalking and grouse shooting, especially in the Scottish Highlands. The process was named after Balmoral Estate, purchased by Queen Victoria in 1848, that fueled the romanticisation of upland Scotland and initiated an influx of the newly wealthy acquiring similar estates in the following decades. In the late 19th century just 118 people owned half of Scotland, with nearly 60 per cent of the whole country being part of shooting estates. While their relative importance has somewhat declined due to changing recreational interests throughout the 20th century, deer stalking and grouse shooting remain of prime importance on many private estates in Scotland. Scotland played a major role in the British effort in the First World War. It especially provided manpower, ships, machinery, fish and money. With a population of 4.8 million in 1911, Scotland sent over half a million men to the war, of whom over a quarter died in combat or from disease, and 150,000 were seriously wounded. Field Marshal Sir Douglas Haig was Britain's commander on the Western Front. The war saw the emergence of a radical movement called "Red Clydeside" led by militant trades unionists. Formerly a Liberal stronghold, the industrial districts switched to Labour by 1922, with a base among the Irish Catholic working-class districts. Women were especially active in building neighbourhood solidarity on housing issues. However, the "Reds" operated within the Labour Party and had little influence in Parliament and the mood changed to passive despair by the late 1920s. The shipbuilding industry expanded by a third and expected renewed prosperity, but instead, a serious depression hit the economy by 1922 and it did not fully recover until 1939. The interwar years were marked by economic stagnation in rural and urban areas, and high unemployment. Indeed, the war brought with it deep social, cultural, economic, and political dislocations. Thoughtful Scots pondered their declension, as the main social indicators such as poor health, bad housing, and long-term mass unemployment, pointed to terminal social and economic stagnation at best, or even a downward spiral. Service abroad on behalf of the Empire lost its allure to ambitious young people, who left Scotland permanently. The heavy dependence on obsolescent heavy industry and mining was a central problem, and no one offered workable solutions. The despair reflected what Finlay (1994) describes as a widespread sense of hopelessness that prepared local business and political leaders to accept a new orthodoxy of centralised government economic planning when it arrived during the Second World War. During the Second World War, Scotland was targeted by Nazi Germany largely due to its factories, shipyards, and coal mines. Cities such as Glasgow and Edinburgh were targeted by German bombers, as were smaller towns mostly located in the central belt of the country. Perhaps the most significant air-raid in Scotland was the Clydebank Blitz of March 1941, which intended to destroy naval shipbuilding in the area. 528 people were killed and 4,000 homes totally destroyed. Perhaps Scotland's most unusual wartime episode occurred in 1941 when Rudolf Hess flew to Renfrewshire, possibly intending to broker a peace deal through the Duke of Hamilton. Before his departure from Germany, Hess had given his adjutant, Karlheinz Pintsch, a letter addressed to Hitler that detailed his intentions to open peace negotiations with the British. Pintsch delivered the letter to Hitler at the Berghof around noon on 11 May. Albert Speer later said Hitler described Hess's departure as one of the worst personal blows of his life, as he considered it a personal betrayal. Hitler worried that his allies, Italy and Japan, would perceive Hess's act as an attempt by Hitler to secretly open peace negotiations with the British. As in World War I, Scapa Flow in Orkney served as an important Royal Navy base. Attacks on Scapa Flow and Rosyth gave RAF fighters their first successes downing bombers in the Firth of Forth and East Lothian. The shipyards and heavy engineering factories in Glasgow and Clydeside played a key part in the war effort, and suffered attacks from the Luftwaffe, enduring great destruction and loss of life. As transatlantic voyages involved negotiating north-west Britain, Scotland played a key part in the battle of the North Atlantic. Shetland's relative proximity to occupied Norway resulted in the Shetland bus by which fishing boats helped Norwegians flee the Nazis, and expeditions across the North Sea to assist resistance. Scottish industry came out of the depression slump by a dramatic expansion of its industrial activity, absorbing unemployed men and many women as well. The shipyards were the centre of more activity, but many smaller industries produced the machinery needed by the British bombers, tanks and warships. Agriculture prospered, as did all sectors except for coal mining, which was operating mines near exhaustion. Real wages, adjusted for inflation, rose 25% and unemployment temporarily vanished. Increased income, and the more equal distribution of food, obtained through a tight rationing system, dramatically improved the health and nutrition. After 1945, Scotland's economic situation worsened due to overseas competition, inefficient industry, and industrial disputes. Only in recent decades has the country enjoyed something of a cultural and economic renaissance. Economic factors contributing to this recovery included a resurgent financial services industry, electronics manufacturing, (see Silicon Glen), and the North Sea oil and gas industry. The introduction in 1989 by Margaret Thatcher's government of the Community Charge (widely known as the Poll Tax) one year before the rest of Great Britain, contributed to a growing movement for Scottish control over domestic affairs. Following a referendum on devolution proposals in 1997, the Scotland Act 1998 was passed by the UK Parliament, which established a devolved Scottish Parliament and Scottish Government with responsibility for most laws specific to Scotland. The Scottish Parliament was reconvened in Edinburgh on 4 July 1999. The first to hold the office of first minister of Scotland was Donald Dewar, who served until his sudden death in 2000. The Scottish Parliament Building at Holyrood itself did not open until October 2004, after lengthy construction delays and running over budget. The Scottish Parliament has a form of proportional representation (the additional member system), which normally results in no one party having an overall majority. The pro-independence Scottish National Party led by Alex Salmond achieved this in the 2011 election, winning 69 of the 129 seats available. The success of the SNP in achieving a majority in the Scottish Parliament paved the way for the September 2014 referendum on Scottish independence. The majority voted against the proposition, with 55% voting no to independence. More powers, particularly in relation to taxation, were devolved to the Scottish Parliament after the referendum, following cross-party talks in the Smith Commission. The mainland of Scotland comprises the northern third of the land mass of the island of Great Britain, which lies off the north-west coast of Continental Europe. The total area is , comparable to the size of the Czech Republic. Scotland's only land border is with England, and runs for between the basin of the River Tweed on the east coast and the Solway Firth in the west. The Atlantic Ocean borders the west coast and the North Sea is to the east. The island of Ireland lies only from the south-western peninsula of Kintyre; Norway is to the east and the Faroe Islands, to the north. The territorial extent of Scotland is generally that established by the 1237 Treaty of York between Scotland and the Kingdom of England and the 1266 Treaty of Perth between Scotland and Norway. Important exceptions include the Isle of Man, which having been lost to England in the 14th century is now a crown dependency outside of the United Kingdom; the island groups Orkney and Shetland, which were acquired from Norway in 1472; and Berwick-upon-Tweed, lost to England in 1482 The geographical centre of Scotland lies a few miles from the village of Newtonmore in Badenoch. Rising to above sea level, Scotland's highest point is the summit of Ben Nevis, in Lochaber, while Scotland's longest river, the River Tay, flows for a distance of . The whole of Scotland was covered by ice sheets during the Pleistocene ice ages and the landscape is much affected by glaciation. From a geological perspective, the country has three main sub-divisions. The Highlands and Islands lie to the north and west of the Highland Boundary Fault, which runs from Arran to Stonehaven. This part of Scotland largely comprises ancient rocks from the Cambrian and Precambrian, which were uplifted during the later Caledonian orogeny. It is interspersed with igneous intrusions of a more recent age, remnants of which formed mountain massifs such as the Cairngorms and Skye Cuillins. A significant exception to the above are the fossil-bearing beds of Old Red Sandstones found principally along the Moray Firth coast. The Highlands are generally mountainous and the highest elevations in the British Isles are found here. Scotland has over 790 islands divided into four main groups: Shetland, Orkney, and the Inner Hebrides and Outer Hebrides. There are numerous bodies of freshwater including Loch Lomond and Loch Ness. Some parts of the coastline consist of machair, a low-lying dune pasture land. The Central Lowlands is a rift valley mainly comprising Paleozoic formations. Many of these sediments have economic significance for it is here that the coal and iron bearing rocks that fuelled Scotland's industrial revolution are found. This area has also experienced intense volcanism, Arthur's Seat in Edinburgh being the remnant of a once much larger volcano. This area is relatively low-lying, although even here hills such as the Ochils and Campsie Fells are rarely far from view. The Southern Uplands are a range of hills almost long, interspersed with broad valleys. They lie south of a second fault line (the Southern Uplands fault) that runs from Girvan to Dunbar. The geological foundations largely comprise Silurian deposits laid down some 400–500 million years ago. The high point of the Southern Uplands is Merrick with an elevation of . The Southern Uplands is home to Scotland's highest village, Wanlockhead ( above sea level). The climate of most of Scotland is temperate and oceanic, and tends to be very changeable., As it is warmed by the Gulf Stream from the Atlantic, it has much milder winters (but cooler, wetter summers) than areas on similar latitudes, such as Labrador, southern Scandinavia, the Moscow region in Russia, and the Kamchatka Peninsula on the opposite side of Eurasia. However, temperatures are generally lower than in the rest of the UK, with the coldest ever UK temperature of recorded at Braemar in the Grampian Mountains, on 11 February 1895. Winter maxima average in the Lowlands, with summer maxima averaging . The highest temperature recorded was at Greycrook, Scottish Borders on 9 August 2003. The west of Scotland is usually warmer than the east, owing to the influence of Atlantic ocean currents and the colder surface temperatures of the North Sea. Tiree, in the Inner Hebrides, is one of the sunniest places in the country: it had more than 300 hours of sunshine in May 1975. Rainfall varies widely across Scotland. The western highlands of Scotland are the wettest, with annual rainfall in a few places exceeding . In comparison, much of lowland Scotland receives less than annually. Heavy snowfall is not common in the lowlands, but becomes more common with altitude. Braemar has an average of 59 snow days per year, while many coastal areas average fewer than 10 days of lying snow per year. Scotland's wildlife is typical of the north-west of Europe, although several of the larger mammals such as the lynx, brown bear, wolf, elk and walrus were hunted to extinction in historic times. There are important populations of seals and internationally significant nesting grounds for a variety of seabirds such as gannets. The golden eagle is something of a national icon. On the high mountain tops, species including ptarmigan, mountain hare and stoat can be seen in their white colour phase during winter months. Remnants of the native Scots pine forest exist and within these areas the Scottish crossbill, the UK's only endemic bird species and vertebrate, can be found alongside capercaillie, Scottish wildcat, red squirrel and pine marten. Various animals have been re-introduced, including the white-tailed sea eagle in 1975, the red kite in the 1980s, and there have been experimental projects involving the beaver and wild boar. Today, much of the remaining native Caledonian Forest lies within the Cairngorms National Park and remnants of the forest remain at 84 locations across Scotland. On the west coast, remnants of ancient Celtic Rainforest still remain, particularly on the Taynish peninsula in Argyll, these forests are particularly rare due to high rates of deforestation throughout Scottish history. The flora of the country is varied incorporating both deciduous and coniferous woodland as well as moorland and tundra species. However, large scale commercial tree planting and the management of upland moorland habitat for the grazing of sheep and field sport activities like deer stalking and driven grouse shooting impacts upon the distribution of indigenous plants and animals. The UK's tallest tree is a grand fir planted beside Loch Fyne, Argyll in the 1870s, and the Fortingall Yew may be 5,000 years old and is probably the oldest living thing in Europe. Although the number of native vascular plants is low by world standards, Scotland's substantial bryophyte flora is of global importance. The population of Scotland at the 2001 Census was 5,062,011. This rose to 5,295,400, the highest ever, at the 2011 Census. The most recent ONS estimate, for mid-2017, was 5,424,800. In the 2011 Census, 62% of Scotland's population stated their national identity as 'Scottish only', 18% as 'Scottish and British', 8% as 'British only', and 4% chose 'other identity only'. Although Edinburgh is the capital of Scotland, the largest city is Glasgow, which has just over 584,000 inhabitants. The Greater Glasgow conurbation, with a population of almost 1.2 million, is home to nearly a quarter of Scotland's population. The Central Belt is where most of the main towns and cities are located, including Glasgow, Edinburgh, Dundee, and Perth. Scotland's only major city outside the Central Belt is Aberdeen. The Scottish Lowlands host 80% of the total population, where the Central Belt accounts for 3.5 million people. In general, only the more accessible and larger islands remain inhabited. Currently, fewer than 90 remain inhabited. The Southern Uplands are essentially rural in nature and dominated by agriculture and forestry. Because of housing problems in Glasgow and Edinburgh, five new towns were designated between 1947 and 1966. They are East Kilbride, Glenrothes, Cumbernauld, Livingston, and Irvine. Immigration since World War II has given Glasgow, Edinburgh, and Dundee small South Asian communities. In 2011, there were an estimated 49,000 ethnically Pakistani people living in Scotland, making them the largest non-White ethnic group. Since the Enlargement of the European Union more people from Central and Eastern Europe have moved to Scotland, and the 2011 census indicated that 61,000 Poles live there. Scotland has three officially recognised languages: English, Scots, and Scottish Gaelic. Scottish Standard English, a variety of English as spoken in Scotland, is at one end of a bipolar linguistic continuum, with broad Scots at the other. Scottish Standard English may have been influenced to varying degrees by Scots. The 2011 census indicated that 63% of the population had "no skills in Scots". Others speak Highland English. Gaelic is mostly spoken in the Western Isles, where a large proportion of people still speak it; however, nationally its use is confined to just 1% of the population. The number of Gaelic speakers in Scotland dropped from 250,000 in 1881 to 60,000 in 2008. There are many more people with Scottish ancestry living abroad than the total population of Scotland. In the 2000 Census, 9.2 million Americans self-reported some degree of Scottish descent. Ulster's Protestant population is mainly of lowland Scottish descent, and it is estimated that there are more than 27 million descendants of the Scots-Irish migration now living in the US. In Canada, the Scottish-Canadian community accounts for 4.7 million people. About 20% of the original European settler population of New Zealand came from Scotland. In August 2012, the Scottish population reached an all-time high of 5.25 million people. The reasons given were that, in Scotland, births were outnumbering the number of deaths, and immigrants were moving to Scotland from overseas. In 2011, 43,700 people moved from Wales, Northern Ireland or England to live in Scotland. The total fertility rate (TFR) in Scotland is below the replacement rate of 2.1 (the TFR was 1.73 in 2011). The majority of births are to unmarried women (51.3% of births were outside of marriage in 2012). Life expectancy for those born in Scotland between 2012 and 2014 is 77.1 years for males and 81.1 years for females. This is the lowest of any of the four countries of the UK. Just over half (54%) of the Scottish population reported being a Christian while nearly 37% reported not having a religion in a 2011 census. Since the Scottish Reformation of 1560, the national church (the Church of Scotland, also known as The Kirk) has been Protestant in classification and Reformed in theology. Since 1689 it has had a Presbyterian system of church government and enjoys independence from the state. Its membership is 398,389, about 7.5% of the total population, though according to the 2014 Scottish Annual Household Survey, 27.8%, or 1.5 million adherents, identified the Church of Scotland as the church of their religion. The Church operates a territorial parish structure, with every community in Scotland having a local congregation. Scotland also has a significant Roman Catholic population, 19% professing that faith, particularly in Greater Glasgow and the north-west. After the Reformation, Roman Catholicism in Scotland continued in the Highlands and some western islands like Uist and Barra, and it was strengthened during the 19th century by immigration from Ireland. Other Christian denominations in Scotland include the Free Church of Scotland, and various other Presbyterian offshoots. Scotland's third largest church is the Scottish Episcopal Church. Islam is the largest non-Christian religion (estimated at around 75,000, which is about 1.4% of the population), and there are also significant Jewish, Hindu and Sikh communities, especially in Glasgow. The Samyé Ling monastery near Eskdalemuir, which celebrated its 40th anniversary in 2007, is the first Buddhist monastery in western Europe. The head of state of the United Kingdom is the monarch, currently Queen Elizabeth II (since 1952). The monarchy of the United Kingdom continues to use a variety of styles, titles and other royal symbols of statehood specific to pre-union Scotland, including: the Royal Standard of Scotland, the Royal coat of arms used in Scotland together with its associated Royal Standard, royal titles including that of Duke of Rothesay, certain Great Officers of State, the chivalric Order of the Thistle and, since 1999, reinstating a ceremonial role for the Crown of Scotland after a 292-year hiatus. Elizabeth II's regnal numbering caused controversy in 1953 because there had never been an Elizabeth I in Scotland. A legal action was brought in Scotland to contest the right of the Queen to entitle herself "Elizabeth II" within Scotland, but the Crown won the case. Scotland has limited self-government within the United Kingdom, as well as representation in the UK Parliament. Executive and legislative powers respectively have been devolved to the Scottish Government and the Scottish Parliament at Holyrood in Edinburgh since 1999. The UK Parliament retains control over reserved matters specified in the Scotland Act 1998, including UK taxes, social security, defence, international relations and broadcasting. The Scottish Parliament has legislative authority for all other areas relating to Scotland. It initially had only a limited power to vary income tax, but powers over taxation and social security were significantly expanded by the Scotland Acts of 2012 and 2016. The Scottish Parliament can give legislative consent over devolved matters back to the UK Parliament by passing a Legislative Consent Motion if United Kingdom-wide legislation is considered more appropriate for a certain issue. The programmes of legislation enacted by the Scottish Parliament have seen a divergence in the provision of public services compared to the rest of the UK. For instance, university education and care services for the elderly are free at point of use in Scotland, while fees are paid in the rest of the UK. Scotland was the first country in the UK to ban smoking in enclosed public places. The Scottish Parliament is a unicameral legislature with 129 members (MSPs): 73 of them represent individual constituencies and are elected on a first-past-the-post system; the other 56 are elected in eight different electoral regions by the additional member system. MSPs serve for a four-year period (exceptionally five years from 2011–16). The Parliament nominates one of its Members, who is then appointed by the monarch to serve as first minister. Other ministers are appointed by the first minister and serve at his/her discretion. Together they make up the Scottish Government, the executive arm of the devolved government. The Scottish Government is headed by the first minister, who is accountable to the Scottish Parliament and is the minister of charge of the Scottish Government. The first minister is also the political leader of Scotland. The Scottish Government also comprises the deputy first minister, currently John Swinney MSP, who deputises for the first minister during a period of absence of overseas visits. Alongside the deputy first minister's requirements as Deputy, the minister also has a cabinet ministerial responsibility. Swinney is also currently Cabinet Secretary for Education and Skills. The Scottish Government's cabinet comprises nine cabinet secretaries, who form the Cabinet of Scotland. There are also twelve other ministers, who work alongside the cabinet secretaries in their appointed areas. In the 2016 election, the Scottish National Party (SNP) won 63 of the 129 seats available. Nicola Sturgeon, the leader of the SNP, has been the first minister since November 2014. The Conservative Party became the largest opposition party in the 2016 elections, with the Labour Party, Liberal Democrats and the Green Party also represented in the Parliament. The next Scottish Parliament election is due to be held on 6 May 2021. Scotland is represented in the British House of Commons by 59 MPs elected from territory-based Scottish constituencies. In the 2019 general election, the SNP won 48 of the 59 seats. This represented a significant increase from the 2017 general election, when the SNP won 35 seats. Conservative, Labour and Liberal Democrat parties also represent Scottish constituencies in the House of Commons. The next United Kingdom general election is scheduled for 2 May 2024. The Scotland Office represents the UK government in Scotland on reserved matters and represents Scottish interests within the UK government. The Scotland Office is led by the Secretary of State for Scotland, who sits in the Cabinet of the United Kingdom. Conservative MP Alister Jack has held the position since July 2019. The relationships between the central UK Government and devolved governments of Scotland, Wales and Northern Ireland are based on the extra-statutory principles and agreements with the main elements are set out in a "Memorandum of Understanding" between the UK government and the devolved governments of Scotland, Wales and Northern Ireland. The MOU lays emphasis on the principles of good communication, consultation and co-operation. Since devolution in 1999, Scotland has devolved stronger working relations across the two other devolved governments, the Welsh Government and Northern Ireland Executive. Whilst there are no formal concordats between the Scottish Government, Welsh Government and Northern Ireland Executive, ministers from each devolved government meet at various points throughout the year at various events such as the British-Irish Council and also meet to discuss matters and issues that are devolved to each government. Scotland, along with the Welsh Government, British Government as well as the Northern Ireland executive, participate in the Joint Ministerial Committee (JMC) which allows each government to discuss policy issues together and work together across each government to find solutions. The Scottish Government considers the successful re-establishment of the Plenary, and establishment of the Domestic fora to be important facets of the relationship with the UK Government and the other devolved administrations. In the aftermath of the United Kingdom's decision to withdraw from the European Union in 2016, the Scottish Government has called for there to be a joint approach from each of the devolved governments. In early 2017, the devolved governments met to discuss Brexit and agree on Brexit strategies from each devolved government which lead for Theresa May to issue a statement that claims that the devolved governments will not have a central role or decision making process in the Brexit process, but that the UK Government plans to "fully engage" Scotland in talks alongside the governments of Wales and Northern Ireland. Whilst foreign policy remains a reserved matter, the Scottish Government still has the power and ability to strengthen and develop Scotland, the economy and Scottish interests on the world stage and encourage foreign businesses, international devolved, regional and central governments to invest in Scotland. Whilst the first minister usually undertakes a number of foreign and international visits to promote Scotland, international relations, European and Commonwealth relations are also included within the portfolios of both the Cabinet Secretary for Culture, Tourism and External Affairs (responsible for international development) and the Minister for International Development and Europe (responsible for European Union relations and international relations). During the G8 Summit in 2005, First Minister Jack McConnell welcomed each head of government of the G8 nations to the countries Glasgow Prestwick Airport on behalf of then UK Prime Minister Tony Blair. At the same time, McConnell and the then Scottish Executive pioneered the way forward to launch what would become the Scotland Malawi Partnership which co-ordinates Scottish activities to strengthen existing links with Malawi. During McConnell's time as first minister, several relations with Scotland, including Scottish and Russian relations strengthened following a visit by President of Russia Vladimir Putin to Edinburgh. McConnell, speaking at the end, highlighted that the visit by Putin was a "post-devolution" step towards "Scotland regaining its international identity". Under the Salmond administration, Scotland's trade and investment deals with countries such as China and Canada, where Salmond established the Canada Plan 2010–2015 which aimed to strengthen "the important historical, cultural and economic links" between both Canada and Scotland. To promote Scotland's interests and Scottish businesses in North America, there is a Scottish Affairs Office located in Washington, D.C. with the aim to promoting Scotland in both the United States and Canada. During a 2017 visit to the United States, First Minister Nicola Sturgeon met with Jerry Brown, Governor of California, where both signed an agreement committing both the Government of California and the Scottish Government to work together to tackle climate change, as well as Sturgeon signing a £6.3 million deal for Scottish investment from American businesses and firms promoting trade, tourism and innovation. During an official visit to the Republic of Ireland in 2016, Sturgeon claimed that is it "important for Ireland and Scotland and the whole of the British Isles that Ireland has a strong ally in Scotland". During the same engagement, Sturgeon became the first head of government to address the Seanad Éireann, the Upper House of the Irish Parliament. A policy of devolution had been advocated by the three main UK parties with varying enthusiasm during recent history. A previous Labour leader. John Smith, described the revival of a Scottish parliament as the "settled will of the Scottish people". The devolved Scottish Parliament was created after a referendum in 1997 found majority support for both creating the Parliament and granting it limited powers to vary income tax. The Scottish National Party (SNP), which supports Scottish independence, was first elected to form the Scottish Government in 2007. The new government established a "National Conversation" on constitutional issues, proposing a number of options such as increasing the powers of the Scottish Parliament, federalism, or a referendum on Scottish independence from the United Kingdom. In rejecting the last option, the three main opposition parties in the Scottish Parliament created a commission to investigate the distribution of powers between devolved Scottish and UK-wide bodies. The Scotland Act 2012, based on proposals by the commission, was subsequently enacted devolving additional powers to the Scottish Parliament. In August 2009 the SNP proposed a bill to hold a referendum on independence in November 2010. Opposition from all other major parties led to an expected defeat. After the 2011 elections gave the SNP an overall majority in the Scottish Parliament, a referendum on independence for Scotland was held on 18 September 2014. The referendum resulted in a rejection of independence, by 55.3% to 44.7%. During the campaign, the three main parties in the UK Parliament pledged to extend the powers of the Scottish Parliament. An all-party commission chaired by Lord Smith of Kelvin was formed, which led to a further devolution of powers through the Scotland Act 2016. Following a referendum on the UK's membership of the European Union on 23 June 2016, where a UK-wide majority voted to withdraw from the EU whilst a majority within Scotland voted to remain, Scotland's first minister, Nicola Sturgeon, announced that as a result a new independence referendum was "highly likely". Historical subdivisions of Scotland included the mormaerdom, stewartry, earldom, burgh, parish, county and regions and districts. Some of these names are still sometimes used as geographical descriptors. Modern Scotland is subdivided in various ways depending on the purpose. In local government, there have been 32 single-tier council areas since 1996, whose councils are responsible for the provision of all local government services. Decisions are made by councillors who are elected at local elections every five years. The head of each council is usually the Lord Provost alongside the Leader of the Council, with a Chief Executive being appointed as director of the council area. Community Councils are informal organisations that represent specific sub-divisions within each council area. In the Scottish Parliament, there are 73 constituencies and eight regions. For the Parliament of the United Kingdom, there are 59 constituencies. Until 2013, the Scottish fire brigades and police forces were based on a system of regions introduced in 1975. For healthcare and postal districts, and a number of other governmental and non-governmental organisations such as the churches, there are other long-standing methods of subdividing Scotland for the purposes of administration. City status in the United Kingdom is conferred by letters patent. There are seven cities in Scotland: Aberdeen, Dundee, Edinburgh, Glasgow, Inverness, Stirling and Perth. Scots law has a basis derived from Roman law, combining features of both uncodified civil law, dating back to the "Corpus Juris Civilis", and common law with medieval sources. The terms of the Treaty of Union with England in 1707 guaranteed the continued existence of a separate legal system in Scotland from that of England and Wales. Prior to 1611, there were several regional law systems in Scotland, most notably Udal law in Orkney and Shetland, based on old Norse law. Various other systems derived from common Celtic or Brehon laws survived in the Highlands until the 1800s. Scots law provides for three types of courts responsible for the administration of justice: civil, criminal and heraldic. The supreme civil court is the Court of Session, although civil appeals can be taken to the Supreme Court of the United Kingdom (or before 1 October 2009, the House of Lords). The High Court of Justiciary is the supreme criminal court in Scotland. The Court of Session is housed at Parliament House, in Edinburgh, which was the home of the pre-Union Parliament of Scotland with the High Court of Justiciary and the Supreme Court of Appeal currently located at the Lawnmarket. The sheriff court is the main criminal and civil court, hearing most cases. There are 49 sheriff courts throughout the country. District courts were introduced in 1975 for minor offences and small claims. These were gradually replaced by Justice of the Peace Courts from 2008 to 2010. The Court of the Lord Lyon regulates heraldry. For three centuries the Scots legal system was unique for being the only national legal system without a parliament. This ended with the advent of the Scottish Parliament in 1999, which legislates for Scotland. Many features within the system have been preserved. Within criminal law, the Scots legal system is unique in having three possible verdicts: "guilty", "not guilty" and ""not proven"". Both "not guilty" and "not proven" result in an acquittal, typically with no possibility of retrial in accordance with the rule of double jeopardy. There is, however, the possibility of a retrial where new evidence emerges at a later date that might have proven conclusive in the earlier trial at first instance, where the person acquitted subsequently admits the offence or where it can be proved that the acquittal was tainted by an attempt to pervert the course of justice – see the provisions of the Double Jeopardy (Scotland) Act 2011. Many laws differ between Scotland and the other parts of the United Kingdom, and many terms differ for certain legal concepts. Manslaughter, in England and Wales, is broadly similar to culpable homicide in Scotland, and arson is called wilful fire raising. Indeed, some acts considered crimes in England and Wales, such as forgery, are not so in Scotland. Procedure also differs. Scots juries, sitting in criminal cases, consist of fifteen jurors, which is three more than is typical in many countries. The Scottish Prison Service (SPS) manages the prisons in Scotland, which collectively house over 8,500 prisoners. The Cabinet Secretary for Justice is responsible for the Scottish Prison Service within the Scottish Government. Health care in Scotland is mainly provided by NHS Scotland, Scotland's public health care system. This was founded by the National Health Service (Scotland) Act 1947 (later repealed by the National Health Service (Scotland) Act 1978) that took effect on 5 July 1948 to coincide with the launch of the NHS in England and Wales. However, even prior to 1948, half of Scotland's landmass was already covered by state-funded health care, provided by the Highlands and Islands Medical Service. Healthcare policy and funding is the responsibility of the Scottish Government's Health Directorates. The current Cabinet Secretary for Health and Sport is Jeane Freeman and the Director-General (DG) Health and chief executive, NHS Scotland is Paul Gray. In 2008, the NHS in Scotland had around 158,000 staff including more than 47,500 nurses, midwives and health visitors and over 3,800 consultants. There are also more than 12,000 doctors, family practitioners and allied health professionals, including dentists, opticians and community pharmacists, who operate as independent contractors providing a range of services within the NHS in return for fees and allowances. These fees and allowances were removed in May 2010, and prescriptions are entirely free, although dentists and opticians may charge if the patient's household earns over a certain amount, about £30,000 per annum. Scotland has a Western-style open mixed economy closely linked with the rest of the UK and the wider world. Traditionally, the Scottish economy was dominated by heavy industry underpinned by shipbuilding in Glasgow, coal mining and steel industries. Petroleum related industries associated with the extraction of North Sea oil have also been important employers from the 1970s, especially in the north-east of Scotland. De-industrialisation during the 1970s and 1980s saw a shift from a manufacturing focus towards a more service-oriented economy. Scotland's gross domestic product (GDP), including oil and gas produced in Scottish waters, was estimated at £150 billion for the calendar year 2012. In 2014, Scotland's per capita GDP was one of the highest in the EU. As of April 2019 the Scottish unemployment rate was 3.3%, below the UK rate of 3.8%, and the Scottish employment rate was 75.9%. Edinburgh is the financial services centre of Scotland, with many large finance firms based there, including: Lloyds Banking Group (owners of HBOS); the Government-owned Royal Bank of Scotland and Standard Life. Edinburgh was ranked 15th in the list of world financial centres in 2007, but fell to 37th in 2012, following damage to its reputation, and in 2016 was ranked 56th out of 86. Its status had returned to 17th however by 2020. In 2014, total Scottish exports (excluding intra-UK trade) were estimated to be £27.5 billion. Scotland's primary exports include whisky, electronics and financial services. The United States, Netherlands, Germany, France, and Norway constitute the country's major export markets. Whisky is one of Scotland's more known goods of economic activity. Exports increased by 87% in the decade to 2012 and were valued at £4.3 billion in 2013, which was 85% of Scotland's food and drink exports. It supports around 10,000 jobs directly and 25,000 indirectly. It may contribute £400–682 million to Scotland, rather than several billion pounds, as more than 80% of whisky produced is owned by non-Scottish companies. A briefing published in 2002 by the Scottish Parliament Information Centre (SPICe) for the Scottish Parliament's Enterprise and Life Long Learning Committee stated that tourism accounted for up to 5% of GDP and 7.5% of employment. Although the Bank of England is the central bank for the UK, three Scottish clearing banks issue Sterling banknotes: the Bank of Scotland, the Royal Bank of Scotland and the Clydesdale Bank. The value of the Scottish banknotes in circulation in 2013 was £3.8 billion, underwritten by the Bank of England using funds deposited by each clearing bank, under the Banking Act 2009, in order to cover the total value of such notes in circulation. Of the money spent on UK defence, about £3.3 billion can be attributed to Scotland as of 2013. Although Scotland has a long military tradition predating the Treaty of Union with England, its armed forces now form part of the British Armed Forces, with the exception of the Atholl Highlanders, Europe's only legal private army. In 2006, the infantry regiments of the Scottish Division were amalgamated to form the Royal Regiment of Scotland. Other distinctively Scottish regiments in the British Army include the Scots Guards, the Royal Scots Dragoon Guards and the 154 (Scottish) Regiment RLC, an Army Reserve Regiment of the Royal Logistic Corps. Because of their topography and perceived remoteness, parts of Scotland have housed many sensitive defence establishments. Between 1960 and 1991, the Holy Loch was a base for the US fleet of Polaris ballistic missile submarines. Today, Her Majesty's Naval Base Clyde, north-west of Glasgow, is the base for the four Trident-armed ballistic missile submarines that comprise the UK's nuclear deterrent. Scapa Flow was the major Fleet base for the Royal Navy until 1956. A single front-line Royal Air Force base is located in Scotland. RAF Lossiemouth, located in Moray, is the most northerly air defence fighter base in the United Kingdom and is home to three fast-jet squadrons equipped with the Eurofighter Typhoon. The Scottish education system has always been distinct from the rest of the United Kingdom, with a characteristic emphasis on a broad education. In the 15th century, the Humanist emphasis on education cumulated with the passing of the Education Act 1496, which decreed that all sons of barons and freeholders of substance should attend grammar schools to learn "perfyct Latyne", resulting in an increase in literacy among a male and wealthy elite. In the Reformation, the 1560 "First Book of Discipline" set out a plan for a school in every parish, but this proved financially impossible. In 1616 an act in Privy council commanded every parish to establish a school. By the late seventeenth century there was a largely complete network of parish schools in the lowlands, but in the Highlands basic education was still lacking in many areas. Education remained a matter for the church rather than the state until the Education (Scotland) Act 1872. The "Curriculum for Excellence", Scotland's national school curriculum, presently provides the curricular framework for children and young people from age 3 to 18. All 3- and 4-year-old children in Scotland are entitled to a free nursery place. Formal primary education begins at approximately 5 years old and lasts for 7 years (P1–P7); children in Scotland study Standard Grades, or Intermediate qualifications between the ages of 14 and 16. These are being phased out and replaced by the National Qualifications of the Curriculum for Excellence. The school leaving age is 16, after which students may choose to remain at school and study for Access, Intermediate or Higher Grade and Advanced Higher qualifications. A small number of students at certain private, independent schools may follow the English system and study towards GCSEs and A and AS-Levels instead. There are fifteen Scottish universities, some of which are amongst the oldest in the world. These include the University of St Andrews, the University of Glasgow, the University of Aberdeen and the University of Edinburgh—many of which are ranked amongst the best in the UK. Scotland had more universities per capita in QS' World University Rankings' top 100 in 2012 than any other nation. The country produces 1% of the world's published research with less than 0.1% of the world's population, and higher education institutions account for 9% of Scotland's service sector exports. Scotland's University Courts are the only bodies in Scotland authorised to award degrees. Tuition is handled by the Student Awards Agency Scotland (SAAS), which does not charge fees to what it defines as "Young Students". Young Students are defined as those under 25, without children, marriage, civil partnership or cohabiting partner, who have not been outside of full-time education for more than three years. Fees exist for those outside the young student definition, typically from £1,200 to £1,800 for undergraduate courses, dependent on year of application and type of qualification. Postgraduate fees can be up to £3,400. The system has been in place since 2007 when graduate endowments were abolished. Labour's education spokesperson Rhona Brankin criticised the Scottish system for failing to address student poverty. Scotland's universities are complemented in the provision of Further and Higher Education by 43 colleges. Colleges offer National Certificates, Higher National Certificates, and Higher National Diplomas. These Group Awards, alongside Scottish Vocational Qualifications, aim to ensure Scotland's population has the appropriate skills and knowledge to meet workplace needs. In 2014, research reported by the Office for National Statistics found that Scotland was the most highly educated country in Europe and among the most well-educated in the world in terms of tertiary education attainment, with roughly 40% of people in Scotland aged 16–64 educated to NVQ level 4 and above. Based on the original data for EU statistical regions, all four Scottish regions ranked significantly above the European average for completion of tertiary-level education by 25- to 64-year-olds. Kilmarnock Academy in East Ayrshire is one of only two schools in the UK, and the only school in Scotland, to have educated two Nobel Prize Laureates – Alexander Fleming, discoverer of Penicillin, and John Boyd Orr, 1st Baron Boyd-Orr, for his scientific research into nutrition and his work as the first Director-General of the United Nations Food and Agriculture Organization (FAO). Scottish music is a significant aspect of the nation's culture, with both traditional and modern influences. A famous traditional Scottish instrument is the Great Highland bagpipe, a wind instrument consisting of three drones and a melody pipe (called the chanter), which are fed continuously by a reservoir of air in a bag. Bagpipe bands, featuring bagpipes and various types of drums, and showcasing Scottish music styles while creating new ones, have spread throughout the world. The clàrsach (harp), fiddle and accordion are also traditional Scottish instruments, the latter two heavily featured in Scottish country dance bands. There are many successful Scottish bands and individual artists in varying styles including Annie Lennox, Amy Macdonald, Runrig, Belle and Sebastian, Boards of Canada, Camera Obscura, Cocteau Twins, Deacon Blue, Franz Ferdinand, Susan Boyle, Emeli Sandé, Texas, The View, The Fratellis, Twin Atlantic and Biffy Clyro. Other Scottish musicians include Shirley Manson, Paolo Nutini, Andy Stewart and Calvin Harris. Scotland has a literary heritage dating back to the early Middle Ages. The earliest extant literature composed in what is now Scotland was in Brythonic speech in the 6th century, but is preserved as part of Welsh literature. Later medieval literature included works in Latin, Gaelic, Old English and French. The first surviving major text in Early Scots is the 14th-century poet John Barbour's epic "Brus", focusing on the life of Robert I, and was soon followed by a series of vernacular romances and prose works. In the 16th century, the crown's patronage helped the development of Scots drama and poetry, but the accession of James VI to the English throne removed a major centre of literary patronage and Scots was sidelined as a literary language. Interest in Scots literature was revived in the 18th century by figures including James Macpherson, whose Ossian Cycle made him the first Scottish poet to gain an international reputation and was a major influence on the European Enlightenment. It was also a major influence on Robert Burns, whom many consider the national poet, and Walter Scott, whose Waverley Novels did much to define Scottish identity in the 19th century. Towards the end of the Victorian era a number of Scottish-born authors achieved international reputations as writers in English, including Robert Louis Stevenson, Arthur Conan Doyle, J. M. Barrie and George MacDonald. In the 20th century the Scottish Renaissance saw a surge of literary activity and attempts to reclaim the Scots language as a medium for serious literature. Members of the movement were followed by a new generation of post-war poets including Edwin Morgan, who would be appointed the first Scots Makar by the inaugural Scottish government in 2004. From the 1980s Scottish literature enjoyed another major revival, particularly associated with a group of writers including Irvine Welsh. Scottish poets who emerged in the same period included Carol Ann Duffy, who, in May 2009, was the first Scot named UK Poet Laureate. As one of the Celtic nations, Scotland and Scottish culture are represented at interceltic events at home and over the world. Scotland hosts several music festivals including Celtic Connections (Glasgow), and the Hebridean Celtic Festival (Stornoway). Festivals celebrating Celtic culture, such as Festival Interceltique de Lorient (Brittany), the Pan Celtic Festival (Ireland), and the National Celtic Festival (Portarlington, Australia), feature elements of Scottish culture such as language, music and dance. The image of St. Andrew, martyred while bound to an X-shaped cross, first appeared in the Kingdom of Scotland during the reign of William I. Following the death of King Alexander III in 1286 an image of Andrew was used on the seal of the Guardians of Scotland who assumed control of the kingdom during the subsequent interregnum. Use of a simplified symbol associated with Saint Andrew, the saltire, has its origins in the late 14th century; the Parliament of Scotland decreeing in 1385 that Scottish soldiers should wear a white Saint Andrew's Cross on the front and back of their tunics. Use of a blue background for the Saint Andrew's Cross is said to date from at least the 15th century. Since 1606 the saltire has also formed part of the design of the Union Flag. There are numerous other symbols and symbolic artefacts, both official and unofficial, including the thistle, the nation's floral emblem (celebrated in the song, The Thistle o' Scotland), the Declaration of Arbroath, incorporating a statement of political independence made on 6 April 1320, the textile pattern tartan that often signifies a particular Scottish clan and the royal Lion Rampant flag. Highlanders can thank James Graham, 3rd Duke of Montrose, for the repeal in 1782 of the Act of 1747 prohibiting the wearing of tartans. Although there is no official national anthem of Scotland, "Flower of Scotland" is played on special occasions and sporting events such as football and rugby matches involving the Scotland national teams and since 2010 is also played at the Commonwealth Games after it was voted the overwhelming favourite by participating Scottish athletes. Other currently less popular candidates for the National Anthem of Scotland include "Scotland the Brave", "Highland Cathedral", "Scots Wha Hae" and "A Man's A Man for A' That". St Andrew's Day, 30 November, is the national day, although Burns' Night tends to be more widely observed, particularly outside Scotland. In 2006, the Scottish Parliament passed the St Andrew's Day Bank Holiday (Scotland) Act 2007, designating the day an official bank holiday. Tartan Day is a recent innovation from Canada. The national animal of Scotland is the unicorn, which has been a Scottish heraldic symbol since the 12th century. Scottish cuisine has distinctive attributes and recipes of its own but shares much with wider British and European cuisine as a result of local and foreign influences, both ancient and modern. Traditional Scottish dishes exist alongside international foodstuffs brought about by migration. Scotland's natural larder of game, dairy products, fish, fruit, and vegetables is the chief factor in traditional Scots cooking, with a high reliance on simplicity and a lack of spices from abroad, as these were historically rare and expensive. Irn-Bru is the most common Scottish carbonated soft drink, often described as "Scotland's other national drink" (after whisky). During the Late Middle Ages and early modern era, French cuisine played a role in Scottish cookery due to cultural exchanges brought about by the "Auld Alliance", especially during the reign of Mary, Queen of Scots. Mary, on her return to Scotland, brought an entourage of French staff who are considered responsible for revolutionising Scots cooking and for some of Scotland's unique food terminology. National newspapers such as the "Daily Record", "The Herald", "The Scotsman" and "The National" are all produced in Scotland. Important regional dailies include the Evening News in Edinburgh, "The Courier" in Dundee in the east, and "The Press and Journal" serving Aberdeen and the north. Scotland is represented at the Celtic Media Festival, which showcases film and television from the Celtic countries. Scottish entrants have won many awards since the festival began in 1980. Television in Scotland is largely the same as UK-wide broadcasts, however, the national broadcaster is BBC Scotland, a constituent part of the British Broadcasting Corporation, the publicly funded broadcaster of the United Kingdom. It runs three national television stations, and the national radio stations, "BBC Radio Scotland" and "BBC Radio nan Gàidheal", amongst others. Scotland also has some programming in the Gaelic language. BBC Alba is the national Gaelic-language channel. The main Scottish commercial television station is STV which broadcasts on two of the three ITV regions of Scotland. Scotland has a number of production companies which produce films and television programmes for Scottish, UK and international audiences. Popular films associated with Scotland through Scottish production or being filmed in Scotland include "Braveheart" (1995), "Highlander" (1986), "Trainspotting" (1996), "Red Road" (2006), "Neds" (2010), "The Angel's Share" (2012), "Brave" (2012) and "Outlaw King" (2018). Popular television programmes associated with Scotland include the long running BBC Scotland soap opera "River City" which has been broadcast since 2002, "Still Game", a popular Scottish sitcom broadcast throughout the United Kingdom (2002–2007, revived in 2016), "Rab C. Nesbitt", "Two Doors Down" and "Take the High Road". Wardpark Studios in Cumbernauld is one of Scotland's television and film production studios where the television programme "Outlander" is produced. Dumbarton Studios, located in Dumbarton is largely used for BBC Scotland programming, used for the filming and production of television programmes such as "Still Game", "River City", "Two Doors Down", and "Shetland". Scotland hosts its own national sporting competitions and has independent representation at several international sporting events, including the FIFA World Cup, the Rugby Union World Cup, the Rugby League World Cup, the Cricket World Cup, the Netball World Cup and the Commonwealth Games. Scotland has its own national governing bodies, such as the Scottish Football Association (the second oldest national football association in the world) and the Scottish Rugby Union. Variations of football have been played in Scotland for centuries, with the earliest reference dating back to 1424. The world's first official international association football match was held in 1872 and was the idea of C. W. Alcock of the Football Association which was seeking to promote Association Football in Scotland. The match took place at the West of Scotland Cricket Club's Hamilton Crescent ground in the Partick area of Glasgow. The match was between Scotland and England and resulted in a 0–0 draw. Following this, the newly developed football became the most popular sport in Scotland. The Scottish Cup was first contested in 1873. Queen's Park F.C., in Glasgow, is probably the oldest association football club in the world outside England. The Scottish Football Association (SFA), the second-oldest national football association in the world, is the main governing body for Scottish association football, and a founding member of the International Football Association Board (IFAB) which governs the Laws of the Game. As a result of this key role in the development of the sport Scotland is one of only four countries to have a permanent representative on the IFAB; the other four representatives being appointed for set periods by FIFA. The SFA also has responsibility for the Scotland national football team, whose supporters are commonly known as the "Tartan Army". , Scotland are ranked as the 50th best national football team in the FIFA World Rankings. The national team last attended the World Cup in France in 1998, but finished last in their group stage. The Scotland women's team have achieved more recent success, qualifying for both Euro 2017 and the 2019 World Cup. , they were ranked as the 22nd best women's national team in the FIFA Rankings. Scottish clubs have achieved some success in European competitions, with Celtic winning the European Cup in 1967, Rangers and Aberdeen winning the UEFA Cup Winners' Cup in 1972 and 1983 respectively, and Aberdeen also winning the UEFA Super Cup in 1983. Celtic, Rangers and Dundee United have also reached European finals, the most recent of these being Rangers in 2008. With the modern game of golf originating in 15th-century Scotland, the country is promoted as the home of golf. To many golfers the Old Course in the Fife town of St Andrews, an ancient links course dating to before 1552, is considered a site of pilgrimage. In 1764, the standard 18-hole golf course was created at St Andrews when members modified the course from 22 to 18 holes. The world's oldest golf tournament, and golf's first major, is The Open Championship, which was first played on 17 October 1860 at Prestwick Golf Club, in Ayrshire, Scotland, with Scottish golfers winning the earliest majors. There are many other famous golf courses in Scotland, including Carnoustie, Gleneagles, Muirfield, and Royal Troon. Other distinctive features of the national sporting culture include the Highland games, curling and shinty. In boxing, Scotland has had 13 world champions, including Ken Buchanan, Benny Lynch and Jim Watt. Scotland has also been successful in motorsport, particularly in Formula One. Notable drivers include; David Coulthard, Jim Clark, Paul Di Resta, and Jackie Stewart. In IndyCar, Dario Franchitti has won 4 consecutive IndyCar world championships. Scotland has competed at every Commonwealth Games since 1930 and has won 356 medals in total—91 Gold, 104 Silver and 161 Bronze. Edinburgh played host to the Commonwealth Games in 1970 and 1986, and most recently Glasgow in 2014. Scotland's primary sources for energy are provided though renewable energy (42%), nuclear (35%) and fossil fuel generation (22%). The Scottish Government has a target to have the equivalent of 50% of the energy for Scotland's heat, transport and electricity consumption to be supplied from renewable sources by 2030. Scotland has five international airports operating scheduled services to Europe, North America and Asia, as well domestic services to England, Northern Ireland and Wales. Highlands and Islands Airports operates eleven airports across the Highlands, Orkney, Shetland and the Western Isles, which are primarily used for short distance, public service operations, although Inverness Airport has a number of scheduled flights to destinations across the UK and mainland Europe. Edinburgh Airport is currently Scotland's busiest airport handling over 13 million passengers in 2017. It is also the UK's 6th busiest airport. British Airways, easyJet, flybe, Jet2, and Ryanair operate the majority of flights between Scotland and other major UK and European airports. Four airlines are based in Scotland: Network Rail owns and operates the fixed infrastructure assets of the railway system in Scotland, while the Scottish Government retains overall responsibility for rail strategy and funding in Scotland. Scotland's rail network has 359 railway stations and around of track. In 2018–19 there were 102million passenger journies on Scottish railways. The East Coast and West Coast main railway lines connect the major cities and towns of Scotland with each other and with the rail network in England. London North Eastern Railway provides inter-city rail journeys between Glasgow, Edinburgh, Aberdeen and Inverness to London. Domestic rail services within Scotland are operated by Abellio ScotRail. During the time of British Rail, the West Coast Main Line from London Euston to Glasgow Central was electrified in the early 1970s, followed by the East Coast Main Line in the late 1980s. British Rail created the ScotRail brand. When British Rail existed, many railway lines in Strathclyde were electrified. Strathclyde Passenger Transport Executive was at the forefront with the acclaimed "largest electrified rail network outside London". Some parts of the network are electrified, but there are no electrified lines in the Highlands, Angus, Aberdeenshire, the cities of Dundee or Aberdeen, or Perth & Kinross, and none of the islands has a rail link (although the railheads at Kyle of Lochalsh and Mallaig principally serve the islands). The East Coast Main Line crosses the Firth of Forth by the Forth Bridge. Completed in 1890, this cantilever bridge has been described as "the one internationally recognised Scottish landmark". Scotland's rail network is managed by Transport Scotland. The Scottish motorways and major trunk roads are managed by Transport Scotland. The remainder of the road network is managed by the Scottish local authorities in each of their areas. Regular ferry services operate between the Scottish mainland and outlying islands. Ferries serving both the inner and outer Hebrides are principally operated by the state-owned enterprise Caledonian MacBrayne. Services to the Northern Isles are operated by Serco. Other routes, served by multiple companies, connect southwest Scotland to Northern Ireland. DFDS Seaways operated a freight-only Rosyth – Zeebrugge ferry service, until a fire damaged the vessel DFDS were using. A passenger service was also operated between 2002–2010. Additional routes are operated by local authorities.
https://en.wikipedia.org/wiki?curid=26994
Scientist A scientist is someone who conducts scientific research to advance knowledge in an area of interest. In classical antiquity, there was no real ancient analog of a modern scientist. Instead, philosophers engaged in the philosophical study of nature called natural philosophy, a precursor of natural science. It was not until the 19th century that the term "scientist" came into regular use after it was coined by the theologian, philosopher, and historian of science William Whewell in 1833. In modern times, many scientists have advanced degrees in an area of science and pursue careers in various sectors of the economy such as academia, industry, government, and nonprofit environments."" The roles of "scientists", and their predecessors before the emergence of modern scientific disciplines, have evolved considerably over time. Scientists of different eras (and before them, natural philosophers, mathematicians, natural historians, natural theologians, engineers, and others who contributed to the development of science) have had widely different places in society, and the social norms, ethical values, and epistemic virtues associated with scientists—and expected of them—have changed over time as well. Accordingly, many different historical figures can be identified as early scientists, depending on which characteristics of modern science are taken to be essential. Some historians point to the Scientific Revolution that began in 16th century as the period when science in a recognizably modern form developed. It wasn't until the 19th century that sufficient socioeconomic changes occurred for scientists to emerge as a major profession. Knowledge about nature in classical antiquity was pursued by many kinds of scholars. Greek contributions to science—including works of geometry and mathematical astronomy, early accounts of biological processes and catalogs of plants and animals, and theories of knowledge and learning—were produced by philosophers and physicians, as well as practitioners of various trades. These roles, and their associations with scientific knowledge, spread with the Roman Empire and, with the spread of Christianity, became closely linked to religious institutions in most of European countries. Astrology and astronomy became an important area of knowledge, and the role of astronomer/astrologer developed with the support of political and religious patronage. By the time of the medieval university system, knowledge was divided into the "trivium"—philosophy, including natural philosophy—and the "quadrivium"—mathematics, including astronomy. Hence, the medieval analogs of scientists were often either philosophers or mathematicians. Knowledge of plants and animals was broadly the province of physicians. Science in medieval Islam generated some new modes of developing natural knowledge, although still within the bounds of existing social roles such as philosopher and mathematician. Many proto-scientists from the Islamic Golden Age are considered polymaths, in part because of the lack of anything corresponding to modern scientific disciplines. Many of these early polymaths were also religious priests and theologians: for example, Alhazen and al-Biruni were mutakallimiin; the physician Avicenna was a hafiz; the physician Ibn al-Nafis was a hafiz, muhaddith and ulema; the botanist Otto Brunfels was a theologian and historian of Protestantism; the astronomer and physician Nicolaus Copernicus was a priest. During the Italian Renaissance scientists like Leonardo Da Vinci, Michelangelo, Galileo Galilei and Gerolamo Cardano have been considered as the most recognizable polymaths. During the Renaissance, Italians made substantial contributions in science. Leonardo Da Vinci made significant discoveries in paleontology and anatomy. The Father of modern Science, Galileo Galilei, made key improvements on the thermometer and telescope which allowed him to observe and clearly describe the solar system. Descartes was not only a pioneer of analytic geometry but formulated a theory of mechanics and advanced ideas about the origins of animal movement and perception. Vision interested the physicists Young and Helmholtz, who also studied optics, hearing and music. Newton extended Descartes' mathematics by inventing calculus (at the same time as Leibniz). He provided a comprehensive formulation of classical mechanics and investigated light and optics. Fourier founded a new branch of mathematics — infinite, periodic series — studied heat flow and infrared radiation, and discovered the greenhouse effect. Girolamo Cardano, Blaise Pascal Pierre de Fermat, Von Neumann, Turing, Khinchin, Markov and Wiener, all mathematicians, made major contributions to science and probability theory, including the ideas behind computers, and some of the foundations of statistical mechanics and quantum mechanics. Many mathematically inclined scientists, including Galileo, were also musicians. There are many compelling stories in medicine and biology, such as the development of ideas about the circulation of blood from Galen to Harvey. During the age of Enlightenment, Luigi Galvani, the pioneer of the bioelectromagnetics, discovered the animal electricity. He discovered that a charge applied to the spinal cord of a frog could generate muscular spasms throughout its body. Charges could make frog legs jump even if the legs were no longer attached to a frog. While cutting a frog leg, Galvani's steel scalpel touched a brass hook that was holding the leg in place. The leg twitched. Further experiments confirmed this effect, and Galvani was convinced that he was seeing the effects of what he called animal electricity, the life force within the muscles of the frog. At the University of Pavia, Galvani's colleague Alessandro Volta was able to reproduce the results, but was sceptical of Galvani's explanation. Lazzaro Spallanzani is one of the most influential figures in experimental physiology and the natural sciences. His investigations have exerted a lasting influence on the medical sciences. He made important contributions to the experimental study of bodily functions and animal reproduction. Francesco Redi discovered that microorganisms can cause disease. Until the late 19th or early 20th century, scientists were still referred to as "natural philosophers" or "men of science". English philosopher and historian of science William Whewell coined the term "scientist" in 1833, and it first appeared in print in Whewell's anonymous 1834 review of Mary Somerville's "On the Connexion of the Physical Sciences" published in the "Quarterly Review". Whewell's suggestion of the term was partly satirical, a response to changing conceptions of science itself in which natural knowledge was increasingly seen as distinct from other forms of knowledge. Whewell wrote of "an increasing proclivity of separation and dismemberment" in the sciences; while highly specific terms proliferated—chemist, mathematician, naturalist—the broad term "philosopher" was no longer satisfactory to group together those who pursued science, without the caveats of "natural" or "experimental" philosopher. Members of the British Association for the Advancement of Science had been complaining about the lack of a good term at recent meetings, Whewell reported in his review; alluding to himself, he noted that "some ingenious gentleman proposed that, by analogy with "artist", they might form [the word] "scientist", and added that there could be no scruple in making free with this term since we already have such words as "economist", and "atheist"—but this was not generally palatable". Whewell proposed the word again more seriously (and not anonymously) in his 1840 "The Philosophy of the Inductive Sciences": He also proposed the term "physicist" at the same time, as a counterpart to the French word "physicien". Neither term gained wide acceptance until decades later; "scientist" became a common term in the late 19th century in the United States and around the turn of the 20th century in Great Britain. By the twentieth century, the modern notion of science as a special brand of information about the world, practiced by a distinct group and pursued through a unique method, was essentially in place. Ramón y Cajal won the Nobel Prize in 1906 for his remarkable observations in neuroanatomy. Marie Curie became the first female to win the Nobel Prize and the first person to win it twice. Her efforts led to the development of nuclear energy and Radiotherapy for the treatment of cancer. In 1922, she was appointed a member of the International Commission on Intellectual Co-operation by the Council of the League of Nations. She campaigned for scientist's right to patent their discoveries and inventions. She also campaigned for free access to international scientific literature and for internationally recognized scientific symbols. As a profession, the scientist of today is widely recognized. In modern times, many professional scientists are trained in an academic setting (e.g., universities and research institutes), mostly at the level of graduate schools. Upon completion, they would normally attain an academic degree, with the highest degree being a doctorate such as a Doctor of Philosophy (PhD). Although graduate education for scientists varies among institutions and countries, some common training requirements include specializing in an area of interest, publishing research findings in peer-reviewed scientific journals and presenting them at scientific conferences, giving lectures or teaching, and defending a thesis (or dissertation) during an oral examination. To aid them in this endeavor, graduate students often work under the guidance of a mentor, usually a senior scientist, which may continue after the completion of their doctorates whereby they work as postdoctoral researchers. After the completion of their training, many scientists pursue careers in a variety of work settings and conditions. In 2017, the British scientific journal "Nature" published the results of a large-scale survey of more than 5,700 doctoral students worldwide, asking them which sectors of the economy they would like to work in. A little over half of the respondents wanted to pursue a career in academia, with smaller proportions hoping to work in industry, government, and nonprofit environments. Scientists are motivated to work in several ways. Many have a desire to understand why the world is as we see it and how it came to be. They exhibit a strong curiosity about reality. Other motivations are recognition by their peers and prestige. The Nobel Prize, a widely regarded prestigious award, is awarded annually to those who have achieved scientific advances in the fields of medicine, physics, chemistry, and economics. Some scientists have a desire to apply scientific knowledge for the benefit of people's health, the nations, the world, nature, or industries (academic scientist and industrial scientist). Scientists tend to be less motivated by direct financial reward for their work than other careers. As a result, scientific researchers often accept lower average salaries when compared with many other professions which require a similar amount of training and qualification. Although there have been exceptions, most scientists tend to do their best research when they are relatively young, in their 30s. Scientists include experimentalists who mainly perform experiments to test hypotheses, and theoreticians who mainly develop models to explain existing data and predict new results. There is a continuum between two activities and the division between them is not clear-cut, with many scientists performing both tasks. Those considering science as a career often look to the frontiers. These include cosmology and biology, especially molecular biology and the human genome project. Other areas of active research include the exploration of matter at the scale of elementary particles as described by high-energy physics, and materials science, which seeks to discover and design new materials. Although there have been remarkable discoveries with regard to brain function and neurotransmitters, the nature of the mind and human thought still remains unknown. The number of scientists is vastly different from country to country. For instance, there are only four full-time scientists per 10,000 workers in India, while this number is 79 for the United Kingdom, and 85 for the United States. According to the National Science Foundation, 4.7 million people with science degrees worked in the United States in 2015, across all disciplines and employment sectors. The figure included twice as many men as women. Of that total, 17% worked in academia, that is, at universities and undergraduate institutions, and men held 53% of those positions. 5% of scientists worked for the federal government, and about 3.5% were self-employed. Of the latter two groups, two-thirds were men. 59% of scientists in the United States were employed in industry or business, and another 6% worked in non-profit positions. Scientist and engineering statistics are usually intertwined, but they indicate that women enter the field far less than men, though this gap is narrowing. The number of science and engineering doctorates awarded to women rose from a mere 7 percent in 1970 to 34 percent in 1985 and in engineering alone the numbers of bachelor's degrees awarded to women rose from only 385 in 1975 to more than 11000 in 1985.
https://en.wikipedia.org/wiki?curid=26997
Smog Smog is a type of intense air pollution. The word "smog" was coined in the early 20th century, and is a contraction (portmanteau) of the words smoke and fog to refer to smoky fog; its opacity, and odor. The word was then intended to refer to what was sometimes known as pea soup fog, a familiar and serious problem in London from the 19th century to the mid-20th century. This kind of visible air pollution is composed of nitrogen oxides, sulphur oxides, ozone, smoke and other particulates. Man-made smog is derived from coal combustion emissions, vehicular emissions, industrial emissions, forest and agricultural fires and photochemical reactions of these emissions. Smog is often categorized as being either summer smog or winter smog. Summer smog is primarily associated with the photochemical formation of ozone. During the summer season when the temperatures are warmer and there is more sunlight present, photochemical smog is the dominant type of smog formation. During the winter months when the temperatures are colder, and atmospheric inversions are common, there is an increase in coal and other fossil fuel usage to heat homes and buildings. These combustion emissions, together with the lack of pollutant dispersion under inversions, characterize winter smog formation. While photochemical smog is the main smog formation mechanism during summer months, winter smog episodes are still common. Smog formation in general relies on both primary and secondary pollutants. Primary pollutants are emitted directly from a source, such as emissions of sulfur dioxide from coal combustion. Secondary pollutants, such as ozone, are formed when primary pollutants undergo chemical reactions in the atmosphere. Photochemical smog, as found for example in Los Angeles, is a type of air pollution derived from vehicular emission from internal combustion engines and industrial fumes. These pollutants react in the atmosphere with sunlight to form secondary pollutants that also combine with the primary emissions to form photochemical smog. In certain other cities, such as Delhi, smog severity is often aggravated by stubble burning in neighboring agricultural areas. The atmospheric pollution levels of Los Angeles, Beijing, Delhi, Lahore, Mexico City, Tehran and other cities are often increased by an inversion that traps pollution close to the ground. The developing smog is usually toxic to humans and can cause severe sickness, a shortened life span, or premature death. Coinage of the term "smog" is often attributed to Dr. Henry Antoine Des Voeux in his 1905 paper, "Fog and Smoke" for a meeting of the Public Health Congress. The 26 July 1905 edition of the London newspaper "Daily Graphic" quoted Des Voeux, "He said it required no science to see that there was something produced in great cities which was not found in the country, and that was smoky fog, or what was known as 'smog'." The following day the newspaper stated that "Dr. Des Voeux did a public service in coining a new word for the London fog." However, the term appears fifteen years earlier than Dr. Voeux's paper, in a column in the July 3, 1880, Santa Cruz Weekly Sentinel. Coal fires can emit significant clouds of smoke that contribute to the formation of winter smog. Coal fires can be used to heat individual buildings or to provide energy in a power-producing plant. Air pollution from this source has been reported in England since the Middle Ages.London, in particular, was notorious up through the mid-20th century for its coal-caused smogs, which were nicknamed 'pea-soupers.' Air pollution of this type is still a problem in areas that generate significant smoke from burning coal. The emissions from coal combustion are one of the main causes of air pollution in China. Especially during autumn and winter when coal-fired heating ramps up, the amount of produced smoke at times forces some Chinese cities to close down roads, schools or airports. One prominent example for this was China's Northeastern city of Harbin in 2013. Traffic emissions – such as from trucks, buses, and automobiles– also contribute to the formation of smog. Airborne by-products from vehicle exhaust systems cause air pollution and are a major ingredient in the creation of smog in some large cities. The major culprits from transportation sources are carbon monoxide (CO),nitrogen oxides (NO and NO2),volatile organic compounds, and hydrocarbons (hydrocarbons are the main component of petroleum fuels such as gasoline and diesel fuel). Transportation emissions also include sulfur dioxides and particulate matter but in much smaller quantities than the pollutants mentioned previously. The nitrogen oxides and volatile organic compounds can undergo a series of chemical reactions with sunlight, heat, ammonia, moisture, and other compounds to form the noxious vapors, ground level ozone, and particles that comprise smog. Photochemical smog, often referred to as "summer smog", is the chemical reaction of sunlight, nitrogen oxides and volatile organic compounds in the atmosphere, which leaves airborne particles and ground-level ozone. Photochemical smog depends on primary pollutants as well as the formation of secondary pollutants. These primary pollutants include nitrogen oxides, particularly nitric oxide (NO) and nitrogen dioxide (NO2), and volatile organic compounds. The relevant secondary pollutants include peroxylacyl nitrates (PAN), tropospheric ozone, and aldehydes. An important secondary pollutant for photochemical smog is ozone, which is formed when hydrocarbons (HC) and nitrogen oxides (NOx) combine in the presence of sunlight; nitrogen dioxide (NO2), which is formed as nitric oxide (NO) combines with oxygen (O2) in the air. In addition, when SO2 and NOx are emitted they eventually are oxidized in the troposphere to nitric acid and sulfuric acid, which, when mixed with water, form the main components of acid rain. All of these harsh chemicals are usually highly reactive and oxidizing. Photochemical smog is therefore considered to be a problem of modern industrialization. It is present in all modern cities, but it is more common in cities with sunny, warm, dry climates and a large number of motor vehicles. Because it travels with the wind, it can affect sparsely populated areas as well. The composition and chemical reactions involved in photochemical smog were not understood until the 1950s. In 1948, flavor chemist Arie Haagen-Smit adapted some of his equipment to collect chemicals from polluted air, and identified ozone as a component of Los Angeles smog. Haagen-Smit went on to discover that nitrogen oxides from automotive exhausts and gaseous hydrocarbons from cars and oil refineries, exposed to sunlight, were key ingredients in the formation of ozone and photochemical smog. Haagen-Smit worked with Arnold Beckman, who developed various equipment for detecting smog, ranging from an "Apparatus for recording gas concentrations in the atmosphere" patented on October 7, 1952, to "air quality monitoring vans" for use by government and industry. During the morning rush hour, a high concentration of nitric oxide and hydrocarbons are emitted to the atmosphere, mostly via on-road traffic but also from industrial sources. Some hydrocarbons are rapidly oxidized by OH· and form peroxy radicals, which convert nitric oxide (NO) to nitrogen dioxide (NO2). (1) R{.} + O2 + M -> RO2{.} + M (2) RO2{.} + NO -> NO2 + RO{.} (3) HO2{.} + NO -> NO2 + OH{.} Nitrogen dioxide (NO2) and nitric oxide (NO) further react with ozone (O3) in a series of chemical reactions: (4) NO2 + hv -> O(^3P) + NO, formula_1 (5) O(^3P) + O2 + M-> O3 + M(heat) (6) O3 + NO -> NO2 + O2 This series of equations is referred to as the photostationary state (PSS). However, because of the presence of Reaction 2 and 3, NOx and ozone are not in a perfect steady state. By replacing Reaction 6 with Reaction 2 and Reaction 3, the O3 molecule is no longer destroyed. Therefore, the concentration of ozone keeps increasing throughout the day. This mechanism can escalate the formation of ozone in smog. Other reactions such as the photooxidation of formaldehyde (HCHO), a common secondary pollutant, can also contribute to the increased concentration of ozone and NO2. Photochemical smog is more prevalent during summer days since incident solar radiation fluxes are high, which favors the formation of ozone (reactions 4 and 5). The presence of a temperature inversion layer is another important factor. That is because it prevents the vertical convective mixing of the air and thus allows the pollutants, including ozone, to accumulate near the ground level, which again favors the formation of photochemical smog. There are certain reactions that can limit the formation of O3 in smog. The main limiting reaction in polluted areas is: (7) NO2 + OH{.} + M -> HNO3 + M This reaction removes NO2 which limits the amount of O3 that can be produced from its photolysis (reaction 4). HNO3 is a sticky compound that can easily be removed onto surfaces (dry deposition) or dissolved in water and be rained out (wet deposition). Both ways are common in the atmosphere and can efficiently remove the radicals and nitrogen dioxide. An erupting volcano can emit high levels of sulphur dioxide along with a large quantity of particulates matter; two key components to the creation of smog. However, the smog created as a result of a volcanic eruption is often known as vog to distinguish it as a natural occurrence. The chemical reactions that form smog following a volcanic eruption are different than the reactions that form photochemical smog. The term smog encompasses the effect when a large amount of gas phase molecules and particulate matter are emitted to the atmosphere, creating a visible haze. The event causing a large amount of emissions can vary but still result in the formation of smog. Plants are another natural source of hydrocarbons that could undergo reactions in the atmosphere and produce smog. Globally both plants and soil contribute a substantial amount to the production of hydrocarbons, mainly by producing isoprene and terpenes. Hydrocarbons released by plants can often be more reactive than man-made hydrocarbons. For example when plants release isoprene, the isoprene reacts very quickly in the atmosphere with hydroxyl radicals. These reactions produce hydroperoxides which increase ozone formation. Smog is a serious problem in many cities and continues to harm human health. Ground-level ozone, sulphur dioxide, nitrogen dioxide and carbon monoxide are especially harmful for senior citizens, children, and people with heart and lung conditions such as emphysema, bronchitis, and asthma. It can inflame breathing passages, decrease the lungs' working capacity, cause shortness of breath, pain when inhaling deeply, wheezing, and coughing. It can cause eye and nose irritation and it dries out the protective membranes of the nose and throat and interferes with the body's ability to fight infection, increasing susceptibility to illness. Hospital admissions and respiratory deaths often increase during periods when ozone levels are high. There is a lack of knowledge on the long-term effects of air pollution exposure and the origin of asthma. An experiment was carried out using intense air pollution similar to that of the 1952 Great Smog of London. The results from this experiment concluded that there is a link between early-life pollution exposure that leads to the development of asthma, proposing the ongoing effect of the Great Smog. Modern studies continue to find links between mortality and the presence of smog. One study, published in Nature magazine, found that smog episodes in the city of Jinan, a large city in eastern China, during 2011–15, were associated with a 5.87% (95% CI 0.16–11.58%) increase in the rate of overall mortality. This study highlights the effect of exposure to air pollution on the rate of mortality in China. The U.S. EPA has developed an air quality index to help explain air pollution levels to the general public. 8 hour average ozone concentrations of 85 to 104 ppbv are described as "Unhealthy for Sensitive Groups", 105 ppbv to 124 ppbv as "unhealthy" and 125 ppb to 404 ppb as "very unhealthy". The "very unhealthy" range for some other pollutants are: 355 μg m−3 – 424 μg m−3 for PM10; 15.5 ppm – 30.4ppm for CO and 0.65 ppm – 1.24 ppm for NO2. In 2016, the Ontario Medical Association announced that smog is responsible for an estimated 9,500 premature deaths in the province each year. A 20-year American Cancer Society study found that cumulative exposure also increases the likelihood of premature death from a respiratory disease, implying the 8-hour standard may be insufficient. Tiny magnetic particles from air pollution have for the first time been discovered to be lodged in human brains– and researchers think they could be a possible cause of Alzheimer's disease. Researchers at Lancaster University found abundant magnetite nanoparticles in the brain tissue from 37 individuals aged three to 92-years-old who lived in Mexico City and Manchester. This strongly magnetic mineral is toxic and has been implicated in the production of reactive oxygen species (free radicals) in the human brain, which are associated with neurodegenerative diseases including Alzheimer's disease. A study examining 806 women who had babies with birth defects between 1997 and 2006, and 849 women who had healthy babies, found that smog in the San Joaquin Valley area of California was linked to two types of neural tube defects: spina bifida (a condition involving, among other manifestations, certain malformations of the spinal column), and anencephaly (the underdevelopment or absence of part or all of the brain, which if not fatal usually results in profound impairment). According to a study published in The Lancet, even a very small (5 μg) change in PM2.5 exposure was associated with an increase (18%) in risk of a low birth weight at delivery, and this relationship held even below the current accepted safe levels. Smog can form in almost any climate where industries or cities release large amounts of air pollution, such as smoke or gases. However, it is worse during periods of warmer, sunnier weather when the upper air is warm enough to inhibit vertical circulation. It is especially prevalent in geologic basins encircled by hills or mountains. It often stays for an extended period of time over densely populated cities or urban areas, and can build up to dangerous levels. According to the Canadian Science Smog Assessment published in 2012, smog is responsible for detrimental effects on human and ecosystem health, as well as socioeconomic well-being across the country. It was estimated that the province of Ontario sustains $201 million in damages annually for selected crops, and an estimated tourism revenue degradation of $7.5 million in Vancouver and $1.32 million in The Fraser Valley due to decreased visibility. Air pollution in British Columbia is of particular concern, especially in the Fraser Valley, because of a meteorological effect called inversion which decreases air dispersion and leads to smog concentration. For the past few years, cities in northern India have been covered in a thick layer of winter smog. The situation has turned quite drastic in the National Capital, Delhi. This smog is caused by the collection of Particulate Matter (a very fine type of dust and toxic gases) in the air due to stagnant movement of air during winters. Delhi is the most polluted city in the world and according to one estimate, air pollution causes the death of about 10,500 people in Delhi every year. During 2013–14, peak levels of fine particulate matter (PM) in Delhi increased by about 44%, primarily due to high vehicular and industrial emissions, construction work and crop burning in adjoining states. Delhi has the highest level of the airborne particulate matter, PM2.5 considered most harmful to health, with 153 micrograms. Rising air pollution level has significantly increased lung-related ailments (especially asthma and lung cancer) among Delhi's children and women. The dense smog in Delhi during winter season results in major air and rail traffic disruptions every year. According to Indian meteorologists, the average maximum temperature in Delhi during winters has declined notably since 1998 due to rising air pollution. Environmentalists have criticised the Delhi government for not doing enough to curb air pollution and to inform people about air quality issues. Most of Delhi's residents are unaware of alarming levels of air pollution in the city and the health risks associated with it. Since the mid-1990s, Delhi has undertaken some measures to curb air pollution – Delhi has the third highest quantity of trees among Indian cities and the Delhi Transport Corporation operates the world's largest fleet of environmentally friendly compressed natural gas (CNG) buses. In 1996, the Centre for Science and Environment (CSE) started a public interest litigation in the Supreme Court of India that ordered the conversion of Delhi's fleet of buses and taxis to run on CNG and banned the use of leaded petrol in 1998. In 2003, Delhi won the United States Department of Energy's first 'Clean Cities International Partner of the Year' award for its "bold efforts to curb air pollution and support alternative fuel initiatives". The Delhi Metro has also been credited for significantly reducing air pollutants in the city. However, according to several authors, most of these gains have been lost, especially due to stubble burning, rise in market share of diesel cars and a considerable decline in bus ridership. According to CUE and System of Air Quality Weather Forecasting and Research (SAFER), burning of agricultural waste in nearby Punjab, Haryana and Uttar Pradesh regions results in severe intensification of smog over Delhi. The state government of adjoining Uttar Pradesh is considering imposing a ban on crop burning to reduce pollution in Delhi NCR and an environmental panel has appealed to India's Supreme Court to impose a 30% cess on diesel cars. Joint research between American and Chinese researchers in 2006 concluded that much of the city's pollution comes from surrounding cities and provinces. On average 35–60% of the ozone can be traced to sources outside the city. Shandong Province and Tianjin Municipality have a "significant influence on Beijing's air quality", partly due to the prevailing south/southeasterly flow during the summer and the mountains to the north and northwest. In 1306, concerns over air pollution were sufficient for Edward I to (briefly) ban coal fires in London. In 1661, John Evelyn's "Fumifugium" suggested burning fragrant wood instead of mineral coal, which he believed would reduce coughing. The "" the same year describes how the smoke "does our lungs and spirits choke, Our hanging spoil, and rust our iron." Severe episodes of smog continued in the 19th and 20th centuries, mainly in the winter, and were nicknamed "pea-soupers," from the phrase "as thick as pea soup". The Great Smog of 1952 darkened the streets of London and killed approximately 4,000 people in the short time of four days (a further 8,000 died from its effects in the following weeks and months). Initially a flu epidemic was blamed for the loss of life. In 1956 the Clean Air Act started legally enforcing smokeless zones in the capital. There were areas where no soft coal was allowed to be burned in homes or in businesses, only coke, which produces no smoke. Because of the smokeless zones, reduced levels of sooty particulates eliminated the intense and persistent London smog. It was after this that the great clean-up of London began. One by one, historical buildings which, during the previous two centuries had gradually completely blackened externally, had their stone facades cleaned and restored to their original appearance. Victorian buildings whose appearance changed dramatically after cleaning included the British Museum of Natural History. A more recent example was the Palace of Westminster, which was cleaned in the 1980s. A notable exception to the restoration trend was 10 Downing Street, whose bricks upon cleaning in the late 1950s proved to be naturally "yellow"; the smog-derived black colour of the façade was considered so iconic that the bricks were painted black to preserve the image. Smog caused by traffic pollution, however, does still occur in modern London. Other areas of the United Kingdom were affected by smog, especially heavily industrialised areas. The cities of Glasgow and Edinburgh, in Scotland, suffered smoke-laden fogs in 1909. Des Voeux, commonly credited with creating the "smog" moniker, presented a paper in 1911 to the Manchester Conference of the Smoke Abatement League of Great Britain about the fogs and resulting deaths. One Birmingham resident described near black-out conditions in the 1900s before the Clean Air Act, with visibility so poor that cyclists had to dismount and walk in order to stay on the road. On 29 April 2015, the UK Supreme Court ruled that the government must take immediate action to cut air pollution, following a case brought by environmental lawyers at ClientEarth. Due to its location in a highland "bowl", cold air sinks down onto the urban area of Mexico City, trapping industrial and vehicle pollution underneath, and turning it into the most infamously smog-plagued city of Latin America. Within one generation, the city has changed from being known for some of the cleanest air of the world into one with some of the worst pollution, with pollutants like nitrogen dioxide being double or even triple international standards. Similar to Mexico City, the air pollution of Santiago valley, located between the Andes and the Chilean Coast Range, turn it into the most infamously smog-plagued city of South America. Other aggravates of the situation reside in its high latitude (31 degrees South) and dry weather during most of the year. In December 2005, schools and public offices had to close in Tehran and 1600 people were taken to hospital, in a severe smog blamed largely on unfiltered car exhaust. Smog was brought to the attention of the general U.S. public in 1933 with the publication of the book "Stop That Smoke", by Henry Obermeyer, a New York public utility official, in which he pointed out the effect on human life and even the destruction of of a farmer's spinach crop. Since then, the United States Environmental Protection Agency has designated over 300 U.S. counties to be non-attainment areas for one or more pollutants tracked as part of the National Ambient Air Quality Standards. These areas are largely clustered around large metropolitan areas, with the largest contiguous non-attainment zones in California and the Northeast. Various U.S. and Canadian government agencies collaborate to produce real-time air quality maps and forecasts. To combat smog conditions, localities may declare "smog alert" days, such as in the Spare the Air program in the San Francisco Bay Area. In the United States, smog pollution kills 24,000 Americans every year. The U.S. is among the dirtier countries in terms of smog, ranked 123 out of 195 countries measured, where 1 is cleanest and 195 is most smog polluted. Because of their locations in low basins surrounded by mountains, Los Angeles and the San Joaquin Valley are notorious for their smog. Heavy automobile traffic, combined with the additional effects of the San Francisco Bay and Los Angeles/Long Beach port complexes, frequently contribute to further air pollution. Los Angeles in particular is strongly predisposed to accumulation of smog, because of peculiarities of its geography and weather patterns. Los Angeles is situated in a flat basin with ocean on one side and mountain ranges on three sides. A nearby cold ocean current depresses surface air temperatures in the area, resulting in an inversion layer: a phenomenon where air temperature increases, instead of decreasing, with altitude, suppressing thermals and restricting vertical convection. All taken together, this results in a relatively thin, enclosed layer of air above the city that cannot easily escape out of the basin and tends to accumulate pollution. Los Angeles was one of the best known cities suffering from transportation smog for much of the 20th century, so much so that it was sometimes said that "Los Angeles" was a synonym for "smog." In 1970, when the Clean Air Act was passed, Los Angeles was the most polluted basin in the country, and California was unable to create a State Implementation Plan that would enable it to meet the new air quality standards. However, ensuing strict regulations by state and federal government agencies overseeing this problem (such as the California Air Resources Board and the United States Environmental Protection Agency), including tight restrictions on allowed emissions levels for all new cars sold in California and mandatory regular emission tests of older vehicles, resulted in significant improvements in air quality. For example, air concentrations of volatile organic compounds declined by a factor of 50 between 1962 and 2012. Concentrations of air pollutants such as nitrous oxides and ozone declined by 70% to 80% over the same period of time. In the late 1990s, massive immigration to Ulaanbaatar from the countryside began. An estimated 150,000 households, mainly living in traditional Mongolian gers on the outskirts of Ulaanbaatar, burn wood and coal (some poor families burn even car tires and trash) to heat themselves during the harsh winter, which lasts from October to April, since these outskirts are not connected to the city's central heating system. A temporary solution to decrease smog was proposed in the form of stoves with improved efficiency, although with no visible results. Coal-fired ger stoves release high levels of ash and other particulate matter (PM). When inhaled, these particles can settle in the lungs and respiratory tract and cause health problems. At two to 10 times above Mongolian and international air quality standards, Ulaanbaatar's PM rates are among the worst in the world, according to a December 2009 World Bank report. The Asian Development Bank (ADB) estimates that health costs related to this air pollution account for as much as 4 percent of Mongolia's GDP. Smog is a regular problem in Southeast Asia caused by land and forest fires in Indonesia, especially Sumatra and Kalimantan, although the term haze is preferred in describing the problem. Farmers and plantation owners are usually responsible for the fires, which they use to clear tracts of land for further plantings. Those fires mainly affect Brunei, Indonesia, Philippines, Malaysia, Singapore and Thailand, and occasionally Guam and Saipan. The economic losses of the fires in 1997 have been estimated at more than US$9 billion. This includes damages in agriculture production, destruction of forest lands, health, transportation, tourism, and other economic endeavours. Not included are social, environmental, and psychological problems and long-term health effects. The second-latest bout of haze to occur in Malaysia, Singapore and the Malacca Straits is in October 2006, and was caused by smoke from fires in Indonesia being blown across the Straits of Malacca by south-westerly winds. A similar haze has occurred in June 2013, with the PSI setting a new record in Singapore on June 21 at 12pm with a reading of 401, which is in the "Hazardous" range. The Association of Southeast Asian Nations (ASEAN) reacted. In 2002, the Agreement on Transboundary Haze Pollution was signed between all ASEAN nations. ASEAN formed a Regional Haze Action Plan (RHAP) and established a co-ordination and support unit (CSU). RHAP, with the help of Canada, established a monitoring and warning system for forest/vegetation fires and implemented a Fire Danger Rating System (FDRS). The Malaysian Meteorological Department (MMD) has issued a daily rating of fire danger since September 2003. Indonesia has been ineffective at enforcing legal policies on errant farmers. Since start of winter season heavy smog loaded with pollutants covered major part of Punjab especially the city of Lahore, causing breathing problems and disrupting normal traffic. Doctors advised residents to stay indoors and wear facemasks outside. The severity of smog is often measured using automated optical instruments such as Nephelometers, as haze is associated with visibility and traffic control in ports. Haze however can also be an indication of poor air quality though this is often better reflected using accurate purpose built air indexes such as the American Air Quality Index, the Malaysian API (Air Pollution Index) and the Singaporean Pollutant Standards Index. In hazy conditions, it is likely that the index will report the suspended particulate level. The disclosure of the responsible pollutant is mandated in some jurisdictions. The Malaysian API does not have a capped value; hence its most hazardous readings can go above 500. Above 500, a state of emergency is declared in the affected area. Usually, this means that non-essential government services are suspended, and all ports in the affected area are closed. There may also be prohibitions on private sector commercial and industrial activities in the affected area excluding the food sector. So far, state of emergency rulings due to hazardous API levels were applied to the Malaysian towns of Port Klang, Kuala Selangor and the state of Sarawak during the 2005 Malaysian haze and the 1997 Southeast Asian haze.
https://en.wikipedia.org/wiki?curid=27000
Smoke Smoke is a collection of airborne particulates and gases emitted when a material undergoes combustion or pyrolysis, together with the quantity of air that is entrained or otherwise mixed into the mass. It is commonly an unwanted by-product of fires (including stoves, candles, internal combustion engines, oil lamps, and fireplaces), but may also be used for pest control (fumigation), communication (smoke signals), defensive and offensive capabilities in the military (smoke screen), cooking, or smoking (tobacco, cannabis, etc.). It is used in rituals where incense, sage, or resin is burned to produce a smell for spiritual or magical purposes. It can also be a flavoring agent and preservative for various foodstuffs. Smoke inhalation is the primary cause of death in victims of indoor fires. The smoke kills by a combination of thermal damage, poisoning and pulmonary irritation caused by carbon monoxide, hydrogen cyanide and other combustion products. Smoke is an aerosol (or mist) of solid particles and liquid droplets that are close to the ideal range of sizes for Mie scattering of visible light. The composition of smoke depends on the nature of the burning fuel and the conditions of combustion. Fires with high availability of oxygen burn at a high temperature and with a small amount of smoke produced; the particles are mostly composed of ash, or with large temperature differences, of condensed aerosol of water. High temperature also leads to production of nitrogen oxides. Sulfur content yields sulfur dioxide, or in case of incomplete combustion, hydrogen sulfide. Carbon and hydrogen are almost completely oxidized to carbon dioxide and water. Fires burning with lack of oxygen produce a significantly wider palette of compounds, many of them toxic. Partial oxidation of carbon produces carbon monoxide, while nitrogen-containing materials can yield hydrogen cyanide, ammonia, and nitrogen oxides. Hydrogen gas can be produced instead of water. Contents of halogens such as chlorine (e.g. in polyvinyl chloride or brominated flame retardants) may lead to the production of hydrogen chloride, phosgene, dioxin, and chloromethane, bromomethane and other halocarbons. Hydrogen fluoride can be formed from fluorocarbons, whether fluoropolymers subjected to fire or halocarbon fire suppression agents. Phosphorus and antimony oxides and their reaction products can be formed from some fire retardant additives, increasing smoke toxicity and corrosivity. Pyrolysis of polychlorinated biphenyls (PCB), e.g. from burning older transformer oil, and to lower degree also of other chlorine-containing materials, can produce 2,3,7,8-tetrachlorodibenzodioxin, a potent carcinogen, and other polychlorinated dibenzodioxins. Pyrolysis of fluoropolymers, e.g. teflon, in presence of oxygen yields carbonyl fluoride (which hydrolyzes readily to HF and CO2); other compounds may be formed as well, e.g. carbon tetrafluoride, hexafluoropropylene, and highly toxic perfluoroisobutene (PFIB). Pyrolysis of burning material, especially incomplete combustion or smoldering without adequate oxygen supply, also results in production of a large amount of hydrocarbons, both aliphatic (methane, ethane, ethylene, acetylene) and aromatic (benzene and its derivates, polycyclic aromatic hydrocarbons; e.g. benzo[a]pyrene, studied as a carcinogen, or retene), terpenes. Heterocyclic compounds may be also present. Heavier hydrocarbons may condense as tar; smoke with significant tar content is yellow to brown. Presence of such smoke, soot, and/or brown oily deposits during a fire indicates a possible hazardous situation, as the atmosphere may be saturated with combustible pyrolysis products with concentration above the upper flammability limit, and sudden inrush of air can cause flashover or backdraft. Presence of sulfur can lead to formation of e.g. hydrogen sulfide, carbonyl sulfide, sulfur dioxide, carbon disulfide, and thiols; especially thiols tend to get adsorbed on surfaces and produce a lingering odor even long after the fire. Partial oxidation of the released hydrocarbons yields in a wide palette of other compounds: aldehydes (e.g. formaldehyde, acrolein, and furfural), ketones, alcohols (often aromatic, e.g. phenol, guaiacol, syringol, catechol, and cresols), carboxylic acids (formic acid, acetic acid, etc.). The visible particulate matter in such smokes is most commonly composed of carbon (soot). Other particulates may be composed of drops of condensed tar, or solid particles of ash. The presence of metals in the fuel yields particles of metal oxides. Particles of inorganic salts may also be formed, e.g. ammonium sulfate, ammonium nitrate, or sodium chloride. Inorganic salts present on the surface of the soot particles may make them hydrophilic. Many organic compounds, typically the aromatic hydrocarbons, may be also adsorbed on the surface of the solid particles. Metal oxides can be present when metal-containing fuels are burned, e.g. solid rocket fuels containing aluminium. Depleted uranium projectiles after impacting the target ignite, producing particles of uranium oxides. Magnetic particles, spherules of magnetite-like ferrous ferric oxide, are present in coal smoke; their increase in deposits after 1860 marks the beginning of the Industrial Revolution. (Magnetic iron oxide nanoparticles can be also produced in the smoke from meteorites burning in the atmosphere.) Magnetic remanence, recorded in the iron oxide particles, indicates the strength of Earth's magnetic field when they were cooled beyond their Curie temperature; this can be used to distinguish magnetic particles of terrestrial and meteoric origin. Fly ash is composed mainly of silica and calcium oxide. Cenospheres are present in smoke from liquid hydrocarbon fuels. Minute metal particles produced by abrasion can be present in engine smokes. Amorphous silica particles are present in smokes from burning silicones; small proportion of silicon nitride particles can be formed in fires with insufficient oxygen. The silica particles have about 10 nm size, clumped to 70-100 nm aggregates and further agglomerated to chains. Radioactive particles may be present due to traces of uranium, thorium, or other radionuclides in the fuel; hot particles can be present in case of fires during nuclear accidents (e.g. Chernobyl disaster) or nuclear war. Smoke particulates, like other aerosols, are categorized into three modes based on particle size: Most of the smoke material is primarily in coarse particles. Those undergo rapid dry precipitation, and the smoke damage in more distant areas outside of the room where the fire occurs is therefore primarily mediated by the smaller particles. Aerosol of particles beyond visible size is an early indicator of materials in a preignition stage of a fire. Burning of hydrogen-rich fuel produces water; this results in smoke containing droplets of water vapor. In absence of other color sources (nitrogen oxides, particulates...), such smoke is white and cloud-like. Smoke emissions may contain characteristic trace elements. Vanadium is present in emissions from oil fired power plants and refineries; oil plants also emit some nickel. Coal combustion produces emissions containing aluminium, arsenic, chromium, cobalt, copper, iron, mercury, selenium, and uranium. Traces of vanadium in high-temperature combustion products form droplets of molten vanadates. These attack the passivation layers on metals and cause high temperature corrosion, which is a concern especially for internal combustion engines. Molten sulfate and lead particulates also have such effect. Some components of smoke are characteristic of the combustion source. Guaiacol and its derivatives are products of pyrolysis of lignin and are characteristic of wood smoke; other markers are syringol and derivates, and other methoxy phenols. Retene, a product of pyrolysis of conifer trees, is an indicator of forest fires. Levoglucosan is a pyrolysis product of cellulose. Hardwood vs softwood smokes differ in the ratio of guaiacols/syringols. Markers for vehicle exhaust include polycyclic aromatic hydrocarbons, hopanes, steranes, and specific nitroarenes (e.g. 1-nitropyrene). The ratio of hopanes and steranes to elemental carbon can be used to distinguish between emissions of gasoline and diesel engines. Many compounds can be associated with particulates; whether by being adsorbed on their surfaces, or by being dissolved in liquid droplets. Hydrogen chloride is well absorbed in the soot particles. Inert particulate matter can be disturbed and entrained into the smoke. Of particular concern are particles of asbestos. Deposited hot particles of radioactive fallout and bioaccumulated radioisotopes can be reintroduced into the atmosphere by wildfires and forest fires; this is a concern in e.g. the Zone of alienation containing contaminants from the Chernobyl disaster. Polymers are a significant source of smoke. Aromatic side groups, e.g. in polystyrene, enhance generation of smoke. Aromatic groups integrated in the polymer backbone produce less smoke, likely due to significant charring. Aliphatic polymers tend to generate the least smoke, and are non-self-extinguishing. However presence of additives can significantly increase smoke formation. Phosphorus-based and halogen-based flame retardants decrease production of smoke. Higher degree of cross-linking between the polymer chains has such effect too. The naked eye detects particle sizes greater than 7 µm (micrometres). Visible particles emitted from a fire are referred to as smoke. Invisible particles are generally referred to as gas or fumes. This is best illustrated when toasting bread in a toaster. As the bread heats up, the products of combustion increase in size. The fumes initially produced are invisible but become visible if the toast is burnt. An ionization chamber type smoke detector is technically a product of combustion detector, not a smoke detector. Ionization chamber type smoke detectors detect particles of combustion that are invisible to the naked eye. This explains why they may frequently false alarm from the fumes emitted from the red-hot heating elements of a toaster, before the presence of visible smoke, yet they may fail to activate in the early, low-heat smoldering stage of a fire. Smoke from a typical house fire contains hundreds of different chemicals and fumes. As a result, the damage caused by the smoke can often exceed that caused by the actual heat of the fire. In addition to the physical damage caused by the smoke of a fire – which manifests itself in the form of stains – is the often even harder to eliminate problem of a smoky odor. Just as there are contractors that specialize in rebuilding/repairing homes that have been damaged by fire and smoke, fabric restoration companies specialize in restoring fabrics that have been damaged in a fire. Smoke from oxygen-deprived fires contains a significant concentration of compounds that are flammable. A cloud of smoke, in contact with atmospheric oxygen, therefore has the potential of being ignited – either by another open flame in the area, or by its own temperature. This leads to effects like backdraft and flashover. Smoke inhalation is also a danger of smoke that can cause serious injury and death. Many compounds of smoke from fires are highly toxic and/or irritating. The most dangerous is carbon monoxide leading to carbon monoxide poisoning, sometimes with the additive effects of hydrogen cyanide and phosgene. Smoke inhalation can therefore quickly lead to incapacitation and loss of consciousness. Sulfur oxides, hydrogen chloride and hydrogen fluoride in contact with moisture form sulfuric, hydrochloric and hydrofluoric acid, which are corrosive to both lungs and materials. When asleep the nose does not sense smoke nor does the brain, but the body will wake up if the lungs become enveloped in smoke and the brain will be stimulated and the person will be awoken. This does not work if the person is incapacitated or under the influence of drugs and/or alcohol. Cigarette smoke is a major modifiable risk factor for lung disease, heart disease, and many cancers. Smoke can also be a component of ambient air pollution due to the burning of coal in power plants, forest fires or other sources, although the concentration of pollutants in ambient air is typically much less than that in cigarette smoke. One day of exposure to PM2.5 at a concentration of 880 μg/m3, such as occurs in Beijing, China, is the equivalent of smoking one or two cigarettes in terms of particulate inhalation by weight. The analysis is complicated, however, by the fact that the organic compounds present in various ambient particulates may have a higher carcinogenicity than the compounds in cigarette smoke particulates. Secondhand tobacco smoke is the combination of both sidestream and mainstream smoke emissions from a burning tobacco product. These emissions contain more than 50 carcinogenic chemicals. According to the Surgeon General's 2006 report on the subject, "Short exposures to secondhand [tobacco] smoke can cause blood platelets to become stickier, damage the lining of blood vessels, decrease coronary flow velocity reserves, and reduce heart variability, potentially increasing the risk of a heart attack". The American Cancer Society lists "heart disease, lung infections, increased asthma attacks, middle ear infections, and low birth weight" as ramifications of smoker's emission. Smoke can obscure visibility, impeding occupant exiting from fire areas. In fact, the poor visibility due to the smoke that was in the Worcester Cold Storage Warehouse fire in Worcester, Massachusetts was the reason why the trapped rescue firefighters couldn't evacuate the building in time. Because of the striking similarity that each floor shared, the dense smoke caused the firefighters to become disoriented. Smoke contains a wide variety of chemicals, many of them aggressive in nature. Examples are hydrochloric acid and hydrobromic acid, produced from halogen-containing plastics and fire retardants, hydrofluoric acid released by pyrolysis of fluorocarbon fire suppression agents, sulfuric acid from burning of sulfur-containing materials, nitric acid from high-temperature fires where nitrous oxide gets formed, phosphoric acid and antimony compounds from P and Sb based fire retardants, and many others. Such corrosion is not significant for structural materials, but delicate structures, especially microelectronics, are strongly affected. Corrosion of circuit board traces, penetration of aggressive chemicals through the casings of parts, and other effects can cause an immediate or gradual deterioration of parameters or even premature (and often delayed, as the corrosion can progress over long time) failure of equipment subjected to smoke. Many smoke components are also electrically conductive; deposition of a conductive layer on the circuits can cause crosstalks and other deteriorations of the operating parameters or even cause short circuits and total failures. Electrical contacts can be affected by corrosion of surfaces, and by deposition of soot and other conductive particles or nonconductive layers on or across the contacts. Deposited particles may adversely affect the performance of optoelectronics by absorbing or scattering the light beams. Corrosivity of smoke produced by materials is characterized by the corrosion index (CI), defined as material loss rate (angstrom/minute) per amount of material gasified products (grams) per volume of air (m3). It is measured by exposing strips of metal to flow of combustion products in a test tunnel. Polymers containing halogen and hydrogen (polyvinyl chloride, polyolefins with halogenated additives, etc.) have the highest CI as the corrosive acids are formed directly with water produced by the combustion, polymers containing halogen only (e.g. polytetrafluoroethylene) have lower CI as the formation of acid is limited to reactions with airborne humidity, and halogen-free materials (polyolefins, wood) have the lowest CI. However, some halogen-free materials can also release significant amount of corrosive products. Smoke damage to electronic equipment can be significantly more extensive than the fire itself. Cable fires are of special concern; low smoke zero halogen materials are preferable for cable insulation. When smoke comes into contact with the surface of any substance or structure, the chemicals contained in it are transferred to it. The corrosive properties of the chemicals cause the substance or structure to decompose at a rapid rate. Certain materials or structures absorb these chemicals, which is why clothing, unsealed surfaces, potable water, piping, wood, etc., are replaced in most cases of structural fires. As early as the 15th century Leonardo da Vinci commented at length on the difficulty of assessing smoke, and distinguished between black smoke (carbonized particles) and white 'smoke' which is not a smoke at all but merely a suspension of harmless water particulates. Smoke from heating appliances is commonly measured in one of the following ways: In-line capture. A smoke sample is simply sucked through a filter which is weighed before and after the test and the mass of smoke found. This is the simplest and probably the most accurate method, but can only be used where the smoke concentration is slight, as the filter can quickly become blocked. The ASTM smoke pump is a simple and widely used method of in-line capture where a measured volume of smoke is pulled through a filter paper and the dark spot so formed is compared with a standard. Filter/dilution tunnel. A smoke sample is drawn through a tube where it is diluted with air, the resulting smoke/air mixture is then pulled through a filter and weighed. This is the internationally recognized method of measuring smoke from combustion. Electrostatic precipitation. The smoke is passed through an array of metal tubes which contain suspended wires. A (huge) electrical potential is applied across the tubes and wires so that the smoke particles become charged and are attracted to the sides of the tubes. This method can over-read by capturing harmless condensates, or under-read due to the insulating effect of the smoke. However, it is the necessary method for assessing volumes of smoke too great to be forced through a filter, i.e., from bituminous coal. Ringelmann scale. A measure of smoke color. Invented by Professor Maximilian Ringelmann in Paris in 1888, it is essentially a card with squares of black, white and shades of gray which is held up and the comparative grayness of the smoke judged. Highly dependent on light conditions and the skill of the observer it allocates a grayness number from 0 (white) to 5 (black) which has only a passing relationship to the actual quantity of smoke. Nonetheless, the simplicity of the Ringelmann scale means that it has been adopted as a standard in many countries. Optical scattering. A light beam is passed through the smoke. A light detector is situated at an angle to the light source, typically at 90°, so that it receives only light reflected from passing particles. A measurement is made of the light received which will be higher as the concentration of smoke particles becomes higher. Optical obscuration. A light beam is passed through the smoke and a detector opposite measures the light. The more smoke particles are present between the two, the less light will be measured. Combined optical methods. There are various proprietary optical smoke measurement devices such as the 'nephelometer' or the 'aethalometer' which use several different optical methods, including more than one wavelength of light, inside a single instrument and apply an algorithm to give a good estimate of smoke. It has been claimed that these devices can differentiate types of smoke and so their probable source can be inferred, though this is disputed. Inference from carbon monoxide. Smoke is incompletely burned fuel, carbon monoxide is incompletely burned carbon, therefore it has long been assumed that measurement of CO in flue gas (a cheap, simple and very accurate procedure) will provide a good indication of the levels of smoke. Indeed, several jurisdictions use CO measurement as the basis of smoke control. However it is far from clear how accurate the correspondence is. Throughout recorded history, humans have used the smoke of medicinal plants to cure illness. A sculpture from Persepolis shows Darius the Great (522–486 BC), the king of Persia, with two censers in front of him for burning Peganum harmala and/or sandalwood Santalum album, which was believed to protect the king from evil and disease. More than 300 plant species in 5 continents are used in smoke form for different diseases. As a method of drug administration, smoking is important as it is a simple, inexpensive, but very effective method of extracting particles containing active agents. More importantly, generating smoke reduces the particle size to a microscopic scale thereby increasing the absorption of its active chemical principles.
https://en.wikipedia.org/wiki?curid=27001
Tobacco pipe A tobacco pipe, often called simply a pipe, is a device specifically made to smoke tobacco. It comprises a chamber (the bowl) for the tobacco from which a thin hollow stem (shank) emerges, ending in a mouthpiece. Pipes can range from very simple machine-made briar models to highly prized hand-made artisanal implements made by renowned pipemakers, which are often very expensive collector's items. Pipe smoking is the oldest known traditional form of tobacco smoking. Some Native American cultures smoke tobacco in ceremonial pipes, and have done so since long before the arrival of Europeans. For instance the Lakota people use a ceremonial pipe called čhaŋnúŋpa. Other American Indian cultures smoke tobacco socially. The tobacco plant is native to South America but spread into North America long before Europeans arrived. Tobacco was introduced to Europe from the Americas in the 16th century and spread around the world rapidly. As tobacco was not introduced to the Old World until the 16th century, the older pipes outside of the Americas were usually used to smoke various other substances, including hashish, a rare and expensive substance outside areas of the Middle East, Central Asia and India, where it was then produced. A pipe's fundamental function is to provide a relatively safe, manipulable volume in which to incompletely combust a smokable substance. Typically this is accomplished by connecting a refractory 'bowl' to some sort of 'stem' which extends and may also cool the smoke mixture drawn through the combusting organic mass (see below). The broad anatomy of a pipe typically comprises mainly the bowl and the stem. The "bowl" (1) which is the cup-like outer shell, the part hand-held while packing, holding and smoking a pipe, is also the part "knocked" top-down to loosen and release impacted spent tobacco. On being sucked, the general stem delivers the smoke from the bowl to the user's mouth. Inside the bowl is an inner "chamber" (2) space holding tobacco pressed into it. This "draught hole" (3), is for air flow where air has travelled through the tobacco in the chamber, taking the smoke with it, up the "shank" (4). At the end of the shank, the pipe's "mortise" (5) and "tenon" (6) join is an air-tight, simple connection of two detachable parts where the mortise is a hole met by the tenon, a tight-fitting "tongue" at the start of the "stem" (7). Known as the "bore" (10), the inner shaft of this second section stays uniform throughout while the outer stem tapers down to the mouthpiece or "bit" (8) held in the smoker's teeth, and finally ends in the "lip" (9), attenuated for comfort. The bowls of tobacco pipes are commonly made of briar wood, meerschaum, corncob, pear-wood, rose-wood or clay. Less common are other dense-grained woods such as cherry, olive, maple, mesquite, oak, and bog-wood. Minerals such as catlinite and soapstone have also been used. Pipe bowls are sometimes decorated by carving, and moulded clay pipes often had simple decoration in the mould. Unusual, but still noteworthy pipe materials include gourds, as in the famous calabash pipe, and pyrolytic graphite. Metal and glass are uncommon materials for tobacco pipes, but are common for pipes intended for other substances, such as cannabis. The stem needs a long channel of constant position and diameter running through it for a proper draw, although filter pipes have varying diameters and can be successfully smoked even without filters or adapters. Because it is molded rather than carved, clay may make up the entire pipe or just the bowl, but most other materials have stems made separately and detachable. Stems and bits of tobacco pipes are usually made of moldable materials like Ebonite, Lucite, Bakelite, and soft plastic. Less common are stems made of reeds, bamboo, or hollowed out pieces of wood. Expensive pipes once had stems made of amber, though this is rare now. Calabash gourds (usually with meerschaum or porcelain bowls set inside them) have long made prized pipes, but they are labour-intensive and, today, quite expensive. Because of this expense, pipes with bodies made of wood (usually mahogany) instead of gourd, but with the same classic shape, are sold as calabashes. Both wood and gourd pipes are functionally the same (with the important exception that the dried gourd, usually being noticeably lighter, sits more comfortably in the mouth). They consist of a downward curve that ends with an upcurve where the bowl sits. Beneath the bowl is an air chamber which serves to cool, dry, and mellow the smoke. There are also briar pipes being sold as calabashes. These typically do not have an air chamber and are so named only because of their external shape. A calabash pipe is rather large and easy to recognize as a pipe when used on a stage in dramatic productions. Although a British newspaper cartoon of the early 1900s depicts the British actor H. A. Saintsbury as the Great Detective smoking what may be a calabash pipe, its now-stereotypical identification with Sherlock Holmes remains a mystery. Some commentators have erroneously associated the calabash with William Gillette, the first actor to become universally recognized as the embodiment of the detective. Gillette actually introduced the curving or bent pipe for use by Holmes, but his pipe was an ornate briar. Gillette chose a bent pipe, more easily clenched in the teeth when delivering lines. While there are promotional stills of Basil Rathbone smoking calabash pipes as Holmes for other projects, most notably his radio show, in his first two outings as Holmes produced by 20th Century-Fox as taking place in the Victorian era, Rathbone smoked an apple-bowled, black briar with a half bend, made by Dunhill, the company known for making the best pipes at that time. In the next dozen films, the series produced by Universal Studios, with Holmes and Watson updated to the 1940s, Rathbone smokes a much less expensive Peterson half bend with a billiard-shaped bowl. A calabash is introduced in "The Spider Woman" but Holmes does not smoke it. In the original chronicles, such as "The Adventure of the Copper Beeches", Sherlock Holmes is described as smoking a long-stemmed cherrywood (but not a churchwarden pipe) which he favored "when in a disputatious, rather than a meditative mood." Holmes smokes an old briar-root pipe on occasion, "The Sign of the Four" for one, and an "unsavory" and "disreputable" black and oily clay pipe in several stories, notably in "The Red-Headed League". Dr Watson declares it to be the detective's preferred pipe: “It was to him as a counsellor” ("A Case of Identity"); the “companion of his deepest meditations" ("The Valley of Fear").. Bowls are made of varying shapes and materials to allow the smoker to try different characteristics or to dedicate particular bowls for particular tobaccos. Bowls are not interchangeable between manufacturers. A "hookah", "ghelyan", or "narghile", is a Middle Eastern water pipe that cools the smoke by filtering it through a water chamber. Often ice, cough-drops, milk, or fruit juice is added to the water. Traditionally, the tobacco is mixed with a sweetener, such as honey or molasses. Fruit flavors have also become popular. Modern hookah smokers, especially in the US, smoke "me'assel", "moassel", "molasses" or "shisha", all names for the same wet mixture of tobacco, molasses/honey, glycerine, and often, flavoring. This style of tobacco is smoked in a bowl with foil or a screen (metal or glass) on top of the bowl. More traditional tobaccos are "tombiek" (a dry unflavored tobacco, which the user moistens in water, squeezes out the extra liquid, and places coals directly on top) or "jarak" (more of a paste of tobacco with fruit to flavor the smoke). The majority of pipes sold today, whether handmade or machine-made, are fashioned from briar (). Briar is a particularly well suited wood for pipe making for a number of reasons. The first and most important characteristic is its natural resistance to fire. The second is its inherent ability to absorb moisture. The burl absorbs water in nature to supply the tree in the dry times and likewise will absorb the moisture that is a byproduct of combustion. Briar is cut from the root burl of the tree heath ("Erica arborea"), which is native to the rocky and sandy soils of the Mediterranean region. Briar burls are cut into two types of blocks; ebauchon and plateaux. Ebauchon is taken from the heart of the burl while plateaux is taken from the outer part of the burl. While both types of blocks can produce pipes of the highest quality, most artisan pipemakers prefer to use plateaux because of their superior graining. Ceramic pipes, made of moulded and then fired clay, were used almost universally by Europeans between the introduction of tobacco in the 16th century, and the introduction of cheap cigarettes at the end of the nineteenth. The material is not very strong and the early varieties had long thin stems, so they frequently broke, but were cheap to replace. It has been claimed that this fragility was somewhat intentional as it was utilized by Colonial American tavern keepers, for example, in renting the clay pipes to patrons. When the patron was done smoking the pipe and returned it to the keeper, the end of the stem was simply broken off so as to be ready for the next patron. However, there is no documentary evidence for this practice; it is known that communal pipes used in taverns were cleansed by being heated in an oven on special iron racks. Forming the pipe involved making them in moulds with the bore created by pushing an oiled wire inside the stem. The preferred material was pipeclay or "tobacco pipe clay", which fires to a white colour and is found in only certain locations. In North America, many clay pipes were historically made from more typical terracotta-coloured clays. According to one British writer in 1869, the French preferred old pipes and the English new, the middle class preferred long stems and the working class preferred short. Short stemmed pipes, sometimes called "cuttys" or "nose warmers" in England, were preferred by those doing manual work as they could be gripped between the teeth, leaving both of the smoker's hands free. Later low-quality clay pipes were made by slip casting in a mould. Higher quality pipes are made in a labour-intensive hand shaping process. Traditionally, clay pipes are un-glazed. Clays burn "hot" in comparison to other types of pipes, so they are often difficult for most pipe-smokers to use. Their proponents claim that, unlike other materials, a well-made clay pipe gives a "pure" smoke with no flavour addition from the pipe bowl. In addition to aficionados, reproductions of historical clay styles are used by some historical re-enactors. Clay pipes were once very popular in Ireland, where they were called "dudeen"s. Broken fragments of clay pipe can be useful as dating evidence for archaeologists. In the 1950s, the American archaeologist J. C. Harrington noted that the bore of pipe stems decreased over time, so a late sixteenth or early seventeenth centuries pipe would have a stem bore diameter of around , but a late eighteenth century pipe would have a bore diameter of around . The size of bowls also increased over time as tobacco became a cheaper commodity, and later pipes tend to be more decorated. The specifically American style of pipes made from corncobs are cheap and effective, even if some regard them as inelegant. The cobs are first dried for two years. Then they are hollowed out to make a bowl shape. The bowls are dipped in a plaster-based mixture and varnished or lacquered on the outside. Shanks made from birch wood are then inserted into the bowls. The first and largest manufacturer of corncob pipes is Missouri Meerschaum, located in Washington, Missouri, in the United States. Missouri Meerschaum has produced the pipes since 1869. General Douglas MacArthur and Mark Twain were perhaps the most famous smokers of this type of pipe, along with the cartoon characters Popeye and Frosty the Snowman. Corncob pipes remain popular today because they are inexpensive and require no "break-in" period like briar pipes. For these two reasons, corncob pipes are often recommended as a "beginner's pipe." However, corncob pipes are equally valued by both learners and experienced smokers who simply desire a cool, clean smoke. Pipesmokers who wish to sample a wide variety of different tobaccos and blends also might keep a stock of corncobs on hand to permit them to try new flavors without "carryover" from an already-used pipe, or to keep a potentially bad-tasting tobacco from adding its flavor to a more expensive or favored pipe. Meerschaum (hydrated magnesium silicate), a mineral found in small shallow deposits mainly around the city of Eskişehir in central Turkey, is prized for the properties which allow it to be carved into finely detailed decorative and figural shapes. It has been used since the 17th century and, with clay pipes, represented the most common medium for pipes before the introduction of briar as the material of choice in the 19th century. The word "meerschaum" means "sea foam" in German, alluding to its natural white color and its surprisingly low weight. Meerschaum is a very porous mineral that absorbs elements of the tobacco during the smoking process, and gradually changes color to a golden brown. Old, well-smoked meerschaum pipes are valued by collectors for their distinctive coloring. Meerschaum pipes can either be carved from a block of meerschaum, or made from meerschaum dust collected after carving and mixed with a binder then pressed into a pipe shape. The latter are far less absorbent, color in blotches, and lack the smoking quality of the block carved pipe. A variety of other materials may also be used for pipes. The Redmanol corporation manufactured pipes with translucent stems in the 1920s and a series of pipes were manufactured and distributed by the Tar Gard (later Venturi) Corporation of San Francisco from 1965-1975. Marketed under names such as "the pipe," "THE SMOKE" and "Venturi," they used materials such as pyrolytic graphite, phenolic resin, nylon, Bakelite and other synthetics, allowing for higher temperatures in the bowl, reduced tar, and aesthetic variations of color and style. After Venturi stopped making pipes, several companies continue to make pipes from Brylon, a composite of nylon and wood flour, as a cheaper substitute for briar. Used to absorb moisture, tar and nicotine. Made of: Filters can be single- or double-sided. Double-sided filter has both ends ceramic that can withstand hot smoke. Single-sided filter has ceramic end to the bowl and plastic end to the stem. Smoking a pipe requires more apparatus and technique than cigarette or even cigar smoking. In addition to the pipe itself and matches or a pipe lighter, smokers usually require a pipe tool for packing, adjusting, and emptying the tobacco in the bowl, and a regular supply of pipe cleaners. Tobaccos for smoking in pipes are often carefully treated and blended to achieve flavour nuances not available in other tobacco products. Many of these are blends using staple ingredients of variously cured Burley and Virginia tobaccos which are enhanced by spice tobaccos, among them many Oriental or Balkan varietals, Latakia (a fire-cured spice tobacco of Syrian origin, but now made in other regions, such as, Cyprus and Lebanon ), Perique (uniquely grown in St. James Parish, Louisiana) which is also an old method of fermentation, or blends of Virginia and Burley tobaccos of African, Indian, or South American origins. Traditionally, many U.S. blends are made of American Burley with sweeteners and flavorings added to create an "aromatic" flavor, whereas "English" blends are based on natural Virginia tobaccos enhanced with Oriental and other natural tobaccos. There is a growing tendency towards "natural" tobaccos which derive their aromas from artful blending with selected spice tobaccos only and careful, often historically-based, curing processes. Pipe tobacco can be purchased in several forms, which vary both in flavour (leading to many blends and opportunities for smokers to blend their own tobaccos) and in the physical shape and size to which the tobacco has been reduced. Most pipe tobaccos are less mild than cigarette tobacco, substantially more moist and cut much more coarsely. Too finely cut tobacco does not allow enough air to flow through the pipe, and overly dry tobacco burns too quickly with little flavour. Pipe tobacco must be kept in an airtight container, such as a canning jar or sealed tin, to keep from drying out. Some pipe tobaccos are cut into long narrow ribbons. Some are pressed into flat plugs which are sliced into flakes. Others are tightly wound into long ropes, then sliced into discs. Plug tobacco is maintained in its pressed block form and sold in small blocks. The plug will be sliced into thin flakes by the smoker and then prepared in a similar fashion to flake tobacco. It is considered that plug tobacco holds its flavor better than rubbed or flake tobacco. Flake tobacco (sliced cakes or ropes) may be prepared in several ways. Generally it is rubbed out with the fingers and palms until it is loose enough to pack. It can also be crumbled or simply folded and stuffed into a pipe. Some people also prefer to dice up very coarse tobaccos before using them, making them easier to pack. In the most common method of packing, tobacco is added to the bowl of the pipe in several batches, each one pressed down until the mixture has a uniform density that optimizes airflow (something that it is difficult to gauge without practice). This can be done with a finger or thumb, but if the tobacco needs to be repacked later, while it is burning, the tamper on a pipe tool is sometimes used. If it needs to be loosened, the reamer, or any similar long pin can be used. A traditional way of packing the pipe is to fill the bowl and then pack gently to about full, fill again and pack slightly more firmly to about full, and then pack more firmly still to the top. An alternative packing technique called the Frank method involves lightly dropping tobacco in the pipe, after which a large plug is gingerly pushed into the bowl all at once. Matches, or separately lit slivers of wood are often considered preferable to lighters because of lower burning temperature. Butane lighters made specifically for pipes emit flame sideways or at an angle to make it easier to direct flame into the bowl. Torch-style lighters should never be used to light a pipe because their flames are too hot and can char the rim of the pipe bowl. Matches should be allowed to burn for several seconds to allow the sulfur from the tip to burn away and the match to produce a full flame. A naphtha fueled lighter should also be allowed to burn a few seconds to get rid of stray naphtha vapors that could give a foul taste to the smoke. When a flame has been produced, it is then moved in circles above the rim of the bowl while the smoker puffs to draw the flame down and light the tobacco. Packing method and humidity can affect how often a pipe must be relit. With care, a briar pipe can last a very long time without burning out. However, due to aggressive (hot) smoking, imperfections in the wood, a hole can be burned in the tobacco chamber of the pipe. There are several methods used to help prevent a wood pipe from burning out. These generally involve coating the chamber with any of a variety of substances, or by gently smoking a new pipe to build up a cake (a mixture of ash, unburned tobacco, oils, sugars, and other residue) on the walls. These coatings may include honey and water; powdered sugar and water; cigar ash and water; and sour cream, buttermilk, and activated charcoal among many others. Many modern briar pipes are pre-treated by the manufacturer to resist burning. If smoked correctly, the cake will build up properly on its own. Another technique is to alternate a half-bowl and a full-bowl the first several times the pipe is used to build an even cake. Burley is often recommended to help a new pipe build cake. The effectiveness of these methods is by no means universally agreed upon. The caked layer that helps prevent burning through the bottom or sides of a briar wood pipe may damage other pipes, such as meerschaum or clay. As the cake layer heats up, it expands and may cause cracks or breaks in non-briar pipes. Pipe smoke, like cigar smoke, is usually not inhaled. It is merely brought into the mouth, pumped around oral and nasal cavities to permit absorption of nicotine toward the brain through the mucous membranes, and released. It is normal to have to relight a pipe periodically. If it is smoked too slowly, this will happen more often. If it is smoked too quickly, it can produce excess moisture causing a gurgling sound in the pipe and an uncomfortable sensation on the tongue (referred to as "pipe tongue", or more commonly, "tongue bite"). A pipe cleaner can be used to dry out the bowl and, wetted with alcohol, the inner channel. The bowl of the pipe can also become uncomfortably hot, depending on the material and the rate of smoking. For this reason, clay pipes in particular are often held by the stem. Meerschaum pipes are held in a square of chamois leather, with gloves, or else by the stem in order to prevent uneven coloring of the material. The ash and the last bits of unburned tobacco, known as dottle, should be cleaned out with a suitable pipe tool. A soft or bristle pipe cleaner, which may be moistened with strong spirits is then run through the airways of the stem and shank to remove any moisture, ash, and other residue before the pipe is allowed to dry. A pipe should be allowed to cool before removing the stem to avoid the possibility of warping it. A cake of ash eventually develops inside the bowl. This is generally considered desirable for controlling overall heat. However, if it becomes too thick, it may expand faster than the bowl of the pipe itself when heated, cracking the bowl. Before reaching this point, it needs to be scraped down with a reamer. It is generally recommended to keep the cake at approximately the thickness of a U.S. dime (about 1/20th of an inch or 1.5 mm), though sometimes the cake is removed entirely as part of efforts to eliminate flavors or aromas. Cake is considered undesirable in meerschaum pipes because it can easily crack the bowl and/or interfere with the mineral's natural porosity. Meerschaum also softens when heated so it is recommended to allow meerschaum pipes to cool before cleaning as people have been known to push pipe cleaners through the walls of heated pipes. Regardless if a pipe is cleaned after every smoke, over time there is a buildup of cake in the bowl and tars in the internals of a smoking pipe. The cake can be controlled by gentle reaming, but a buildup of tars in the shank and airway of a pipe is more difficult to deal with. This may require the services of a professional pipe restorer to properly clean and sanitize the pipe. When tobacco is burned, oils from adjoining not yet ignited particles vaporize and condense into the existing cake on the walls of the bowl and shank. Over time, these oils can oxidize and turn rancid, causing the pipe to give a sour or bitter smoke. A purported countermeasure involves filling the bowl with kosher salt and carefully wetting it with strong spirits. It is important to not use iodized salt, as the iodine and other additives may impart an unpleasant flavor. Regularly wiping out the bowl with spirits such as vodka or rum is helpful in preventing souring. Commercial pipe-sweetening products are also available.
https://en.wikipedia.org/wiki?curid=27002
Swiss cheese (North America) Swiss cheese is the name for a variety of cheese that resembles Emmental cheese, a yellow, medium-hard cheese that originated in the area around Emmental, Switzerland. Some types of Swiss cheese have a distinctive appearance, as the blocks or rounds of the cheese are riddled with holes known as "eyes". Swiss cheese without eyes is known as "blind". Swiss cheese is now manufactured in many countries, including the United States, Finland, Estonia, and Ireland. It is sometimes made with pasteurized or part-skim milk, unlike the original from Switzerland made with raw milk. The United States Department of Agriculture uses the terms Swiss cheese and Emmentaler cheese interchangeably. In Australia, both terms are used, along with Swiss-style cheese, in some cases differentiating the two. The term Swiss cheese is sometimes used in India, although it is also often referred to as Emmental, which is the more common name in Europe. Three types of bacteria are used in the production of Swiss cheese: "Streptococcus salivarius" subspecies "thermophilus" (also known as "Streptococcus thermophilus"), "Lactobacillus" ("Lactobacillus helveticus" or "Lactobacillus delbrueckii" subspecies "bulgaricus"), and "Propionibacterium" ("Propionibacterium freudenreichii" subspecies "shermani"). In a late stage of cheese production, the propionibacteria consume the lactic acid excreted by the other bacteria and release acetate, propionic acid, and carbon dioxide gas. The carbon dioxide slowly forms the bubbles that develop the "eyes". The acetate and propionic acid give Swiss its nutty and sweet flavor. A hypothesis proposed by Swiss researchers in 2015 notes that particulate matter may also play a role in the holes' development and that modern sanitation eliminated debris such as hay dust in the milk played a role in reduced hole size in Swiss cheeses, or even "blind cheese". Historically, the holes were seen as a sign of imperfection and cheese makers originally tried to avoid them by pressing during production. In modern times, the holes have become an identifier of the cheese. In general, the larger the eyes in a Swiss cheese, the more pronounced its flavor because a longer fermentation period gives the bacteria more time to act. This poses a problem, however, because cheese with large eyes does not slice well and comes apart in mechanical slicers. As a result, U.S. industry regulators have reduced the minimum eye size with which Swiss cheese can receive the Grade A stamp. In 2014, 297.8 million pounds of Swiss cheese was reportedly produced in the United States. Baby Swiss and Lacy Swiss are two varieties of American Swiss cheeses. Both have small holes and a mild flavor. Baby Swiss is made from whole milk, and Lacy Swiss is made from low fat milk. Baby Swiss was developed in the mid-1960s outside of Charm, Ohio, by the Guggisberg Cheese Company, owned by Alfred Guggisberg.
https://en.wikipedia.org/wiki?curid=27003
Serendipity Serendipity is the occurrence of an unplanned fortunate discovery. Serendipity is a common occurrence throughout the history of product invention and scientific discovery. Serendipity is also seen as a potential design principle for online activities that would present a wide array of information and viewpoints, rather than just re-enforcing a user's opinion. The first noted use of "serendipity" in the English language was by Horace Walpole on 28 January 1754. In a letter he wrote to his friend Horace Mann, Walpole explained an unexpected discovery he had made about a lost painting of Bianca Cappello by Giorgio Vasari by reference to a Persian fairy tale, "The Three Princes of Serendip". The princes, he told his correspondent, were "always making discoveries, by accidents and sagacity, of things which they were not in quest of." The name comes from "Serendip", an old name for Sri Lanka (Ceylon), hence "Sarandib" by Arab traders. It is derived from the Sanskrit "Siṃhaladvīpaḥ" (Siṃhalaḥ, Sri Lanka + dvīpaḥ, island). The word has been exported into many other languages, with the general meaning of “unexpected discovery” or “fortunate chance”. The term "serendipity" is often applied to inventions made by chance rather than intent. Andrew Smith, editor of "The" "Oxford Companion to American Food and Drink", has speculated that most everyday products had serendipitous roots, with many early ones related to animals. The origin of cheese, for example, possibly originated in the Nomad practice of storing milk in the stomach of a dead camel that was attached to the saddle of a live one, thereby mixing rennet from the stomach with the milk stored within. Other examples of serendipity in inventions include: Serendipity contributed to entomologist Shaun Winterton discovering "Semachrysa jade", a new species of lacewing, which he found not in its native Malaysia, but on the photo-sharing site Flickr. Winterton's discovery was aided by Flickr's ability to present images that are personalized to a user's interests, thereby increasing the odds he would chance upon the photo. Computer scientist Jaime Teevan has argued that serendipitous discovery is promoted by such personalization, writing that "people don’t know what to do with random new information. Instead, we want information that is at the fringe of what we already know, because that is when we have the cognitive structures to make sense of the new ideas." Serendipity is a design principle for online activity that would present viewpoints that diverge from those participants already hold. Harvard Law professor Cass Sunstein argues that such an "architecture of serendipity" would promote a healthier democracy. Like a great city or university, "a well-functioning information market" provides exposure to new ideas, people, and ways of life, "Serendipity is crucial because it expands your horizons. You need that if you want to be free." The idea has potential application in the design of social media, information searches, and web browsing. William Boyd coined the term zemblanity in the late twentieth century to mean somewhat the opposite of serendipity: "making unhappy, unlucky and expected discoveries occurring by design". A zemblanity is, effectively, an "unpleasant unsurprise". It derives from Novaya Zemlya (or Nova Zembla), a cold, barren land with many features opposite to the lush Sri Lanka (Serendip). Bahramdipity is derived directly from Bahram Gur as characterized in "The Three Princes of Serendip". It describes the "suppression" of serendipitous discoveries or research results by powerful individuals.
https://en.wikipedia.org/wiki?curid=27006
Samuel Morse Samuel Finley Breese Morse (April 27, 1791 – April 2, 1872) was an American inventor and painter. After having established his reputation as a portrait painter, in his middle age Morse contributed to the invention of a single-wire telegraph system based on European telegraphs. He was a co-developer of Morse code and helped to develop the commercial use of telegraphy. Samuel F. B. Morse was born in Charlestown, Massachusetts, the first child of the pastor Jedidiah Morse (1761–1826), who was also a geographer, and his wife Elizabeth Ann Finley Breese (1766–1828). His father was a great preacher of the Calvinist faith and supporter of the American Federalist party. He thought it helped preserve Puritan traditions (strict observance of Sabbath, among other things), and believed in the Federalist support of an alliance with Britain and a strong central government. Morse strongly believed in education within a Federalist framework, alongside the instillation of Calvinist virtues, morals, and prayers for his first son. His first ancestor in America was Samuel Morse, who emigrated to Dedham, Massachusetts in 1635. After attending Phillips Academy in Andover, Massachusetts, Samuel Morse went on to Yale College to receive instruction in the subjects of religious philosophy, mathematics, and science of horses. While at Yale, he attended lectures on electricity from Benjamin Silliman and Jeremiah Day and was a member of the Society of Brothers in Unity. He supported himself by painting. In 1810, he graduated from Yale with Phi Beta Kappa honors. Morse married Lucretia Pickering Walker on September 29, 1818, in Concord, New Hampshire. She died on February 7, 1825, of a heart attack shortly after the birth of their third child. (Susan b. 1819, Charles b. 1823, James b. 1825). He married his second wife, Sarah Elizabeth Griswold on August 10, 1848, in Utica, New York and had four children (Samuel b. 1849, Cornelia b. 1851, William b. 1853, Edward b. 1857). Morse expressed some of his Calvinist beliefs in his painting, "Landing of the Pilgrims", through the depiction of simple clothing as well as the people's austere facial features. His image captured the psychology of the Federalists; Calvinists from England brought to North America ideas of religion and government, thus linking the two countries. This work attracted the attention of the notable artist, Washington Allston. Allston wanted Morse to accompany him to England to meet the artist Benjamin West. Allston arranged—with Morse's father—a three-year stay for painting study in England. The two men set sail aboard the "Libya" on July 15, 1811. In England, Morse perfected his painting techniques under Allston's watchful eye; by the end of 1811, he gained admittance to the Royal Academy. At the Academy, he was moved by the art of the Renaissance and paid close attention to the works of Michelangelo and Raphael. After observing and practicing life drawing and absorbing its anatomical demands, the young artist produced his masterpiece, the "Dying Hercules". (He first made a sculpture as a study for the painting.) To some, the "Dying Hercules" seemed to represent a political statement against the British and also the American Federalists. The muscles symbolized the strength of the young and vibrant United States versus the British and British-American supporters. During Morse's time in Britain, the Americans and British were engaged in the War of 1812. Both societies were conflicted over loyalties. Anti-Federalist Americans aligned themselves with the French, abhorred the British, and believed a strong central government to be inherently dangerous to democracy. As the war raged on, Morse's letters to his parents became more anti-Federalist in tone. In one such letter, Morse wrote: I assert ... that the Federalists in the Northern States have done more injury to their country by their violent opposition measures than a French alliance could. Their proceedings are copied into the English papers, read before Parliament, and circulated through their country, and what do they say of them ... they call them [Federalists] cowards, a base set, say they are traitors to their country and ought to be hanged like traitors. Although Jedidiah Morse did not change Samuel's political views, he continued as an influence. Critics believe that the elder Morse's Calvinist ideas are integral to Morse's "Judgment of Jupiter," another significant work completed in England. Jupiter is shown in a cloud, accompanied by his eagle, with his hand spread above the parties and he is pronouncing judgment. Marpessa, with an expression of compunction and shame, is throwing herself into the arms of her husband. Idas, who tenderly loved Marpessa, is eagerly rushing forward to receive her while Apollo stares with surprise. Critics have suggested that Jupiter represents God's omnipotence—watching every move that is made. Some call the portrait a moral teaching by Morse on infidelity. Although Marpessa fell victim, she realized that her eternal salvation was important and desisted from her wicked ways. Apollo shows no remorse for what he did but stands with a puzzled look. Many American paintings throughout the early nineteenth century had religious themes, and Morse was an early exemplar of this. "Judgment of Jupiter" allowed Morse to express his support of Anti-Federalism while maintaining his strong spiritual convictions. Benjamin West sought to present the "Jupiter" at another Royal Academy exhibition, but Morse's time had run out. He left England on August 21, 1815, to return to the United States and begin his full-time career as a painter. The decade 1815–1825 marked significant growth in Morse's work, as he sought to capture the essence of America's culture and life. He painted the Federalist former President John Adams (1816). The Federalists and Anti-Federalists clashed over Dartmouth College. Morse painted portraits of Francis Brown—the college's president—and Judge Woodward (1817), who was involved in bringing the Dartmouth case before the U.S. Supreme Court. Morse also sought commissions among the elite of Charleston, South Carolina. Morse's 1818 painting of Mrs. Emma Quash symbolized the opulence of Charleston. The young artist was doing well for himself. Between 1819 and 1821, Morse went through great changes in his life, including a decline in commissions due to the Panic of 1819. Morse was commissioned to paint President James Monroe in 1820. He embodied Jeffersonian democracy by favoring the common man over the aristocrat. Morse had moved to New Haven. His commissions for "The House of Representatives" (1821) and a portrait of the Marquis de Lafayette (1825) engaged his sense of democratic nationalism. "The House of Representatives" was designed to capitalize on the success of François Marius Granet's "The Capuchin Chapel in Rome," which toured the United States extensively throughout the 1820s, attracting audiences willing to pay the 25-cent admission fee. The artist chose to paint the House of Representatives, in a similar way, with careful attention to architecture and dramatic lighting. He also wished to select a uniquely American topic that would bring glory to the young nation. His subject did just that, showing American democracy in action. He traveled to Washington D.C. to draw the architecture of the new Capitol and placed eighty individuals within the painting. He chose to portray a night scene, balancing the architecture of the Rotunda with the figures, and using lamplight to highlight the work. Pairs of people, those who stood alone, individuals bent over their desks working, were each painted simply but with faces of character. Morse chose nighttime to convey that Congress' dedication to the principles of democracy transcended day. "The House of Representatives" failed to draw a crowd when exhibited in New York City in 1823. By contrast, John Trumbull's "Declaration of Independence" had won popular acclaim a few years earlier. Viewers may have felt that the architecture of "The House of Representatives" overshadows the individuals, making it hard to appreciate the drama of what was happening. Morse was honored to paint the Marquis de Lafayette, the leading French supporter of the American Revolution. He felt compelled to paint a grand portrait of the man who helped to establish a free and independent America. He features Lafayette against a magnificent sunset. He has positioned Lafayette to the right of three pedestals: one has a bust of Benjamin Franklin, another of George Washington, and the third seems reserved for Lafayette. A peaceful woodland landscape below him symbolized American tranquility and prosperity as it approached the age of fifty. The developing friendship between Morse and Lafayette and their discussions of the Revolutionary War affected the artist after his return to New York City. In 1826, he helped found the National Academy of Design in New York City. He served as the Academy's president from 1826 to 1845 and again from 1861 to 1862. From 1830 to 1832, Morse traveled and studied in Europe to improve his painting skills, visiting Italy, Switzerland, and France. During his time in Paris, he developed a friendship with the writer James Fennimore Cooper. As a project, he painted miniature copies of 38 of the Louvre's famous paintings on a single canvas (6 ft. x 9 ft), which he entitled "The Gallery of the Louvre." He completed the work upon his return to the United States. On a subsequent visit to Paris in 1839, Morse met Louis Daguerre. He became interested in the latter's daguerreotype—the first practical means of photography. Morse wrote a letter to the "New York Observer" describing the invention, which was published widely in the American press and provided broad awareness of the new technology. Mathew Brady, one of the earliest photographers in American history, famous for his depictions of the Civil War, initially studied under Morse and later took photographs of him. Some of Morse's paintings and sculptures are on display at his Locust Grove estate in Poughkeepsie, New York. While returning by ship from Europe in 1832, Morse encountered Charles Thomas Jackson of Boston, a man who was well schooled in electromagnetism. Witnessing various experiments with Jackson's electromagnet, Morse developed the concept of a single-wire telegraph. He set aside his painting, "The Gallery of the Louvre". The original Morse telegraph, submitted with his patent application, is part of the collections of the National Museum of American History at the Smithsonian Institution. In time the Morse code, which he developed, would become the primary language of telegraphy in the world. It is still the standard for rhythmic transmission of data. Meanwhile, William Cooke and Professor Charles Wheatstone had learned of the Wilhelm Weber and Carl Gauss electromagnetic telegraph in 1833. They had reached the stage of launching a commercial telegraph prior to Morse, despite starting later. In England, Cooke became fascinated by electrical telegraphy in 1836, four years after Morse. Aided by his greater financial resources, Cooke abandoned his primary subject of anatomy and built a small electrical telegraph within three weeks. Wheatstone also was experimenting with telegraphy and (most importantly) understood that a single large battery would not carry a telegraphic signal over long distances. He theorized that numerous small batteries were far more successful and efficient in this task. (Wheatstone was building on the primary research of Joseph Henry, an American physicist.) Cooke and Wheatstone formed a partnership and patented the electrical telegraph in May 1837, and within a short time had provided the Great Western Railway with a stretch of telegraph. However, within a few years, Cooke and Wheatstone's multiple-wire signaling method would be overtaken by Morse's cheaper method. In an 1848 letter to a friend, Morse describes how vigorously he fought to be called the sole inventor of the electromagnetic telegraph despite the previous inventions. Morse encountered the problem of getting a telegraphic signal to carry over more than a few hundred yards of wire. His breakthrough came from the insights of Professor Leonard Gale, who taught chemistry at New York University (he was a personal friend of Joseph Henry). With Gale's help, Morse introduced extra circuits or relays at frequent intervals and was soon able to send a message through of wire. This was the great breakthrough he had been seeking. Morse and Gale were soon joined by Alfred Vail, an enthusiastic young man with excellent skills, insights, and money. At the Speedwell Ironworks in Morristown, New Jersey on January 11, 1838, Morse and Vail made the first public demonstration of the electric telegraph. Although Morse and Alfred Vail had done most of the research and development in the ironworks facilities, they chose a nearby factory house as the demonstration site. Without the repeater, the range of the telegraph was limited to , and the inventors had pulled of wires inside the factory house through an elaborate scheme. The first public transmission, with the message, "A patient waiter is no loser", was witnessed by a mostly local crowd. Morse traveled to Washington, D.C. in 1838 seeking federal sponsorship for a telegraph line but was not successful. He went to Europe, seeking both sponsorship and patents, but in London discovered that Cooke and Wheatstone had already established priority. After his return to the US, Morse finally gained financial backing by Maine congressman Francis Ormand Jonathan Smith. This funding may be the first instance of government support to a private researcher, especially funding for applied (as opposed to basic or theoretical) research. Morse made his last trip to Washington, D.C., in December 1842, stringing "wires between two committee rooms in the Capitol, and sent messages back and forth" to demonstrate his telegraph system. Congress appropriated $30,000 in 1843 for construction of an experimental telegraph line between Washington, D.C., and Baltimore along the right-of-way of the Baltimore and Ohio Railroad. An impressive demonstration occurred on May 1, 1844, when news of the Whig Party's nomination of Henry Clay for U.S. president was telegraphed from the party's convention in Baltimore to the Capitol Building in Washington. On May 24, 1844, the line was officially opened as Morse sent the now-famous words, "What hath God wrought," from the Supreme Court chamber in the basement of the U.S. Capitol building in Washington, D.C., to the B&O's Mount Clare Station in Baltimore. Annie Ellsworth chose these words from the Bible (Numbers 23:23); her father, U.S. Patent Commissioner Henry Leavitt Ellsworth, had championed Morse's invention and secured early funding for it. His telegraph could transmit thirty characters per minute. In May 1845, the Magnetic Telegraph Company was formed in order to build telegraph lines from New York City toward Philadelphia; Boston; Buffalo, New York; and the Mississippi. Telegraphic lines rapidly spread throughout the United States in the next few years, with 12,000 miles of wire laid by 1850. Morse at one time adopted Wheatstone and Carl August von Steinheil's idea of broadcasting an electrical telegraph signal through a body of water or down steel railroad tracks or anything conductive. He went to great lengths to win a lawsuit for the right to be called "inventor of the telegraph" and promoted himself as being an inventor. But Alfred Vail also played an important role in the development of the Morse code, which was based on earlier codes for the electromagnetic telegraph. Morse received a patent for the telegraph in 1847, at the old Beylerbeyi Palace (the present Beylerbeyi Palace was built in 1861–1865 on the same location) in Istanbul, which was issued by Sultan Abdülmecid, who personally tested the new invention. He was elected an Associate Fellow of the American Academy of Arts and Sciences in 1849. The original patent went to the Breese side of the family after the death of Samuel Morse. In 1856, Morse went to Copenhagen and visited the Thorvaldsens Museum, where the sculptor's grave is in the inner courtyard. He was received by King Frederick VII, who decorated him with the Order of the Dannebrog for the telegraph. Morse expressed his wish to donate his Thorvaldsen portrait from 1831 in Rome to the king. The Thorvaldsen portrait today belongs to Margrethe II of Denmark. The Morse telegraphic apparatus was officially adopted as the standard for European telegraphy in 1851. Only the United Kingdom (with its extensive overseas empire) kept the needle telegraph of Cooke and Wheatstone. In 1858, Morse introduced wired communication to Latin America when he established a telegraph system in Puerto Rico, then a Spanish Colony. Morse's oldest daughter, Susan Walker Morse (1819–1885), would often visit her uncle Charles Pickering Walker, who owned the Hacienda Concordia in the town of Guayama. During one of her visits, she met Edward Lind, a Danish merchant who worked in his brother-in-law's Hacienda La Henriqueta in the town of Arroyo. They later married. Lind purchased the Hacienda from his sister when she became a widow. Morse, who often spent his winters at the Hacienda with his daughter and son-in-law, set a two-mile telegraph line connecting his son-in-law's Hacienda to their house in Arroyo. The line was inaugurated on March 1, 1859, in a ceremony flanked by the Spanish and American flags. The first words transmitted by Samuel Morse that day in Puerto Rico were: Puerto Rico, beautiful jewel! When you are linked with the other jewels of the Antilles in the necklace of the world's telegraph, yours will not shine less brilliantly in the crown of your Queen! There is an argument amongst historians that Morse may have received the idea of a plausible telegraph from Harrison Gray Dyar some eighteen years earlier than his patent. Morse was a leader in the anti-Catholic and anti-immigration movement of the mid-19th century. In 1836, he ran unsuccessfully for mayor of New York under the anti-immigrant Nativist Party's banner, receiving only 1,496 votes. When Morse visited Rome, he allegedly refused to take his hat off in the presence of the Pope. Morse worked to unite Protestants against Catholic institutions (including schools), wanted to forbid Catholics from holding public office, and promoted changing immigration laws to limit immigration from Catholic countries. On this topic, he wrote, "We must first stop the leak in the ship through which muddy waters from without threaten to sink us." He wrote numerous letters to the New York "Observer" (his brother Sidney was the editor at the time) urging people to fight the perceived Catholic menace. These were widely reprinted in other newspapers. Among other claims, he believed that the Austrian government and Catholic aid organizations were subsidizing Catholic immigration to the United States in order to gain control of the country. In his "Foreign Conspiracy Against the Liberties of the United States", Morse wrote: Surely American Protestants, freemen, have discernment enough to discover beneath them the cloven foot of this subtle foreign heresy. They will see that Popery is now, what it has ever been, a system of the darkest political intrigue and despotism, cloaking itself to avoid attack under the sacred name of religion. They will be deeply impressed with the truth, that Popery is a political as well as a religious system; that in this respect it differs totally from all other sects, from all other forms of religion in the country. In the same book, published in 1835 under the name of "Brutus", in speaking of "the foreign Emissaries of Popery re-warded in their own country," said : "Where is Bishop Kelly of Richmond, Va.? He also sojourns with us until his duties to foreign masters are performed, and then is rewarded by promotion." (Patrick Kelly was a native of Ireland, and the first bishop of Richmond, Virginia. When after a couple of years, differences regarding questions of jurisdiction arose between him and Ambrose Maréchal, Archbishop of Baltimore, Kelly was offered the recently vacant See of Waterford and Lismore in his homeland.) In the 1850s, Morse became well known as a defender of slavery, considering it to be sanctioned by God. This was a position held by many Southerners and others. In his treatise "An Argument on the Ethical Position of Slavery," he wrote: In the United States, Morse held his telegraph patent for many years, but it was both ignored and contested. In 1853, "The Telegraph Patent case – O'Reilly v. Morse" came before the U.S. Supreme Court where, after very lengthy investigation, Chief Justice Roger B. Taney ruled that Morse had been the first to combine the battery, electromagnetism, the electromagnet, and the correct battery configuration into a workable practical telegraph. However, in spite of this clear ruling, Morse still received no official recognition from the United States government. The Supreme Court did not accept all of Morse's claims. The "O'Reilly v. Morse" case has become widely known among patent lawyers because the Supreme Court explicitly denied Morse's claim 8 for any and all use of the electromagnetic force for purposes of transmitting intelligible signals to any distance. The Supreme Court sustained, however, Morse's claim to such telecommunication when effectuated by means of Morse's inventive "repeater" apparatus. This was an electrical circuit in which a cascade of many sets comprising a relay and a battery were connected in series, so that when each relay closed, it closed a circuit to cause the next battery to power the succeeding relay, as suggested in the accompanying figure. This caused Morse's signal to pass along the cascade without degrading into noise as its amplitude decreased with the distance traveled. (Each time the amplitude of the signal approaches the noise level, the repeater [in effect, a nonlinear amplifier] boosts the signal amplitude well above the noise level.) This use of "repeaters" permitted a message to be sent to great distances, which was previously not feasible. The Supreme Court thus held that Morse could properly claim a patent monopoly on the system or process of transmitting signals at any distance by means of the repeater circuitry indicated above, but he could not properly claim a monopoly over any and all uses of electromagnetic force to transmit signals. The apparatus limitation in the former type of claim limited the patent monopoly to what Morse taught and gave the world. The lack of that limitation in the latter type of claim (i.e., claim 8) both gave Morse more than was commensurate with what he had contributed to society and discouraged the inventive efforts of others who might come up with different and/or better ways to send signals at a distance using the electromagnetic force. The problem that Morse faced (the deterioration of the signal with distance) and how he solved it is discussed in more detail in the article "O'Reilly v. Morse". In summary, the solution, as the Supreme Court stated, was the repeater apparatus described in the preceding paragraphs. The importance of this legal precedent in patent law cannot be overstated, as it became the foundation of the law governing the eligibility of computer program-implemented inventions (as well as inventions implementing natural laws) to be granted patents. Assisted by the American ambassador in Paris, the governments of Europe were approached about their long neglect of Morse while their countries were using his invention. There was a widespread recognition that something must be done, and in 1858 Morse was awarded the sum of 400,000 French francs (equivalent to about $80,000 at the time) by the governments of France, Austria, Belgium, the Netherlands, Piedmont, Russia, Sweden, Tuscany, and Turkey, each of which contributed a share according to the number of Morse instruments in use in each country. In 1858, he was also elected a foreign member of the Royal Swedish Academy of Sciences. Morse lent his support to Cyrus West Field's ambitious plan to construct the first transoceanic telegraph line. Morse had experimented with underwater telegraph circuits since 1842. He invested $10,000 in Field's Atlantic Telegraph Company, took a seat on its board of directors, and was appointed honorary "Electrician". In 1856, Morse traveled to London to help Charles Tilston Bright and Edward Whitehouse test a 2,000-mile-length of spooled cable. After the first two cable-laying attempts failed, Field reorganized the project, removing Morse from direct involvement. Though the cable broke three times during the third attempt, it was successfully repaired, and the first transatlantic telegraph messages were sent in 1858. The cable failed after just three months of use. Though Field had to wait out the Civil War, the cable laid in 1866 proved more durable, and the era of reliable transatlantic telegraph service had begun. In addition to the telegraph, Morse invented a marble-cutting machine that could carve three-dimensional sculptures in marble or stone. He could not patent it, however, because of an existing 1820 Thomas Blanchard design. Samuel Morse gave large sums to charity. He also became interested in the relationship of science and religion and provided the funds to establish a lectureship on "the relation of the Bible to the Sciences". Though he was rarely awarded any royalties for the later uses and implementations of his inventions, he was able to live comfortably. Morsemere in Ridgefield, New Jersey takes its name from Morse, who had bought property there to build a home, but died before its completion. He died in New York City on April 2, 1872, and was interred at Green-Wood Cemetery in Brooklyn, New York. By the time of his death, his estate was valued at some $500,000 ($ today). Morse was elected a member of the American Antiquarian Society in 1815. Despite honors and financial awards received from foreign countries, there was no such recognition in the U.S. until he neared the end of his life when on June 10, 1871, a bronze statue of Samuel Morse was unveiled in Central Park, New York City. An engraved portrait of Morse appeared on the reverse side of the United States two-dollar bill silver certificate series of 1896. He was depicted along with Robert Fulton. An example can be seen on the website of the Federal Reserve Bank of San Francisco's website in their "American Currency Exhibit": A blue plaque was erected to commemorate him at 141 Cleveland Street, London, where he lived from 1812 to 1815. According to his "The New York Times" obituary published on April 3, 1872, Morse received respectively the decoration of the Atiq Nishan-i-Iftikhar (English: Order of Glory) [first medal on wearer's right depicted in photo of Morse with medals], set in diamonds, from Sultan Abdülmecid of Turkey (c.1847), a "golden snuff box containing the Prussian gold medal for scientific merit" from the King of Prussia (1851); the "Great Gold Medal of Arts and Sciences" from the King of Württemberg (1852); and the "Great Golden Medal of Science and Arts" from Emperor of Austria (1855); a cross of Chevalier in the Légion d'honneur from the Emperor of France; the "Cross of a Knight" of the Order of the Dannebrog from the King of Denmark (1856); the Cross of Knight Commander of the Order of Isabella the Catholic, from the Queen of Spain, besides being elected member of innumerable scientific and art societies in this [United States] and other countries. Other awards include Order of the Tower and Sword from the kingdom of Portugal (1860), and Italy conferred on him the insignia of chevalier of the Order of Saints Maurice and Lazarus in 1864. Morse's telegraph was recognized as an IEEE Milestone in 1988. In 1975, Morse was inducted into the National Inventors Hall of Fame. On April 1, 2012, Google announced the release of "Gmail Tap", an April Fools' Day joke that allowed users to use Morse Code to send text from their mobile phones. Morse's great-great-grandnephew Reed Morse—a Google engineer—was instrumental in the prank, which became a real product.
https://en.wikipedia.org/wiki?curid=27007
Ship A ship is a large watercraft that travels the world's oceans and other sufficiently deep waterways, carrying goods or passengers, or in support of specialized missions, such as defense, research and fishing. Ships are generally distinguished from boats, based on size, shape, load capacity, and tradition. In the Age of Sail a "ship" was a sailing vessel defined by its sailplan of at least three square riged masts and a full bowsprit. Ships have been important contributors to human migration and commerce. They have supported the spread of colonization and the slave trade, but have also served scientific, cultural, and humanitarian needs. After the 15th century, new crops that had come from and to the Americas via the European seafarers significantly contributed to the world population growth. Ship transport is responsible for the largest portion of world commerce. As of 2016, there were more than 49,000 merchant ships, totaling almost 1.8 billion dead weight tons. Of these 28% were oil tankers, 43% were bulk carriers, and 13% were container ships. Ships are generally larger than boats, but there is no universally accepted distinction between the two. Ships generally can remain at sea for longer periods of time than boats. A legal definition of ship from Indian case law is a vessel that carries goods by sea. A common notion is that a ship can carry a boat, but not "vice versa". A US Navy rule of thumb is that ships heel towards the "outside" of a sharp turn, whereas boats heel towards the "inside" because of the relative location of the center of mass versus the center of buoyancy. American and British 19th century maritime law distinguished "vessels" from other craft; ships and boats fall in one legal category, whereas open boats and rafts are not considered vessels. In the Age of Sail, a full-rigged ship was a sailing vessel with at least three square-rigged masts and a full bowsprit; other types of vessel were also defined by their sailplan, e.g. barque, brigantine, etc. A number of large vessels are usually referred to as boats. Submarines are a prime example. Other types of large vessel which are traditionally called boats are Great Lakes freighters, riverboats, and ferryboats. Though large enough to carry their own boats and heavy cargoes, these vessels are designed for operation on inland or protected coastal waters. In most maritime traditions ships have individual names, and modern ships may belong to a ship class often named after its first ship. In the northern parts of Europe and America a ship is traditionally referred to with a female grammatical gender, represented in English with the pronoun "she", even if named after a man. This is not universal usage and some English language journalistic style guides advise using "it" as referring to ships with female pronouns can be seen as offensive and outdated. In many documents the ship name is introduced with a ship prefix being an abbreviation of the ship class, for example "MS" (motor ship) or "SV" (sailing vessel), making it easier to distinguish a ship name from other individual names in a text. The first sea-going sailing ships were developed by the Austronesian peoples from what is now Southern China and Taiwan. Their invention of catamarans, outriggers, and crab claw sails enabled their ships to sail for vast distances in open ocean. It led to the Austronesian Expansion at around 3000 to 1500 BC. From Taiwan, they rapidly colonized the islands of Maritime Southeast Asia, then sailed further onwards to Micronesia, Island Melanesia, Polynesia, and Madagascar, eventually colonizing a territory spanning half the globe. Austronesian rigs were distinctive in that they had spars supporting both the upper and lower edges of the sails (and sometimes in between), in contrast to western rigs which only had a spar on the upper edge. The sails were also made from woven leaves, usually from pandan plants. These were complemented by paddlers, who usually positioned themselves on platforms on the outriggers in the larger boats. Austronesian ships ranged in complexity from simple dugout canoes with outriggers or lashed together to large edge-pegged plank-built boats built around a keel made from a dugout canoe. Their designs were unique, evolving from ancient rafts to the characteristic double-hulled, single-outrigger, and double-outrigger designs of Austronesian ships. Early Austronesian sailors influenced the development of sailing technologies in Sri Lanka and Southern India through the Austronesian maritime trade network of the Indian Ocean, the precursor to the spice trade route and the maritime silk road, which was established at around 1500 BC. Some scholars believe that the triangular Austronesian crab claw sail may have influenced the development of the lateen sail in western ships due to early contact. The junk rigs of Chinese ships is also believed to be originally Javanese in origin. In the 1st century AD, the people from Nusantara archipelago already made large ships over 50 m long and stood out 4–7 m out of the water. They could carry 700-1000 people and 260 ton cargo. These ships known as "kunlun bo" or "k'unlun po" (崑崙舶, lit. "ship of the Kunlun people") by the Chinese and "kolandiaphonta" by the Greeks. It has 4-7 masts and able to sail against the wind due to the usage of tanja sails. These ships reaching as far as Ghana. In China, miniature models of ships that feature steering oars have been dated to the Warring States period (c. 475–221 BC). By the Han dynasty, a well kept naval fleet was an integral part of the military. Sternpost-mounted rudders started to appear on Chinese ship models starting in the 1st century AD. However, these early Chinese ships were fluvial (riverine), and were not seaworthy. The Chinese only acquired sea-going ship technologies in the 10th century AD Song Dynasty after contact with Southeast Asian djong trading ships, leading to the development of the junks. In 3000 BC, Ancient Egyptians learned how to assemble wooden planks into a hull. They used woven straps to lash the planks together, and reeds or grass stuffed between the planks helped to seal the seams. The Greek historian and geographer Agatharchides had documented ship-faring among the early Egyptians: ""During the prosperous period of the Old Kingdom, between the 30th and 25th centuries BC, the river-routes were kept in order, and Egyptian ships sailed the Red Sea as far as the myrrh-country."" Sneferu's ancient cedar wood ship Praise of the Two Lands is the first reference recorded (2613 BC) to a ship being referred to by name. The ancient Egyptians were perfectly at ease building sailboats. A remarkable example of their shipbuilding skills was the Khufu ship, a vessel in length entombed at the foot of the Great Pyramid of Giza around 2500 BC and found intact in 1954. The oldest discovered sea faring hulled boat is the Late Bronze Age Uluburun shipwreck off the coast of Turkey, dating back to 1300 BC. By 1200 B.C., the Phoenicians were building large merchant ships. In world maritime history, declares Richard Woodman, they are recognized as “the first true seafarers, founding the art of pilotage, cabotage, and navigation” and the architects of “the first true ship, built of planks, capable of carrying a deadweight cargo and being sailed and steered.” At this time, ships were developing in Asia in much the same way as Europe. Japan used defensive naval techniques in the Mongol invasions of Japan in 1281. It is likely that the Mongols of the time took advantage of both European and Asian shipbuilding techniques. During the 15th century, China's Ming dynasty assembled one of the largest and most powerful naval fleets in the world for the diplomatic and power projection voyages of Zheng He. Elsewhere in Japan in the 15th century, one of the world's first iron-clads, "Tekkōsen" (), literally meaning "iron ships", was also developed. In Japan, during the Sengoku era from the fifteenth to 17th century, the great struggle for feudal supremacy was fought, in part, by coastal fleets of several hundred boats, including the atakebune. In Korea, in the early 15th century during the Joseon era, "Geobukseon"(거북선), was developed. The "turtle ship", as it was called is recognized as the first armored ship in the world. Until the Renaissance, navigational technology remained comparatively primitive compared to Austronesian cultures. This absence of technology did not prevent some civilizations from becoming sea powers. Examples include the maritime republics of Genoa and Venice, Hanseatic League, and the Byzantine navy. The Vikings used their knarrs to explore North America, trade in the Baltic Sea and plunder many of the coastal regions of Western Europe. Towards the end of the 14th century, ships like the carrack began to develop towers on the bow and stern. These towers decreased the vessel's stability, and in the 15th century, the caravel, designed by the Portuguese, based on the Arabic "qarib" which could sail closer to the wind, became more widely used. The towers were gradually replaced by the forecastle and sterncastle, as in the carrack "Santa María" of Christopher Columbus. This increased freeboard allowed another innovation: the freeing port, and the artillery associated with it. The carrack and then the caravel were developed in Portugal. After Columbus, European exploration rapidly accelerated, and many new trade routes were established. In 1498, by reaching India, Vasco da Gama proved that the access to the Indian Ocean from the Atlantic was possible. These explorations in the Atlantic and Indian Oceans were soon followed by France, England and the Netherlands, who explored the Portuguese and Spanish trade routes into the Pacific Ocean, reaching Australia in 1606 and New Zealand in 1642. Parallel to the development of warships, ships in service of marine fishery and trade also developed in the period between antiquity and the Renaissance. Maritime trade was driven by the development of shipping companies with significant financial resources. Canal barges, towed by draft animals on an adjacent towpath, contended with the railway up to and past the early days of the industrial revolution. Flat-bottomed and flexible scow boats also became widely used for transporting small cargoes. Mercantile trade went hand-in-hand with exploration, self-financed by the commercial benefits of exploration. During the first half of the 18th century, the French Navy began to develop a new type of vessel known as a ship of the line, featuring seventy-four guns. This type of ship became the backbone of all European fighting fleets. These ships were long and their construction required 2,800 oak trees and of rope; they carried a crew of about 800 sailors and soldiers. During the 19th century the Royal Navy enforced a ban on the slave trade, acted to suppress piracy, and continued to map the world. A clipper was a very fast sailing ship of the 19th century. The clipper routes fell into commercial disuse with the introduction of steam ships with better fuel efficiency, and the opening of the Suez and Panama Canals. Ship designs stayed fairly unchanged until the late 19th century. The industrial revolution, new mechanical methods of propulsion, and the ability to construct ships from metal triggered an explosion in ship design. Factors including the quest for more efficient ships, the end of long running and wasteful maritime conflicts, and the increased financial capacity of industrial powers created an avalanche of more specialized boats and ships. Ships built for entirely new functions, such as firefighting, rescue, and research, also began to appear. In 2019, the world's fleet included 51,684 commercial vessels with gross tonnage of more than 1,000 tons, totaling 1.96 billion tons. Such ships carried 11 billion tons of cargo in 2018, a sum that grew by 2.7% over the previous year. In terms of tonnage, 29% of ships were tankers, 43% are bulk carriers, 13% container ships and 15% were other types. In 2002, there were 1,240 warships operating in the world, not counting small vessels such as patrol boats. The United States accounted for 3 million tons worth of these vessels, Russia 1.35 million tons, the United Kingdom 504,660 tons and China 402,830 tons. The 20th century saw many naval engagements during the two world wars, the Cold War, and the rise to power of naval forces of the two blocs. The world's major powers have recently used their naval power in cases such as the United Kingdom in the Falkland Islands and the United States in Iraq. The size of the world's fishing fleet is more difficult to estimate. The largest of these are counted as commercial vessels, but the smallest are legion. Fishing vessels can be found in most seaside villages in the world. As of 2004, the United Nations Food and Agriculture Organization estimated 4 million fishing vessels were operating worldwide. The same study estimated that the world's 29 million fishermen caught of fish and shellfish that year. Because ships are constructed using the principles of naval architecture that require same structural components, their classification is based on their function such as that suggested by Paulet and Presles, which requires modification of the components. The categories accepted in general by naval architects are: Some of these are discussed in the following sections. Freshwater shipping may occur on lakes, rivers and canals. Ships designed for those venues may be specially adapted to the widths and depths of specific waterways. Examples of freshwater waterways that are navigable in part by large vessels include the Danube, Mississippi, Rhine, Yangtze and Amazon Rivers, and the Great Lakes. Lake freighters, also called lakers, are cargo vessels that ply the Great Lakes. The most well-known is , the latest major vessel to be wrecked on the Lakes. These vessels are traditionally called boats, not ships. Visiting ocean-going vessels are called "salties." Because of their additional beam, very large salties are never seen inland of the Saint Lawrence Seaway. Because the smallest of the Soo Locks is larger than any Seaway lock, salties that can pass through the Seaway may travel anywhere in the Great Lakes. Because of their deeper draft, salties may accept partial loads on the Great Lakes, "topping off" when they have exited the Seaway. Similarly, the largest lakers are confined to the Upper Lakes (Superior, Michigan, Huron, Erie) because they are too large to use the Seaway locks, beginning at the Welland Canal that bypasses the Niagara River. Since the freshwater lakes are less corrosive to ships than the salt water of the oceans, lakers tend to last much longer than ocean freighters. Lakers older than 50 years are not unusual, and as of 2005, all were over 20 years of age. , built in 1906 as "William P Snyder", was the oldest laker still working on the Lakes until its conversion into a barge starting in 2013. Similarly, "E.M. Ford", built in 1898 as "Presque Isle", was sailing the lakes 98 years later in 1996. As of 2007 "E.M. Ford" was still afloat as a stationary transfer vessel at a riverside cement silo in Saginaw, Michigan. Merchant ships are ships used for commercial purposes and can be divided into four broad categories: fishing, cargo ships, passenger ships, and special-purpose ships. The UNCTAD review of maritime transport categorizes ships as: oil tankers, bulk (and combination) carriers, general cargo ships, container ships, and "other ships", which includes "liquefied petroleum gas carriers, liquefied natural gas carriers, parcel (chemical) tankers, specialized tankers, reefers, offshore supply, tugs, dredgers, cruise, ferries, other non-cargo". General cargo ships include "multi-purpose and project vessels and roll-on/roll-off cargo". Modern commercial vessels are typically powered by a single propeller driven by a diesel or, less usually, gas turbine engine., but until the mid-19th century they were predominantly square sail rigged. The fastest vessels may use pump-jet engines. Most commercial vessels have full hull-forms to maximize cargo capacity. Hulls are usually made of steel, although aluminum can be used on faster craft, and fiberglass on the smallest service vessels. Commercial vessels generally have a crew headed by a sea captain, with deck officers and engine officers on larger vessels. Special-purpose vessels often have specialized crew if necessary, for example scientists aboard research vessels. Fishing boats are generally small, often little more than but up to for a large tuna or whaling ship. Aboard a fish processing vessel, the catch can be made ready for market and sold more quickly once the ship makes port. Special purpose vessels have special gear. For example, trawlers have winches and arms, stern-trawlers have a rear ramp, and tuna seiners have skiffs. In 2004, of fish were caught in the marine capture fishery. Anchoveta represented the largest single catch at . That year, the top ten marine capture species also included Alaska pollock, Blue whiting, Skipjack tuna, Atlantic herring, Chub mackerel, Japanese anchovy, Chilean jack mackerel, Largehead hairtail, and Yellowfin tuna. Other species including salmon, shrimp, lobster, clams, squid and crab, are also commercially fished. Modern commercial fishermen use many methods. One is fishing by nets, such as purse seine, beach seine, lift nets, gillnets, or entangling nets. Another is trawling, including bottom trawl. Hooks and lines are used in methods like long-line fishing and hand-line fishing. Another method is the use of fishing trap. Cargo ships transport dry and liquid cargo. Dry cargo can be transported in bulk by bulk carriers, packed directly onto a general cargo ship in break-bulk, packed in intermodal containers as aboard a container ship, or driven aboard as in roll-on roll-off ships. Liquid cargo is generally carried in bulk aboard tankers, such as oil tankers which may include both crude and finished products of oil, chemical tankers which may also carry vegetable oils other than chemicals and gas carriers, although smaller shipments may be carried on container ships in tank containers. Passenger ships range in size from small river ferries to very large cruise ships. This type of vessel includes ferries, which move passengers and vehicles on short trips; ocean liners, which carry passengers from one place to another; and cruise ships, which carry passengers on voyages undertaken for pleasure, visiting several places and with leisure activities on board, often returning them to the port of embarkation. Riverboats and inland ferries are specially designed to carry passengers, cargo, or both in the challenging river environment. Rivers present special hazards to vessels. They usually have varying water flows that alternately lead to high speed water flows or protruding rock hazards. Changing siltation patterns may cause the sudden appearance of shoal waters, and often floating or sunken logs and trees (called snags) can endanger the hulls and propulsion of riverboats. Riverboats are generally of shallow draft, being broad of beam and rather square in plan, with a low freeboard and high topsides. Riverboats can survive with this type of configuration as they do not have to withstand the high winds or large waves that are seen on large lakes, seas, or oceans. Fishing vessels are a subset of commercial vessels, but generally small in size and often subject to different regulations and classification. They can be categorized by several criteria: architecture, the type of fish they catch, the fishing method used, geographical origin, and technical features such as rigging. As of 2004, the world's fishing fleet consisted of some 4 million vessels. Of these, 1.3 million were decked vessels with enclosed areas and the rest were open vessels. Most decked vessels were mechanized, but two-thirds of the open vessels were traditional craft propelled by sails and oars. More than 60% of all existing large fishing vessels were built in Japan, Peru, the Russian Federation, Spain or the United States of America. A weather ship was a ship stationed in the ocean as a platform for surface and upper air meteorological observations for use in marine weather forecasting. Surface weather observations were taken hourly, and four radiosonde releases occurred daily. It was also meant to aid in search and rescue operations and to support transatlantic flights. Proposed as early as 1927 by the aviation community, the establishment of weather ships proved to be so useful during World War II that the International Civil Aviation Organization (ICAO) established a global network of weather ships in 1948, with 13 to be supplied by the United States. This number was eventually negotiated down to nine. The weather ship crews were normally at sea for three weeks at a time, returning to port for 10-day stretches. Weather ship observations proved to be helpful in wind and wave studies, as they did not avoid weather systems like other ships tended to for safety reasons. They were also helpful in monitoring storms at sea, such as tropical cyclones. The removal of a weather ship became a negative factor in forecasts leading up to the Great Storm of 1987. Beginning in the 1970s, their role became largely superseded by weather buoys due to the ships' significant cost. The agreement of the use of weather ships by the international community ended in 1990. The last weather ship was "Polarfront", known as weather station M ("Mike"), which was put out of operation on 1 January 2010. Weather observations from ships continue from a fleet of voluntary merchant vessels in routine commercial operation. Naval vessels are those used by a navy for military purposes. There have been many types of naval vessel. Modern naval vessels can be broken down into three categories: surface warships, submarines, and auxiliary ships. Modern warships are generally divided into seven main categories: aircraft carriers, cruisers, destroyers, frigates, corvettes, submarines and amphibious assault ships. The distinction between cruisers, destroyers, frigates, and corvettes is not rigorous; the same vessel may be described differently in different navies. Battleships were used during the Second World War and occasionally since then (the last battleships were removed from the U.S. Naval Vessel Register in March 2006), but were made obsolete by the use of carrier-borne aircraft and guided missiles. Most military submarines are either attack submarines or ballistic missile submarines. Until the end of World War II the primary role of the diesel/electric submarine was anti-ship warfare, inserting and removing covert agents and military forces, and intelligence-gathering. With the development of the homing torpedo, better sonar systems, and nuclear propulsion, submarines also became able to effectively hunt each other. The development of submarine-launched nuclear and cruise missiles gave submarines a substantial and long-ranged ability to attack both land and sea targets with a variety of weapons ranging from cluster munitions to nuclear weapons. Most navies also include many types of support and auxiliary vessel, such as minesweepers, patrol boats, offshore patrol vessels, replenishment ships, and hospital ships which are designated medical treatment facilities. Fast combat vessels such as cruisers and destroyers usually have fine hulls to maximize speed and maneuverability. They also usually have advanced marine electronics and communication systems, as well as weapons. Some components exist in vessels of any size and purpose. Every vessel has a hull of sorts. Every vessel has some sort of propulsion, whether it's a pole, an ox, or a nuclear reactor. Most vessels have some sort of steering system. Other characteristics are common, but not as universal, such as compartments, holds, a superstructure, and equipment such as anchors and winches. For a ship to float, its weight must be less than that of the water displaced by the ship's hull. There are many types of hulls, from logs lashed together to form a raft to the advanced hulls of America's Cup sailboats. A vessel may have a single hull (called a monohull design), two in the case of catamarans, or three in the case of trimarans. Vessels with more than three hulls are rare, but some experiments have been conducted with designs such as pentamarans. Multiple hulls are generally parallel to each other and connected by rigid arms. Hulls have several elements. The bow is the foremost part of the hull. Many ships feature a bulbous bow. The keel is at the very bottom of the hull, extending the entire length of the ship. The rear part of the hull is known as the stern, and many hulls have a flat back known as a transom. Common hull appendages include propellers for propulsion, rudders for steering, and stabilizers to quell a ship's rolling motion. Other hull features can be related to the vessel's work, such as fishing gear and sonar domes. Hulls are subject to various hydrostatic and hydrodynamic constraints. The key hydrostatic constraint is that it must be able to support the entire weight of the boat, and maintain stability even with often unevenly distributed weight. Hydrodynamic constraints include the ability to withstand shock waves, weather collisions and groundings. Older ships and pleasure craft often have or had wooden hulls. Steel is used for most commercial vessels. Aluminium is frequently used for fast vessels, and composite materials are often found in sailboats and pleasure craft. Some ships have been made with concrete hulls. Propulsion systems for ships fall into three categories: human propulsion, sailing, and mechanical propulsion. Human propulsion includes rowing, which was used even on large galleys. Propulsion by sail generally consists of a sail hoisted on an erect mast, supported by stays and spars and controlled by ropes. Sail systems were the dominant form of propulsion until the 19th century. They are now generally used for recreation and competition, although experimental sail systems, such as the turbosails, rotorsails, and wingsails have been used on larger modern vessels for fuel savings. Mechanical propulsion systems generally consist of a motor or engine turning a propeller, or less frequently, an impeller or wave propulsion fins. Steam engines were first used for this purpose, but have mostly been replaced by two-stroke or four-stroke diesel engines, outboard motors, and gas turbine engines on faster ships. Nuclear reactors producing steam are used to propel warships and icebreakers, and there have been attempts to utilize them to power commercial vessels (see NS "Savannah"). In addition to traditional fixed and controllable pitch propellers there are many specialized variations, such as contra-rotating and nozzle-style propellers. Most vessels have a single propeller, but some large vessels may have up to four propellers supplemented with transverse thrusters for maneuvring at ports. The propeller is connected to the main engine via a propeller shaft and, in case of medium- and high-speed engines, a reduction gearbox. Some modern vessels have a diesel-electric powertrain in which the propeller is turned by an electric motor powered by the ship's generators. For ships with independent propulsion systems for each side, such as manual oars or some paddles, steering systems may not be necessary. In most designs, such as boats propelled by engines or sails, a steering system becomes necessary. The most common is a rudder, a submerged plane located at the rear of the hull. Rudders are rotated to generate a lateral force which turns the boat. Rudders can be rotated by a tiller, manual wheels, or electro-hydraulic systems. Autopilot systems combine mechanical rudders with navigation systems. Ducted propellers are sometimes used for steering. Some propulsion systems are inherently steering systems. Examples include the outboard motor, the bow thruster, and the Z-drive. Larger boats and ships generally have multiple decks and compartments. Separate berthings and heads are found on sailboats over about . Fishing boats and cargo ships typically have one or more cargo holds. Most larger vessels have an engine room, a galley, and various compartments for work. Tanks are used to store fuel, engine oil, and fresh water. Ballast tanks are equipped to change a ship's trim and modify its stability. Superstructures are found above the main deck. On sailboats, these are usually very low. On modern cargo ships, they are almost always located near the ship's stern. On passenger ships and warships, the superstructure generally extends far forward. Shipboard equipment varies from ship to ship depending on such factors as the ship's era, design, area of operation, and purpose. Some types of equipment that are widely found include: Ships float in the water at a level where mass of the displaced water equals the mass of the vessel, such that the downwards force of gravity equals the upward force of buoyancy. As a vessel is lowered into the water its weight remains constant but the corresponding weight of water displaced by its hull increases. If the vessel's mass is evenly distributed throughout, it floats evenly along its length and across its beam (width). A vessel's stability is considered in both this hydrostatic sense as well as a hydrodynamic sense, when subjected to movement, rolling and pitching, and the action of waves and wind. Stability problems can lead to excessive pitching and rolling, and eventually capsizing and sinking. The advance of a vessel through water is resisted by the water. This resistance can be broken down into several components, the main ones being the friction of the water on the hull and wave making resistance. To reduce resistance and therefore increase the speed for a given power, it is necessary to reduce the wetted surface and use submerged hull shapes that produce low amplitude waves. To do so, high-speed vessels are often more slender, with fewer or smaller appendages. The friction of the water is also reduced by regular maintenance of the hull to remove the sea creatures and algae that accumulate there. Antifouling paint is commonly used to assist in this. Advanced designs such as the bulbous bow assist in decreasing wave resistance. A simple way of considering wave-making resistance is to look at the hull in relation to its wake. At speeds lower than the wave propagation speed, the wave rapidly dissipates to the sides. As the hull approaches the wave propagation speed, however, the wake at the bow begins to build up faster than it can dissipate, and so it grows in amplitude. Since the water is not able to "get out of the way of the hull fast enough", the hull, in essence, has to climb over or push through the bow wave. This results in an exponential increase in resistance with increasing speed. This hull speed is found by the formula: or, in metric units: where "L" is the length of the waterline in feet or meters. When the vessel exceeds a speed/length ratio of 0.94, it starts to outrun most of its bow wave, and the hull actually settles slightly in the water as it is now only supported by two wave peaks. As the vessel exceeds a speed/length ratio of 1.34, the hull speed, the wavelength is now longer than the hull, and the stern is no longer supported by the wake, causing the stern to squat, and the bow rise. The hull is now starting to climb its own bow wave, and resistance begins to increase at a very high rate. While it is possible to drive a displacement hull faster than a speed/length ratio of 1.34, it is prohibitively expensive to do so. Most large vessels operate at speed/length ratios well below that level, at speed/length ratios of under 1.0. For large projects with adequate funding, hydrodynamic resistance can be tested experimentally in a hull testing pool or using tools of computational fluid dynamics. Vessels are also subject to ocean surface waves and sea swell as well as effects of wind and weather. These movements can be stressful for passengers and equipment, and must be controlled if possible. The rolling movement can be controlled, to an extent, by ballasting or by devices such as fin stabilizers. Pitching movement is more difficult to limit and can be dangerous if the bow submerges in the waves, a phenomenon called pounding. Sometimes, ships must change course or speed to stop violent rolling or pitching. How it has been convincingly shown in scientific studies of the 21st century, controllability of some vessels decreases dramatically in some cases that are conditioned by effects of the bifurcation memory. This class of vessels includes ships with high manoeuvring capabilities, aircraft and controlled underwater vehicles designed to be unstable in steady-state motion that are interesting in terms of applications. These features must be considered in designing ships and in their control in critical situations. A ship will pass through several stages during its career. The first is usually an initial contract to build the ship, the details of which can vary widely based on relationships between the shipowners, operators, designers and the shipyard. Then, the design phase carried out by a naval architect. Then the ship is constructed in a shipyard. After construction, the vessel is launched and goes into service. Ships end their careers in a number of ways, ranging from shipwrecks to service as a museum ship to the scrapyard. A vessel's design starts with a specification, which a naval architect uses to create a project outline, assess required dimensions, and create a basic layout of spaces and a rough displacement. After this initial rough draft, the architect can create an initial hull design, a general profile and an initial overview of the ship's propulsion. At this stage, the designer can iterate on the ship's design, adding detail and refining the design at each stage. The designer will typically produce an overall plan, a general specification describing the peculiarities of the vessel, and construction blueprints to be used at the building site. Designs for larger or more complex vessels may also include sail plans, electrical schematics, and plumbing and ventilation plans. As environmental laws are becoming more strict, ship designers need to create their design in such a way that the ship, when it nears its end-of-term, can be disassembled or disposed easily and that waste is reduced to a minimum. Ship construction takes place in a shipyard, and can last from a few months for a unit produced in series, to several years to reconstruct a wooden boat like the frigate "Hermione", to more than 10 years for an aircraft carrier. During World War II, the need for cargo ships was so urgent that construction time for Liberty Ships went from initially eight months or longer, down to weeks or even days. Builders employed production line and prefabrication techniques such as those used in shipyards today. Hull materials and vessel size play a large part in determining the method of construction. The hull of a mass-produced fiberglass sailboat is constructed from a mold, while the steel hull of a cargo ship is made from large sections welded together as they are built. Generally, construction starts with the hull, and on vessels over about , by the laying of the keel. This is done in a drydock or on land. Once the hull is assembled and painted, it is launched. The last stages, such as raising the superstructure and adding equipment and accommodation, can be done after the vessel is afloat. Once completed, the vessel is delivered to the customer. Ship launching is often a ceremony of some significance, and is usually when the vessel is formally named. A typical small rowboat can cost under US$100, $1,000 for a small speedboat, tens of thousands of dollars for a cruising sailboat, and about $2,000,000 for a Vendée Globe class sailboat. A trawler may cost $2.5 million, and a 1,000-person-capacity high-speed passenger ferry can cost in the neighborhood of $50 million. A ship's cost partly depends on its complexity: a small, general cargo ship will cost $20 million, a Panamax-sized bulk carrier around $35 million, a supertanker around $105 million and a large LNG carrier nearly $200 million. The most expensive ships generally are so because of the cost of embedded electronics: a costs around $2 billion, and an aircraft carrier goes for about $3.5 billion. Ships undergo nearly constant maintenance during their career, whether they be underway, pierside, or in some cases, in periods of reduced operating status between charters or shipping seasons. Most ships, however, require trips to special facilities such as a drydock at regular intervals. Tasks often done at drydock include removing biological growths on the hull, sandblasting and repainting the hull, and replacing sacrificial anodes used to protect submerged equipment from corrosion. Major repairs to the propulsion and steering systems as well as major electrical systems are also often performed at dry dock. Some vessels that sustain major damage at sea may be repaired at a facility equipped for major repairs, such as a shipyard. Ships may also be converted for a new purpose: oil tankers are often converted into floating production storage and offloading units. Most ocean-going cargo ships have a life expectancy of between 20 and 30 years. A sailboat made of plywood or fiberglass can last between 30 and 40 years. Solid wooden ships can last much longer but require regular maintenance. Carefully maintained steel-hulled yachts can have a lifespan of over 100 years. As ships age, forces such as corrosion, osmosis, and rotting compromise hull strength, and a vessel becomes too dangerous to sail. At this point, it can be scuttled at sea or scrapped by shipbreakers. Ships can also be used as museum ships, or expended to construct breakwaters or artificial reefs. Many ships do not make it to the scrapyard, and are lost in fires, collisions, grounding, or sinking at sea. The Allies lost some 5,150 ships during World War II. One can measure ships in terms of overall length, length between perpendiculars, length of the ship at the waterline, beam (breadth), depth (distance between the crown of the weather deck and the top of the keelson), draft (distance between the highest waterline and the bottom of the ship) and tonnage. A number of different tonnage definitions exist and are used when describing merchant ships for the purpose of tolls, taxation, etc. In Britain until Samuel Plimsoll's Merchant Shipping Act of 1876, ship-owners could load their vessels until their decks were almost awash, resulting in a dangerously unstable condition. Anyone who signed on to such a ship for a voyage and, upon realizing the danger, chose to leave the ship, could end up in jail. Plimsoll, a Member of Parliament, realised the problem and engaged some engineers to derive a fairly simple formula to determine the position of a line on the side of any specific ship's hull which, when it reached the surface of the water during loading of cargo, meant the ship had reached its maximum safe loading level. To this day, that mark, called the "Plimsoll Line", exists on ships' sides, and consists of a circle with a horizontal line through the centre. On the Great Lakes of North America the circle is replaced with a diamond. Because different types of water (summer, fresh, tropical fresh, winter north Atlantic) have different densities, subsequent regulations required painting a group of lines forward of the Plimsoll mark to indicate the safe depth (or freeboard above the surface) to which a specific ship could load in water of various densities. Hence the "ladder" of lines seen forward of the Plimsoll mark to this day. This is called the "freeboard mark" or "load line mark" in the marine industry. Ship pollution is the pollution of air and water by shipping. It is a problem that has been accelerating as trade has become increasingly globalized, posing an increasing threat to the world's oceans and waterways as globalization continues. It is expected that "shipping traffic to and from the United States is projected to double by 2020." Because of increased traffic in ocean ports, pollution from ships also directly affects coastal areas. The pollution produced affects biodiversity, climate, food, and human health. However, the degree to which humans are polluting and how it affects the world is highly debated and has been a hot international topic for the past 30 years. Oil spills have devastating effects on the environment. Crude oil contains polycyclic aromatic hydrocarbons (PAHs) which are very difficult to clean up, and last for years in the sediment and marine environment. Marine species constantly exposed to PAHs can exhibit developmental problems, susceptibility to disease, and abnormal reproductive cycles. By the sheer amount of oil carried, modern oil tankers must be considered something of a threat to the environment. An oil tanker can carry of crude oil, or . This is more than six times the amount spilled in the widely known "Exxon Valdez" incident. In this spill, the ship ran aground and dumped of oil into the ocean in March 1989. Despite efforts of scientists, managers, and volunteers, over 400,000 seabirds, about 1,000 sea otters, and immense numbers of fish were killed. The International Tanker Owners Pollution Federation has researched 9,351 accidental spills since 1974. According to this study, most spills result from routine operations such as loading cargo, discharging cargo, and taking on fuel oil. 91% of the operational oil spills were small, resulting in less than 7 tons per spill. Spills resulting from accidents like collisions, groundings, hull failures, and explosions are much larger, with 84% of these involving losses of over 700 tons. Following the "Exxon Valdez" spill, the United States passed the Oil Pollution Act of 1990 (OPA-90), which included a stipulation that all tankers entering its waters be double-hulled by 2015. Following the sinkings of "Erika" (1999) and "Prestige" (2002), the European Union passed its own stringent anti-pollution packages (known as Erika I, II, and III), which require all tankers entering its waters to be double-hulled by 2010. The Erika packages are controversial because they introduced the new legal concept of "serious negligence". When a large vessel such as a container ship or an oil tanker unloads cargo, seawater is pumped into other compartments in the hull to help stabilize and balance the ship. During loading, this ballast water is pumped out from these compartments. One of the problems with ballast water transfer is the transport of harmful organisms. Meinesz believes that one of the worst cases of a single invasive species causing harm to an ecosystem can be attributed to a seemingly harmless jellyfish. "Mnemiopsis leidyi", a species of comb jellyfish that inhabits estuaries from the United States to the Valdés peninsula in Argentina along the Atlantic coast, has caused notable damage in the Black Sea. It was first introduced in 1982, and thought to have been transported to the Black Sea in a ship's ballast water. The population of the jellyfish shot up exponentially and, by 1988, it was wreaking havoc upon the local fishing industry. "The anchovy catch fell from in 1984 to in 1993; sprat from in 1984 to in 1993; horse mackerel from in 1984 to zero in 1993." Now that the jellyfish have exhausted the zooplankton, including fish larvae, their numbers have fallen dramatically, yet they continue to maintain a stranglehold on the ecosystem. Recently the jellyfish have been discovered in the Caspian Sea. Invasive species can take over once occupied areas, facilitate the spread of new diseases, introduce new genetic material, alter landscapes and jeopardize the ability of native species to obtain food. "On land and in the sea, invasive species are responsible for about 137 billion dollars in lost revenue and management costs in the U.S. each year." Ballast and bilge discharge from ships can also spread human pathogens and other harmful diseases and toxins potentially causing health issues for humans and marine life alike. Discharges into coastal waters, along with other sources of marine pollution, have the potential to be toxic to marine plants, animals, and microorganisms, causing alterations such as changes in growth, disruption of hormone cycles, birth defects, suppression of the immune system, and disorders resulting in cancer, tumors, and genetic abnormalities or even death. Exhaust emissions from ships are considered to be a significant source of air pollution. "Seagoing vessels are responsible for an estimated 14 percent of emissions of nitrogen from fossil fuels and 16 percent of the emissions of sulfur from petroleum uses into the atmosphere." In Europe ships make up a large percentage of the sulfur introduced to the air, "as much sulfur as all the cars, lorries and factories in Europe put together". "By 2010, up to 40% of air pollution over land could come from ships." Sulfur in the air creates acid rain which damages crops and buildings. When inhaled, sulfur is known to cause respiratory problems and increase the risk of a heart attack. Ship breaking or ship demolition is a type of ship disposal involving the breaking up of ships for scrap recycling, with the hulls being discarded in ship graveyards. Most ships have a lifespan of a few decades before there is so much wear that refitting and repair becomes uneconomical. Ship breaking allows materials from the ship, especially steel, to be reused. In addition to steel and other useful materials, however, ships (particularly older vessels) can contain many substances that are banned or considered dangerous in developed countries. Asbestos and polychlorinated biphenyls (PCBs) are typical examples. Asbestos was used heavily in ship construction until it was finally banned in most of the developed world in the mid 1980s. Currently, the costs associated with removing asbestos, along with the potentially expensive insurance and health risks, have meant that ship-breaking in most developed countries is no longer economically viable. Removing the metal for scrap can potentially cost more than the scrap value of the metal itself. In most of the developing world, however, shipyards can operate without the risk of personal injury lawsuits or workers' health claims, meaning many of these shipyards may operate with high health risks. Furthermore, workers are paid very low rates with no overtime or other allowances. Protective equipment is sometimes absent or inadequate. Dangerous vapors and fumes from burning materials can be inhaled, and dusty asbestos-laden areas around such breakdown locations are commonplace. Aside from the health of the yard workers, in recent years, ship breaking has also become an issue of major environmental concern. Many developing nations, in which ship breaking yards are located, have lax or no environmental law, enabling large quantities of highly toxic materials to escape into the environment and causing serious health problems among ship breakers, the local population and wildlife. Environmental campaign groups such as Greenpeace have made the issue a high priority for their campaigns. Model ships Lists Ship sizes
https://en.wikipedia.org/wiki?curid=27008
Soap opera A soap opera is a radio or television serial dealing especially with domestic situations and frequently characterized by melodrama and sentimentality. The term "soap opera" originated from radio dramas being sponsored by soap manufacturers. BBC Radio's "The Archers", first broadcast in 1950, is the world's longest-running radio soap opera. The world's longest-running television soap opera is "Coronation Street", which was first broadcast on ITV in 1960. A crucial element that defines the soap opera is the open-ended serial nature of the narrative, with stories spanning several episodes. One of the defining features that makes a television program a soap opera, according to Albert Moran, is "that form of television that works with a continuous open narrative. Each episode ends with a promise that the storyline is to be continued in another episode". In 2012, "Los Angeles Times" columnist Robert Lloyd wrote of daily dramas: Soap opera storylines run concurrently, intersect and lead into further developments. An individual episode of a soap opera will generally switch between several concurrent narrative threads that may at times interconnect and affect one another or may run entirely independent to each other. Episodes may feature some of the show's current storylines, but not always all of them. Especially in daytime serials and those that are broadcast each weekday, there is some rotation of both storyline and actors so any given storyline or actor will appear in some but usually not all of a week's worth of episodes. Soap operas rarely bring all the current storylines to a conclusion at the same time. When one storyline ends, there are several other story threads at differing stages of development. Soap opera episodes typically end on some sort of cliffhanger, and the season finale (if a soap incorporates a break between seasons) ends in the same way, only to be resolved when the show returns for the start of a new yearly broadcast. Evening soap operas and those that air at a rate of one episode per week are more likely to feature the entire cast in each episode, and to represent all current storylines in each episode. Evening soap operas and serials that run for only part of the year tend to bring things to a dramatic end-of-season cliffhanger. In 1976, "Time" magazine described American daytime television as "TV's richest market", noting the loyalty of the soap opera fan base and the expansion of several half-hour series into hour-long broadcasts in order to maximize ad revenues. The article explained that at that time, many prime time series lost money, while daytime serials earned profits several times more than their production costs. The issue's cover notably featured its first daytime soap stars, Bill Hayes and Susan Seaforth Hayes of "Days of Our Lives", a married couple whose onscreen and real-life romance was widely covered by both the soap opera magazines and the mainstream press at large. The first serial considered to be a "soap opera" was "Painted Dreams", which debuted on October 20, 1930 on Chicago radio station WGN. Early radio series such as "Painted Dreams" were broadcast in weekday daytime slots, usually five days a week. Most of the listeners would be housewives, thus, the shows were aimed at, and consumed by, a predominantly female audience. The first nationally broadcast radio soap opera was "Clara, Lu, and Em", which aired on the NBC Blue Network at 10:30 p.m. Eastern Time on January 27, 1931. The main characteristics that define soap operas are "an emphasis on family life, personal relationships, sexual dramas, emotional and moral conflicts; some coverage of topical issues; set in familiar domestic interiors with only occasional excursions into new locations". Fitting in with these characteristics, most soap operas follow the lives of a group of characters who live or work in a particular place, or focus on a large extended family. The storylines follow the day-to-day activities and personal relationships of these characters. "Soap narratives, like those of film melodramas, are marked by what Steve Neale has described as 'chance happenings, coincidences, missed meetings, sudden conversions, last-minute rescues and revelations, deus ex machina endings.'" These elements may be found across the gamut of soap operas, from "EastEnders" to "Dallas". In many soap operas, in particular daytime serials in the US, the characters are frequently attractive, seductive, glamorous and wealthy. Soap operas from the United Kingdom and Australia tend to focus on more everyday characters and situations, and are frequently set in working-class environments. Many of the soaps produced in those two countries explore social realist storylines such as family discord, marriage breakdown or financial problems. Both UK and Australian soap operas feature comedic elements, often affectionate comic stereotypes such as the gossip or the grumpy old man, presented as a comic foil to the emotional turmoil that surrounds them. This diverges from US soap operas where such comedy is rare. UK soap operas frequently make a claim to presenting "reality" or purport to have a "realistic" style. UK soap operas also frequently foreground their geographic location as a key defining feature of the show while depicting and capitalising on the exotic appeal of the stereotypes connected to the location. As examples, "EastEnders" focuses on the tough and grim life in the East End of London; "Coronation Street" and its characters exhibit the stereotypical characteristic of "northern straight talking". Romance, secret relationships, extramarital affairs, and genuine hate have been the basis for many soap opera storylines. In US daytime serials, the most popular soap opera characters, and the most popular storylines, often involved a romance of the sort presented in paperback romance novels. Soap opera storylines weave intricate, convoluted and sometimes confusing tales of characters who have affairs, meet mysterious strangers and fall in love, and who commit adultery, all of which keeps audiences hooked on the unfolding story. Crimes such as kidnapping, rape, and even murder may go unpunished if the perpetrator is to be retained in the ongoing story. Australian and UK soap operas also feature a significant proportion of romance storylines. In Russia, most popular serials explore the "romantic quality" of criminal and/or oligarch life. In soap opera storylines, previously unknown children, siblings and twins (including the evil variety) of established characters often emerge to upset and reinvigorate the set of relationships examined by the series. Unexpected calamities disrupt weddings, childbirths, and other major life events with unusual frequency. As in comic books – another popular form of linear storytelling pioneered in the US during the 20th century – a character's death is not guaranteed to be permanent. On "The Bold and the Beautiful", Taylor Forrester (Hunter Tylo) was shown to flatline and have a funeral. Once Tylo reprised her character in 2005, a retcon explained that Taylor had actually gone into a coma. Stunts and complex physical action are largely absent, especially from daytime serials. Such story events often take place off-screen and are referred to in dialogue instead of being shown. This is because stunts or action scenes are difficult to adequately depict without complex movements, multiple takes, and post-production editing. When episodes were broadcast live, post-production work was impossible. Though all serials have long switched to being taped, extensive post-production work and multiple takes, while possible, are not feasible due to the tight taping schedules and low budgets. The first daytime TV soap opera in the United States was "These Are My Children" in 1949, though earlier melodramas had aired in the evenings as once-a-week programs. Soap operas quickly became a fixture of American daytime television in the early 1950s, joined by game shows, sitcom reruns and talk shows. In 1988, H. Wesley Kenney, who at the time served as the executive producer of "General Hospital", said to "The New York Times": Many long-running US soap operas established particular environments for their stories. "The Doctors" and "General Hospital", in the beginning, told stories almost exclusively from inside the confines of a hospital. "As the World Turns" dealt heavily with Chris Hughes' law practice and the travails of his wife Nancy who, tired of being "the loyal housewife" in the 1970s, became one of the first older women on the American serials to enter the workforce. "Guiding Light" dealt with Bert Bauer (Charita Bauer) and her alcoholic husband Bill, and their endless marital troubles. When Bert's status shifted to caring mother and town matriarch, her children's marital troubles were showcased. "Search for Tomorrow" mostly told its story through the eyes of Joanne Gardner (Mary Stuart). Even when stories revolved around other characters, Joanne was frequently a key player in their storylines. "Days of Our Lives" initially focused on Dr. Tom Horton and his steadfast wife Alice. The show later branched out to focus more on their five children. "The Edge of Night" featured as its central character Mike Karr, a police detective (later an attorney), and largely dealt with organized crime. "The Young and the Restless" first focused on two families, the prosperous Brooks family with four daughters, and the working class Foster family of a single working mother with three children. Its storylines explored realistic problems including cancer, mental illness, poverty, and infidelity. In contrast, "Dark Shadows" (1966–1971), "Port Charles" (1997–2003) and "Passions" (1999-2008) featured supernatural characters and dealt with fantasy and horror storylines. Their characters included vampires, witches, ghosts, goblins, and angels. The American soap opera "Guiding Light" (originally titled "The Guiding Light" until 1975) started as a radio drama in January 1937 and subsequently transferred to television in June 1952. With the exception of several years in the late 1940s, during which creator Irna Phillips was involved in a dispute with Procter & Gamble, "Guiding Light" was heard or seen nearly every weekday from 1937 to 2009, making it the longest story ever told in a broadcast medium. Originally serials were broadcast as 15-minute installments each weekday in daytime slots. In 1956, "As the World Turns" and "The Edge of Night", both produced by Procter & Gamble Productions, debuted as the first half-hour soap operas on the CBS television network. All soap operas broadcast half-hour episodes by the end of the 1960s. With increased popularity in the 1970s, most soap operas had expanded to an hour in length by the end of the decade ("Another World" even expanded to 90 minutes for a short time). More than half of the serials had expanded to one-hour episodes by 1980. As of 2012, three of the four US serials air one-hour episodes each weekday; only "The Bold and the Beautiful" airs 30-minute episodes. Soap operas were originally broadcast live from the studio, creating what many at the time regarded as a feeling similar to that of a stage play. As nearly all soap operas were originated at that time from New York City, a number of soap actors were also accomplished stage actors who performed live theater during breaks from their soap roles. In the 1960s and 1970s, new serials such as "General Hospital", "Days of our Lives" and "The Young and the Restless" were produced in Los Angeles. Their success made the West Coast a viable alternative to New York-produced soap operas, which were becoming more costly to perform. By the early 1970s, nearly all soap operas had transitioned to being taped. "As the World Turns" and "The Edge of Night" were the last to make the switch, in 1975. "Port Charles" used the practice of running 13-week "story arcs," in which the main events of the arc are played out and wrapped up over the 13 weeks, although some storylines did continue over more than one arc. According to the 2006 Preview issue of "Soap Opera Digest", it was briefly discussed that all ABC shows might do telenovela arcs, but this was rejected. Though U.S. daytime soap operas are not generally rerun by their networks, occasionally they are rebroadcast elsewhere; CBS and ABC have made exceptions to this, airing older episodes (either those aired earlier in the current season or those aired years prior) on major holidays when special event programming is not scheduled. Early episodes of "Dark Shadows" were rerun on PBS member stations in the early 1970s after the show's cancellation, and the entire series (except for a single missing episode) was rerun on the Sci-Fi Channel in the 1990s. After "The Edge of Night"s 1984 cancellation, reruns of the show's final five years were shown late nights on USA Network from 1985 to 1989. On January 20, 2000, a digital cable and satellite network dedicated to the genre, Soapnet, began re-airing soaps that originally aired on ABC, NBC and CBS. Newer broadcast networks since the late 1980s, such as Fox and cable television networks, have largely eschewed soap operas in their daytime schedules, instead running syndicated programming and reruns. No cable television outlet has produced its own daytime serial, although DirecTV's The 101 Network took over existing serial "Passions", continuing production for one season; while TBS and CBN Cable Network respectively aired their own soap operas, "The Catlins" (a primetime soap that utilized the daily episode format of its daytime counterparts) and "Another Life" (a soap that combined standard serial drama with religious overtones), during the 1980s. Fox, the fourth "major network," carried a short lived daytime soap "Tribes" in 1990. Yet other than this and a couple of pilot attempts, Fox mainly stayed away from daytime soaps, and has not attempted them since their ascension to major-network status in 1994 (it did later attempt a series of daily prime time soaps, which aired on newly created sister network MyNetworkTV, but the experiment was largely a failure). Due to the masses of episodes produced for a series, release of soap operas to DVD (a popular venue for distribution of current and vintage television series) is considered impractical. With the exception of occasional specials, daytime soap operas are notable by their absence from DVD release schedules (an exception being the supernatural soap opera, "Dark Shadows", which did receive an essentially complete release on both VHS and DVD; the single lost episode #1219 is reconstructed by means of an off-the-air audio recording, still images, and recap material from adjacent episodes). Due to the longevity of these shows, it is not uncommon for a single character to be played by multiple actors. The key character of Mike Karr on "The Edge of Night" was played by three actors. Conversely, several actors have remained playing the same character for many years, or decades even. Helen Wagner played Hughes family matriarch Nancy Hughes on American soap "As the World Turns" from its April 2, 1956 debut through her death in May 2010. She is listed in the Guinness Book of World Records as the actor with the longest uninterrupted performance in a single role. A number of performers played roles for 20 years or longer, occasionally on more than one show. Rachel Ames played Audrey Hardy on both "General Hospital" and "Port Charles" from 1964 until 2007, and returned in 2009. Susan Lucci played Erica Kane in "All My Children" from the show's debut in January 1970 until it ended its network television run on ABC on September 23, 2011. Erika Slezak played Victoria Lord #3 on "One Life to Live" from 1971 until the show ended its network television run on ABC on January 13, 2012 and resumed the role in its short-lived online revival on April 29, 2013. Other actors have played several characters on different shows. Millette Alexander, Bernard Barrow, Doris Belack, David Canary, Judith Chapman, Jordan Charney, Joan Copeland, Nicolas Coster, Jacqueline Courtney, Louis Edmonds, Dan Hamilton, Don Hastings, Vincent Irizarry, Lenore Kasdorf, Teri Keane, Lois Kibbee, John Loprieno, Maeve McGuire, James Mitchell, Christopher Pennock, Antony Ponzini, William Prince, Louise Shaffer, and Diana van der Vlis, among many others, have all played multiple soap roles. For several decades, most daytime soap operas concentrated on family and marital discord, legal drama and romance. The action rarely left interior settings, and many shows were set in fictional, medium-sized Midwestern towns. Exterior shots were slowly incorporated into the series "The Edge of Night" and "Dark Shadows". Unlike many earlier serials that were set in fictional towns, "The Best of Everything" and "Ryan's Hope" were set in a real-world location, New York City. The first exotic location shoot was made by "All My Children", to St. Croix in 1978. Many other soap operas planned lavish storylines after the success of the "All My Children" shoot. Soap operas "Another World" and "Guiding Light" both went to St. Croix in 1980, the former show culminating a long-running storyline between popular characters Mac, Rachel and Janice, and the latter to serve as an exotic setting for Alan Spaulding and Rita Bauer's torrid affair. "Search for Tomorrow" taped for two weeks in Hong Kong in 1981. Later that year, some of the cast and crew ventured to Jamaica to tape a love consummation storyline between the characters of Garth and Kathy. During the 1980s, perhaps as a reaction to the evening drama series that were gaining high ratings, daytime serials began to incorporate action and adventure storylines, more big-business intrigue, and an increased emphasis on youthful romance. One of the first and most popular couples was Luke Spencer and Laura Webber on "General Hospital". Luke and Laura helped to attract both male and female fans. Even actress Elizabeth Taylor was a fan and at her own request was given a guest role in Luke and Laura's wedding episode. Luke and Laura's popularity led to other soap producers striving to reproduce this success by attempting to create supercouples of their own. With increasingly bizarre action storylines coming into vogue, Luke and Laura saved the world from being frozen, brought a mobster down by finding his black book in a left-handed boy statue, and helped a princess find her Aztec treasure in Mexico. Other soap operas attempted similar adventure storylines, often featuring footage shot on location – frequently in exotic locales. During the 1990s, the mob, action and adventure stories fell out of favor with producers, due to generally declining ratings for daytime soap operas at the time, and the resultant budget cuts. In addition, soap operas were no longer able to go on expensive location shoots overseas as they were able to do in the 1980s. During that decade, soap operas increasingly focused on younger characters and social issues, such as Erica Kane's drug addiction on "All My Children", the re-emergence of Viki Lord's multiple personality disorder on "One Life to Live", and Stuart Chandler dealing with his wife Cindy dying of AIDS on "All My Children". Other social issues included cancer, rape, abortion, homophobia, and racism. Some shows during the 2000s incorporated supernatural and science fiction elements into their storylines. One of the main characters on the earlier soap opera "Dark Shadows" was Barnabas Collins, a vampire, and "One Life to Live" featured an angel named Virgil. Both shows featured characters who traveled to and from the past. Modern U.S. daytime soap operas largely stay true to the original soap opera format. The duration and format of storylines and the visual grammar employed by U.S. daytime serials set them apart from soap operas in other countries and from evening soap operas. Stylistically, UK and Australian soap operas, which are usually produced for early evening timeslots, fall somewhere in between U.S. daytime and evening soap operas. Similar to U.S. daytime soap operas, UK and Australian serials are shot on videotape, and the cast and storylines are rotated across the week's episodes so that each cast member will appear in some but not all episodes. UK and Australian soap operas move through storylines at a faster rate than daytime serials, making them closer to U.S. evening soap operas in this regard. American daytime soap operas feature stylistic elements that set them apart from other shows: Soap opera ratings have significantly fallen in the U.S. since the 2000s. No new major daytime soap opera has been created since "Passions" in 1999, while many have been cancelled. Since January 2012, four daytime soap operas – "General Hospital", "Days of Our Lives", "The Young and the Restless" and "The Bold and the Beautiful" – continue to air on the three major networks, down from a total of 12 during the 1990–91 season and a high of 19 in the 1969–70 season. This marks the first time since 1953 that there have been only four soap operas airing on broadcast television. "The Young and the Restless", the highest-rated soap opera from 1988 to the present, had fewer than 5 million daily viewers as of February 2012, a number exceeded by several non-scripted programs such as "Judge Judy". Circulations of soap opera magazines have decreased and some have even ceased publication. Soapnet, which largely aired soap opera reruns, began to be phased out in 2012 and fully ceased operations the following year. The Daytime Emmy Awards, which honor soap operas and other daytime shows, moved from prime time network television to smaller cable channels in 2012, then failed to get any TV broadcast at all in 2014, 2016, and 2017. Several of the U.S.’s most established soaps ended between 2009 and 2012. The longest-running drama in television and radio history, "Guiding Light", barely reached 2.1 million daily viewers in 2009 and ended on September 18 of that year, after a 72-year run (including radio). "As the World Turns" aired its final episode on September 17, 2010 after a 54-year run. "As the World Turns" was the last of 20 soap operas produced by Procter & Gamble, the soap and consumer goods company from which the genre got its name. "As the World Turns" and "Guiding Light" were also among the last of the soaps that originated from New York City. "All My Children", another New York-based soap, moved its production out to Los Angeles in an effort to reduce costs and raise sagging ratings; however, both it and "One Life to Live", each with a 40-year-plus run, were cancelled in 2011. "All My Children" aired its network finale in September 2011, with "One Life to Live" following suit in January 2012. Both "All My Children" and "One Life to Live" were briefly revived online in 2013, before being cancelled again that same year. In 2019, "Days of Our Lives" was put on "indefinite hiatus" and all of the cast's contracts were terminated. As women increasingly worked outside of the home, daytime television viewing declined. New generations of potential viewers were not raised watching soap operas with their mothers, leaving the shows' long and complex storylines foreign to younger audiences. Now, as viewers age, ratings continue to drop among young adult women, the demographic group that soap opera advertisers pay the most for. Those who might watch in workplace breakrooms are not counted, as Nielsen does not track television viewing outside the home. The rise of cable and the Internet has also provided new sources of entertainment during the day. The genre's decline has additionally been attributed to reality television displacing soap operas as TV's dominant form of melodrama. An early term for the reality TV genre was "docu-soap". A precursor to reality TV, the televised 1994–95 O. J. Simpson murder case, both preempted and competed with an entire season of soaps, transforming viewing habits and leaving soap operas with 10 percent fewer viewers after the trial ended. Daytime programming alternatives such as talk shows, game shows, and court shows cost up to 50% less to produce than scripted dramas, making those formats more profitable and attractive to networks, even if they receive the same or slightly lower ratings than soap operas. A network may even prefer to return a time slot to its local stations to keeping a soap opera with disappointing ratings on the air, as was the case with "Sunset Beach" and "Port Charles". Compounding the financial pressure on scripted programming in the 2007–2010 period was a decline in advertising during the Great Recession, which led shows to reduce their budgets and cast sizes. In addition to these external factors, a litany of production decisions has been cited by soap opera fans as contributing to the genre's decline, such as cliched plots, a lack of diversity that narrowed audience appeal, and the elimination of core families. Serials produced for prime time slots have also found success. The first prime time soap opera was "Faraway Hill" (1946), which aired on October 2, 1946, on the now-defunct DuMont Television Network. "Faraway Hill" ran for 12 episodes and was primarily broadcast live, interspersed with short pre-recorded film clips and still photos to remind the audience of the previous week's episode. The first long-running prime time soap opera was "Peyton Place" (1964–1969) on ABC. It was based in part on the eponymous 1957 film (which, in turn, was based on the 1956 novel). The popularity of "Peyton Place" prompted the CBS network to spin-off popular "As the World Turns" character Lisa Miller into her own evening soap opera, "Our Private World" (originally titled "The Woman Lisa" in its planning stages). "Our Private World" was broadcast from May to September 1965. The character of Lisa (and her portrayer Eileen Fulton) returned to "As The World Turns" after the series ended. The structure of "Peyton Place", with its episodic plots and long-running story arcs, set the mold for the prime time serials of the 1980s, when the format reached its pinnacle. The successful prime time serials of the 1980s included "Dallas", its spin-off "Knots Landing", "Dynasty", and "Falcon Crest". These shows frequently dealt with wealthy families, and their personal and big-business travails. Common characteristics were sumptuous sets and costumes, complex storylines examining business schemes and intrigue, and spectacular disaster cliffhanger situations. Each of these series featured a wealthy, domineering, promiscuous, and passionate antagonist as a key character in the storyline – respectively, J.R. Ewing (Larry Hagman), Alexis Colby (Joan Collins), Abby Cunningham (Donna Mills) and Angela Channing (Jane Wyman). These villainous schemers became immensely popular figures that audiences "loved to hate". Unlike daytime serials, which are shot on video in a studio using the multi-camera setup, these evening series were shot on film using a single camera setup, and featured substantial location-shot footage, often in picturesque locales. "Dallas", its spin-off "Knots Landing", and "Falcon Crest" all initially featured episodes with self-contained stories and specific guest stars who appeared in just that episode. Each story was completely resolved by the end of the episode, and there were no end-of-episode cliffhangers. After the first couple of seasons, all three shows changed their story format to that of a pure soap opera, with interwoven ongoing narratives that ran over several episodes. "Dynasty" featured this format throughout its run. The soap opera's distinctive open plot structure and complex continuity was increasingly incorporated into American prime time television programs of the period. The first significant drama series to do this was "Hill Street Blues". This series, produced by Steven Bochco, featured many elements borrowed from soap operas, such as an ensemble cast, multi-episode storylines, and extensive character development over the course of the series. It and the later "Cagney & Lacey" overlaid the police series formula with ongoing narratives exploring the personal lives and interpersonal relationships of the regular characters. The success of these series prompted other drama series, such as "St. Elsewhere" and situation comedy series, to incorporate serialized stories and story structure to varying degrees. The prime time soap operas and drama series of the 1990s, such as "Beverly Hills, 90210", "Melrose Place", "Party of Five", "The OC", and "Dawson's Creek", focused more on younger characters. In the 2000s, ABC began to revitalize the prime time soap opera format with shows such as "Desperate Housewives", "Grey's Anatomy", "Brothers & Sisters", "Ugly Betty", "Private Practice", and more recently "Revenge", "Nashville", "Scandal", "Mistresses", and formerly "Ringer", which its sister production company ABC Studios co-produced with CBS Television Studios for The CW. While not soaps in the traditional sense, these shows managed to appeal to wide audiences with their high drama mixed with humor, and are soap operas by definition. These successes led to NBC's launching serials, including "Heroes" and "Friday Night Lights". The upstart MyNetworkTV, a sister network of Fox, launched a line of prime time telenovelas (a genre similar to soap operas in terms of content) upon its launch in September 2006, but discontinued its use of the format in 2007 after disappointing ratings. On June 13, 2012, "Dallas", a continuation of the 1978 original series premiered on the cable network, TNT. The revived series, which was canceled after three seasons in 2014, delivered solid ratings for the channel, only losing viewership after the show's most established star, Larry Hagman, died midway through the series. In 2012, Nick at Nite debuted a primetime soap opera, "Hollywood Heights", which aired episodes five nights a week (on Monday through Fridays) in a manner similar to a daytime soap opera, instead of the once-a-week episode output common of other prime time soaps. The series, which was an adaptation of the Mexican telenovela "Alcanzar una estrella", suffered from low ratings (generally receiving less than 1 million viewers) and was later moved to sister cable channel TeenNick halfway through its run to burn off the remaining episodes. In 2015, Fox debuted "Empire", a prime time musical serial centering on the power struggle between family members within the titular recording company. Created by Lee Daniels and Danny Strong and led by Oscar nominees Terrence Howard and Taraji P. Henson, the drama premiered to high ratings. The show is strongly influenced by other works such as William Shakespeare's "King Lear", James Goldman's "The Lion in Winter" and the 1980s soap opera "Dynasty". Also in 2015, E! introduced "The Royals", a series following the life and drama of a fictional English Royal family, which was also inspired by "Dynasty" (even featuring Joan Collins as the Queen's mother). In addition, ABC debuted a prime time soap opera "Blood & Oil", following a young couple looking to make money off the modern-day Williston oil boom, premiering on September 27, 2015 during the 2015-16 TV season. The telenovela, a shorter-form format of serial melodrama, shares some thematic and especially stylistic similarity to the soap opera, enough that the colloquialism "Spanish soap opera" has arisen to describe the format. The chief difference between the two is length of series; while soap operas usually have indefinite runs, telenovelas typically have a central story arc with a prescribed ending within a year or two of the show's launch, requiring more concise storytelling. Spanish-language networks, chiefly Univision and Telemundo, have found success airing telenovelas for the growing U.S. Hispanic market. Both originally produced and imported Latin American dramas are popular features of the networks' daytime and primetime lineups, sometimes beating English-language networks in the ratings. Some web series are soap operas, such as "" or "". In 2013, production company Prospect Park revived "All My Children" and "One Life to Live" for the web, retaining original creator Agnes Nixon as a consultant and keeping many of the same actors (Prospect Park purchased the rights to both series months after their cancellations by ABC in 2011, although it initially suspended plans to relaunch the soaps later that same year due to issues receiving approval from acting and production unions). Each show initially produced four half-hour episodes a week, but quickly cut back to two half-hour episodes each. In the midst of (though not directly related to) a lawsuit between Prospect Park and ABC, the experiment ended that same year, with both shows being canceled again. As of 2017, Turkey is the second largest exporter of television soap operas. In 2016, Turkish TV exports earned $350 million, making it the second largest drama exporter in the world behind the United States. Turkish soap operas have a large following across Asia, the Balkans, Eastern Europe, Latin America, the Middle East, and Africa. Soap operas in the UK began on radio and consequently were associated with the BBC. It had resisted soaps as antithetical to its quality image, but began broadcasting "Front Line Family" in April 1941 on its North American shortwave service to encourage American intervention on Britain's behalf in World War II. The BBC continues to broadcast the world's longest-running radio soap, "The Archers", which first aired in May 1950, and has been running nationally since 1951. It is currently broadcast on BBC Radio 4 and continues to attract over five million listeners, or roughly 25% of the radio listening population of the UK at that time of the evening. In the UK, soap operas are one of the most popular genres, with most being broadcast during prime time. Most UK soap operas focus on everyday, working-class communities, influenced by the conventions of the kitchen sink drama. The most popular British soap operas are "EastEnders", "Coronation Street", "Emmerdale", "Hollyoaks", "Doctors", and the Australian produced "Neighbours" and "Home and Away". The first three of these are consistently among the highest-rated shows on British television. Such is the magnitude of the popularity of the soap genre in the UK that all television serials in the country are reputedly enjoyed by members of the British Royal Family, including Elizabeth II herself. Major events in British culture are often mentioned in the storyline, such as England's participation at the World Cup, and the death of Princess Diana. Since 1999, The British Soap Awards has been televised on ITV. The 1986 Christmas Day episode of "EastEnders" is often referred to as the highest-rated UK soap opera episode ever, with 30.15 million viewers (more than half the population at the time). The figure of 30.15 million was actually a combination of the original broadcast, which had just over 19 million viewers, and the Sunday omnibus edition with 10 million viewers. The combined 30.15 million audience figure makes the aforementioned Christmas Day 1986 episode of "EastEnders" the highest-rated single-channel broadcast in the history of UK television. Overall it ranks third behind the 1966 FIFA World Cup Final (32.3 million viewers) and Princess Diana's funeral in 1997 (32.1 million viewers) which were transmitted on both BBC One and ITV. An early television serial was "The Grove Family" on the BBC, which produced 148 episodes from 1954 to 1957. The programme was broadcast live and only a handful of recordings were retained in the archives. The UK's first twice-weekly serial was ITV's "Emergency - Ward 10", running from 1957 until 1967. In the 1960s, "Coronation Street" revolutionised UK television and quickly became a British institution. On 17 September 2010, it became the world's longest-running television soap opera and was listed in "Guinness World Records". The BBC also produced several serials: "Compact" was about the staff of a women's magazine; "The Newcomers" was about the upheaval caused by a large firm setting up a plant in a small town; "United!" contained 147 episodes and focused on a football team; "199 Park Lane" (1965) was an upper class serial, which ran for only 18 episodes. None of these serials came close to making the same impact as "Coronation Street". Indeed, most of the 1960s BBC serials were largely wiped. During the 1960s, "Coronation Street"s main rival was "Crossroads", a daily serial that began in 1964 and aired on ITV in the early evening. "Crossroads" was set in a Birmingham motel and, although the program was popular, its purported low technical standard and bad acting were much mocked. By the 1980s, its ratings had begun to decline. Several attempts to revamp the program through cast changes and, later, expanding the focus from the motel to the surrounding community were unsuccessful. "Crossroads" was cancelled in 1988 (a new version of "Crossroads" was later produced, running from 2001 until 2003). A later rival to "Coronation Street" was ITV's "Emmerdale Farm" (later renamed "Emmerdale"), which began in 1972 in a daytime slot and was set in rural Yorkshire. Increased viewership resulted in "Emmerdale" being moved to a prime-time slot in the 1980s. "Pobol y Cwm" ("People of the Valley") is a Welsh language serial that has been produced by the BBC since October 1974, and is the longest-running television soap opera produced by the broadcaster. "Pobol y Cwm" was originally broadcast on BBC Wales television from 1974 to 1982; it was then moved to the Welsh-language television station S4C when it opened in November 1982. The program was occasionally shown on BBC1 in London during periods of regional optout in the mid- to late 1970s. "Pobol y Cwm" was briefly shown in the rest of the UK in 1994 on BBC2, with English subtitles; it is consistently the most watched programme each week on S4C. Daytime soap operas were non-existent until the 1970s because there was virtually no daytime television in the UK. ITV introduced "General Hospital", which later moved to a prime time slot. In 1980, Scottish Television debuted "Take the High Road", which lasted for over twenty years. Later, daytime slots were filled with an influx of Australian soap operas such as "The Sullivans" (aired on ITV from 1977), "The Young Doctors" (from 1982), "Sons and Daughters" (from 1983), "A Country Practice" (from 1982), "Richmond Hill" (from 1988 to 1989) and eventually, "Neighbours" was acquired by the BBC in 1986, and "Home and Away" aired on ITV beginning in 1989. These achieved significant levels of popularity; "Neighbours" and "Home and Away" were moved to early-evening slots, helping begin the UK soap opera boom in the late 1980s. The day Channel 4 began operations in 1982 it launched its own soap, the Liverpool-based "Brookside", which would redefine soaps over the next decade. The focus of "Brookside" was different from earlier soap operas in the UK; it was set in a middle-class new-build cul-de-sac, unlike "Coronation Street" and "Emmerdale Farm", which were set in established working-class communities. The characters in "Brookside" were generally either people who had advanced themselves from inner-city council estates, or the upper middle-class who had fallen on hard times. Though "Brookside" was still broadcast in a pre-watershed slot (8.00 p.m. and 8.30 p.m. on weekdays, around 5.00 p.m. for the omnibus on Saturdays), it was more liberal than other soaps of the time: the dialogue regularly included expletives. This stemmed from the overall more liberal policy of the channel during that period. The soap was also heavily politicised. Bobby Grant (Ricky Tomlinson), a militant trade-unionist anti-hero, was the most overtly political character. Storylines were often more sensationalist than on other soaps (throughout the soap's history, there were two armed sieges on the street) and were staged with more violence (particularly, rape) often being featured. In 1985, the BBC's "EastEnders" debuted and became a near instant success with viewers and critics alike, with the first episode attracting over 17 million viewers. The Christmas Day 1986 episode was watched by 30.15 million viewers and contained a scene in which divorce papers were served to Angie Watts (Anita Dobson) by her husband, Queen Vic landlord Den (Leslie Grantham). A notable success in pioneering late-night broadcasting, in October 1984, Yorkshire Television began airing the cult Australian soap opera "Prisoner", which originally ran from 1979 to 1986. It was eventually broadcast on all regions of the UK in differing slots, usually around 23:00 (but never before 22:30 in any region), under the title "Prisoner: Cell Block H". It was probably most popular in the Midlands where Central Television consistently broadcast the serial three times a week from 1987 to 1991. Its airing in the UK was staggered, so different regions of the country saw it at a different pace. The program was immensely successful, regularly achieving 10 million viewers when all regions' ratings per episode were added together. Central bowed to fan pressure to repeat the soap, of which the first 95 episodes aired. Then, rival station Channel 5 also acquired rights to repeat the entire rerun of the program, starting in 1997. All 692 episodes have since been released on DVD in the UK. In 1992, the BBC made "Eldorado" to daily alternate with "EastEnders". The programme was heavily criticised and only lasted one year. Nevertheless, soap operas gained increasing prominence on UK television schedules. In 1995, Channel 4 premiered "Hollyoaks", a soap with a youth focus. When Channel 5 launched in March 1997, it debuted the soap opera "Family Affairs", which was formatted as a week-daily soap, airing Monday through Fridays. "Brookside"s premise evolved during the 1990s, phasing out the politicised stories of the 1980s and shifting the emphasis to controversial and sensationalist stories such as child rape, sibling incest, religious cults and drug addiction, including the infamous 'body under the patio' storyline that ran from 1993 to 1995, and gave the serial its highest ratings ever with 9 million viewers. "Coronation Street" and "Brookside" began releasing straight-to-video features. The "Coronation Street" releases generally kept the pace and style of conventional programs episodes with the action set in foreign locations. The "Brookside" releases were set in the usual locations, but featured stories with adult content not allowed on television pre-watershed, with these releases given '18' certificates. "Emmerdale Farm" was renamed "Emmerdale" in 1989. The series was revamped in 1993 with many changes executed via the crash of a passenger jet that partially destroyed the village and killed several characters. This attracted criticism as it was broadcast near the fifth anniversary of the Lockerbie bombing. The storyline drew the soap its highest ever audience of 18 million viewers. The revamp was a success and "Emmerdale" grew in popularity. Throughout the 1990s, "Brookside", "Coronation Street", "EastEnders" and "Emmerdale" continued to flourish. Each increased the number of episodes that aired on a weekly basis by at least one, further defining soap operas as the leading genre in British television. Since 2000, new soap operas have continued to be developed. Daytime drama "Doctors" began in March 2000, preceding "Neighbours" on BBC One. In 2002, as ratings for the Scottish serial "High Road" (formerly "Take The High Road") continued to decline, BBC Scotland launched "River City", which proved popular and effectively replaced "High Road" when it was cancelled in 2003. The long-running serial "Brookside" ended in November 2003 after 21 years on the air, leaving "Hollyoaks" as Channel 4's flagship serial. A new version of "Crossroads" featuring a mostly new cast was produced by Carlton Television for ITV in 2001. It did not achieve high ratings and was cancelled in 2003. In 2001, ITV also launched a new early-evening serial entitled "Night and Day". This program too attracted low viewership and, after being shifted to a late night time slot, was cancelled in 2003. "Family Affairs", which was broadcast opposite the racier "Hollyoaks", never achieved significantly high ratings leading to several dramatic casting revamps and marked changes in style and even location over its run. By 2004, "Family Affairs" had a larger fan base and won its first awards, but was cancelled in late 2005. In 2008, ITV premiered "The Royal Today", a daily spin-off of popular 1960s-based drama "The Royal", which had been running in a primetime slot since 2002. Just days later, soap opera parody programs "Echo Beach" premiered alongside its sister show, the comedy "Moving Wallpaper". Both "Echo Beach" and "The Royal Today" ended after just one series due to low ratings. Radio soap opera "Silver Street" debuted on the BBC Asian Network in 2004. Poor ratings and criticism of the program led to its cancellation in 2010. UK soap operas for many years usually only aired two nights a week. The exception was the original "Crossroads", which began as a week-daily soap opera in the 1960s, but later had its number of weekly broadcasts reduced. In 1989, "Coronation Street" began airing three times a week. In 1996, it expanded to four episodes a week. "Brookside" premiered in 1982 with two episodes a week. In 1990 it expanded to three episodes a week. "EastEnders" increased its number of episode a week in 1994 and "Emmerdale" did so in 1997. "Family Affairs" debuted as a weekdaily soap in 1997, producing five episodes a week its entire run. In 2004, "Emmerdale" began airing six episodes a week. In a January 2008 overhaul of the ITV network, the Sunday episodes of "Coronation Street" and "Emmerdale" were moved out of their slots. "Coronation Street" added a second episode on Friday evenings at 8:30 p.m. "Emmerdale"s Tuesday edition was extended to an hour, putting it in direct competition with "EastEnders". In July 2009, the schedules of these serials were changed again. On 23 July 2009, "Coronation Street" moved from the Wednesday slot it held for 49 years, to Thursday evenings. "Emmerdale" reverted to running just one 30-minute episode on Tuesday evenings and the other 30-minute installment was moved to Thursday evenings. "Coronation Street" later returned to a Wednesday slot, to air Mondays, Wednesdays and Fridays at 19:30 and 20:30. "Emmerdale" airs at 19:00 every weeknight, and 20:00 on Thursdays. Later, "Coronation Street" (which began airing two episodes on Monday nights in 2002) produced five episodes a week. It was announced in June 2016 that starting late 2017, Coronation Street would air six episodes a week. "Doctors" airs five episodes a week, and is the only soap without a weekend omnibus repeat screening. "Hollyoaks" produces five episodes a week. The imported "Neighbours" screens as five new episodes a week. As of 2019, "EastEnders" produces four episodes a week. UK soap operas are shot on videotape in the studio using a multi-camera setup. In their early years, "Coronation Street" and "Emmerdale" used 16 mm film for footage shot on location. Since the 1980s, UK soap opera have routinely featured scenes shot outdoors in each episode. This footage is shot on videotape on a purpose-built outdoor set that represents the community that the soap focuses on. "Hollyoaks" and "Family Affairs" were taped on high-definition video, and used the filmizing process. Australia has had quite a number of well-known soap operas, some of which have gained cult followings in the United Kingdom, New Zealand and other countries. The majority of Australian television soap operas are produced for early evening or evening timeslots. They usually produce two or two-and-a-half hours of new material each week, either arranged as four or five half-hour episodes a week, or as two one-hour episodes. Stylistically, these series most closely resemble UK soap operas in that they are nearly always shot on videotape, are mainly recorded in a studio and use a multi-camera setup. The original Australian serials were shot entirely in-studio. During the 1970s occasional filmed inserts were used to incorporate sequences shot outdoors. Outdoor shooting later became commonplace and starting in the late 1970s, it became standard practice for some on-location footage to be featured in each episode of any Australian soap opera, often to capitalise on the attractiveness and exotic nature of these locations for international audiences. Most Australian soap operas focus on a mixed age range of middle-class characters and will regularly feature a range of locations where the various, disparate characters can meet and interact, such as the café, the surf club, the wine bar or the school. The genre began in Australia on radio, as it had in the United States and the United Kingdom. One such radio serial, "Big Sister", featured actress Thelma Scott in the cast and aired nationally for five years beginning in 1942. Probably the best known Australian radio serial was the long-running soap opera "Blue Hills", which was created by Gwen Meredith and ran from 1949 to 1976. With the advent of Australian television in 1956, daytime television serials followed. The first Australian television soap opera was "Autumn Affair" (1958) featuring radio personality and "Blue Hills" star Queenie Ashton making the transition to television. Each episode of this serial ran for 15 minutes and aired each weekday on the Seven Network. "Autumn Affair" failed to secure a sponsor and ended in 1959 after 156 episodes. It was followed by "The Story of Peter Grey" (1961), another Seven Network weekday series aired in a daytime slot in 15-minute installments. "The Story of Peter Grey" ran for 164 episodes. The first successful wave of Australian evening television soap operas started in 1967 with "Bellbird", produced by the Australian Broadcasting Corporation. This rural-based serial screened in an early evening slot in 15-minute installments as a lead-in to the evening news. "Bellbird" was a moderate success but built-up a consistent and loyal viewer base, especially in rural areas, and enjoyed a ten-year run. "Motel" (1968) was Australia's first half-hour soap opera; the daytime soap had a short run of 132 episodes. The first major soap opera hit in Australia was the sex-melodrama "Number 96", a nighttime series produced by Cash Harmon Television for Network Ten, which debuted March 1972. The program dealt with such topics as homosexuality, adultery, drug use, rape-within-marriage and racism, which had rarely been explored on Australian television programs before. The series became famous for its sex scenes and nudity and for its comedic characters, many of whom became cult heroes in Australia. By 1973, "Number 96" had become Australia's highest-rated show. In 1974, the sexed-up antics of "Number 96" prompted the creation of "The Box", which rivaled it in terms of nudity and sexual situations and was scheduled in a nighttime slot. Produced by Crawford Productions, many critics considered "The Box" to be a more slickly produced and better written show than "Number 96". "The Box" also aired on the Ten Network, programmed to run right after "Number 96". For 1974 "Number 96" was again the highest rating show on Australian television, and that year "The Box" occupied the number two spot. Also in 1974, the Reg Grundy Organisation created its first soap opera, and significantly Australia's first "teen" soap opera, "Class of '74". With its attempts to hint at the sex and sin shown more openly on "Number 96" and "The Box", its high school setting and early evening timeslot, "Class of '74" came under intense scrutiny from the Broadcasting Control Board, who vetted scripts and altered entire storylines. By 1975, both "Number 96" and "The Box", perhaps as a reaction to declining ratings for both shows, de-emphasised the sex and nudity shifting more towards comedic plots. "Class of '74" was renamed "Class of '75" and also added more slapstick comedy for its second year, but the revamped show's ratings declined, resulting in its cancellation in mid-1975. That year Cash Harmon's newly launched second soap "The Unisexers" failed in its early evening slot and was cancelled after three weeks; the Reg Grundy Organisation's second soap "Until Tomorrow" ran in a daytime slot for 180 episodes. A feature film version of "Bellbird" entitled "Country Town" was produced in 1971 by two of the show's stars, Gary Gray and Terry McDermott, without production involvement by the Australian Broadcasting Corporation. "Number 96" and "The Box" also released feature film versions, both of which had the same title as the series, released in 1974 and 1975 respectively. As Australian television had broadcast in black and white until 1975, these theatrical releases all had the novelty of being in colour. The film versions of "Number 96" and "The Box" also allowed more explicit nudity than could be shown on television at that time. In November 1976 "The Young Doctors" debuted on the Nine Network. This Grundy Organization series eschewed the adult drama of "Number 96" and "The Box", focusing more on relationship drama and romance. It became a popular success but received few critical accolades. A week later "The Sullivans", a carefully produced period serial chronicling the effects of World War II on a Melbourne family, also debuted on Nine. Produced by Crawford Productions, "The Sullivans" became a ratings success, attracted many positive reviews, and won television awards. During this period "Number 96" re-introduced nudity into its episodes, with several much-publicised full-frontal nude scenes, a cast revamp and a new range of shock storylines designed to boost the show's declining ratings. "Bellbird" experienced changes to its broadcast pattern with episodes screening in 60 minute blocks, and later in 30 minute installments. "Bellbird", "Number 96" and "The Box", which had been experiencing declining ratings, were cancelled in 1977. Various attempts to revamp each of the shows with cast reshuffles or spectacular disaster storylines had proved only temporarily successful. "The Young Doctors" and "The Sullivans" continued to be popular. November 1977 saw the launch of successful soap opera/police procedural series "Cop Shop" (1977–1984) produced by Crawford Productions for Channel Seven. In early December 1977 Channel Ten debuted the Reg Grundy Organisation produced "The Restless Years" (1977–1981), a more standard soap drama focusing on several young school leavers. The Seven Network, achieving success with "Cop Shop" produced by Crawford Productions, had Crawfords produce "Skyways", a series with a similar format but set in an airport, to compete with the Nine Network's popular talk show "The Don Lane Show". "Skyways", which debuted in July 1979, emphasised adult situations including homosexuality, marriage problems, adultery, prostitution, drug use and smuggling, crime, suicide, political intrigue, and murder, and featured some nudity. Despite this, the program achieved only moderate ratings and was cancelled in mid-1981. The Reg Grundy Organisation found major success with the women's-prison drama "Prisoner" (1979–1986) on Network Ten, and melodramatic family saga "Sons and Daughters" (1982–1987) on the Seven Network. Both shows achieved high ratings in their original runs, and unusually, found success in repeats after the programs ended. Grundy soap "The Young Doctors" and Crawford Productions' "The Sullivans" continued on the Nine Network until late 1982. Thereafter Nine attempted many new replacement soap operas produced by the Reg Grundy Organisation: "Taurus Rising" (1982), "Waterloo Station" (1983), "Starting Out" (1983) and "Possession" (1985), along with "Prime Time" (1986) produced by Crawford Productions. None of these programs were successful and most were cancelled after only a few months. The Reg Grundy Organisation also created "Neighbours", a suburban-based daily serial devised as a sedate family drama with some comedic and lightweight situations, for the Seven Network in 1985. Produced in Melbourne at the studios of HSV-7, "Neighbours" achieved high ratings in Melbourne, Brisbane and Adelaide, but not in Sydney, where it aired at 5.30 p.m. placing it against the hit dating game show "Perfect Match" on Channel 10. The Seven Network's Sydney station ATN-7 quickly lost interest in "Neighbours" as a result of the low ratings in Sydney. HSV-7 in Melbourne lobbied heavily to keep "Neighbours" on the air, but ATN-7 managed to convince the rest of the network to cancel the show and instead keep ATN-7's own Sydney-based dramas "A Country Practice" and "Sons and Daughters". After the network cancelled "Neighbours", it was immediately picked up by Channel Ten, which revamped the cast and scripts slightly and aired the series in the 7.00 p.m. slot starting 20 January 1986. It initially attracted low audiences; however, after a concerted publicity drive, Ten managed to transform the series into a major success, turning several of its actors into major international stars. The show's popularity eventually declined and it was moved to the 6.30 p.m. slot in 1992. In January 2011 it moved to Eleven and is Australia's longest-running soap opera. The success of "Neighbours" in the 1980s prompted the creation of somewhat similar suburban and family or teen-oriented soap operas such as "Home and Away" (1988–present) on Channel Seven and "Richmond Hill" (1988) on Channel Ten. Both proved popular, however "Richmond Hill" emerged as only a moderate success and was cancelled after one year to be replaced on Ten by "E Street" (1989–1993). Nine continued trying to establish a successful new soap opera, without success. After the failure of family drama "Family and Friends" in 1990, it launched the raunchier and more extreme "Chances" in 1991, which resurrected the sex and melodrama of "Number 96" and "The Box" in an attempt to attract attention. "Chances" achieved only moderate ratings, and was moved to a late-night timeslot. It underwent several revamps that removed much of the original cast, and refocused the storylines to incorporate science-fiction and fantasy elements. The series continued in a late night slot until 1992, when it was cancelled due to low ratings despite the much-discussed fantasy storylines. Several Australian soap operas have also found significant international success. In the UK, starting in the mid-1980s, daytime broadcasts of "The Young Doctors", "The Sullivans", "Sons and Daughters" and "Neighbours" (which itself was subsequently moved to an early-evening slot) achieved significant success. Grundy's "Prisoner" began airing in the United States in 1979 and achieved high ratings in many regions there, however the show ended its run in that country three years into its run. "Prisoner" also aired in late-night timeslots in the UK beginning in the late 1980s, achieving enduring cult success there. The show became so popular in that country that it prompted the creation of two stage plays and a stage musical based on the show, all of which toured the UK, among many other spin-offs. In the late 1990s, Channel 5 repeated "Prisoner" in the UK. Between 1998 and 2005, Channel 5 ran late-night repeats of "Sons and Daughters". During the 1980s, the Australian attempts to emulate big-budget U.S. soap operas such as "Dallas" and "Dynasty" had resulted in the debuts of "Taurus Rising" and "Return to Eden", two slick soap opera dramas with big budgets that were shot entirely on film. Though their middling Australian ratings resulted in the shows running only for a single season, both programs were successfully sold internationally. Other shows to achieve varying levels of international success include "Richmond Hill", "E Street", "Paradise Beach" (1993–1994), and "Pacific Drive" (1995–1997). Indeed, these last two series were designed specifically for international distribution. Channel Seven's "Home and Away", a teen soap developed as a rival to "Neighbours", has also achieved significant and enduring success on UK television. "Something in the Air", a serial examining a range of characters in a small country town ran on the ABC from 2000 to 2002. Attempts to replicate the success of daily teen-oriented serials "Neighbours" and "Home and Away" saw the creation of "Echo Point" (1995) and "Breakers" (1999) on Network Ten. These programs foregrounded youthful attractive casts and appealing locations but the programs were not long-running successes and "Neighbours" and "Home and Away" remained the most visible and consistently successful Australian soap operas in production. In their home country, they both attracted respectable although not spectacular ratings in the early 2000s. By 2004, "Neighbours" was regularly attracting just under a million viewers per episode – considered at that time a low figure for Australian prime time television. By March 2007, the Australian audience for "Neighbours" had fallen to fewer than 700,000 a night. This prompted a revamp of the show's cast, its visual presentation, and a move away from the recently added action-oriented emphasis to refocus the show on the domestic storylines it is traditionally known for. During this period "Neighbours" and "Home and Away" continued to achieve significant ratings in the UK. This and other lucrative overseas markets, along with Australian broadcasting laws that enforce a minimum amount of domestic drama production on commercial television networks, help ensure that both programs remain in production. Both shows get higher total ratings in the UK than in Australia (the UK has three times the total population of Australia) and the UK channels make a major contribution to the production costs. It has been suggested that with their emphasis on the younger, attractive and charismatic characters, "Neighbours" and "Home and Away" have found success in the middle ground between glamorous, fantastic U.S. soaps with their wealthy but tragic heroes and the more grim, naturalistic UK soap operas populated by older, unglamorous characters. The casts of "Neighbours" and "Home and Away" are predominantly younger and more attractive than the casts of UK soaps, and without excessive wealth and glamour of the U.S. daytime serial, a middle-ground in which they have found their lucrative niche. "Neighbours" was carried in the United States on the Oxygen cable channel in March 2004; however it attracted few viewers, perhaps in part due to its scheduling opposite well-established and highly popular U.S. soap operas such as "All My Children" and "The Young and the Restless", and was dropped by the network shortly afterwards due to low ratings. "headLand" made its debut on Channel Seven in November 2005, the series arose out of a proposed spinoff of "Home and Away" that was to have been produced in conjunction with "Home and Away"s UK broadcaster, Five. The idea for the spin-off was scuttled after Five pulled out of the deal, which meant that the show could potentially air on a rival channel in the UK; as such, Five requested that the new show be developed as a standalone series and not be spun off from a series that it owned a stake in. The series premiered in Australia on November 15, 2005, but was not a ratings success and was cancelled two months later on January 23, 2006. The series broadcast on E4 and Channel 4 in the UK. Nickelodeon's "" appeared on July 2006 on Network Ten. Since Connie considered this mention as a torrid soap opera, this was mentioned in the Steven Universe episode "Love Letters". After losing the UK television rights to "Neighbours" to Five, the BBC commissioned a replacement serial "Out of the Blue", which was produced in Australia. It debuted as part of BBC One's weekday afternoon schedule on 28 April 2008 but low ratings prompted its move to BBC Two on 19 May 2008. The series was cancelled after its first season. "Neighbours"' continued low ratings in Australia resulted in it being moved to Ten's new digital channel, Eleven on January 11, 2011. However, it continues to achieve reasonable ratings on Channel 5 in the United Kingdom, and as of March 2013 still reportedly achieved significant international sales. Pioneering series "Pukemanu" aired over two years (1971–72) and was the NZBC's first continuing drama. It followed the goings-on of a North Island timber town. "Close to Home" is a New Zealand television soap opera that ran on TVNZ 1 from 1975 to 1983. At its peak in 1977 nearly one million viewers tuned in twice weekly to watch the series co-created by Michael Noonan and Tony Isaac (who had initially only agreed to make the show on the condition they would get to make "The Governor"). "Gloss" is a television drama series that screened from 1987 to 1990. The series is about a fictional publishing empire run by the Redfern family. Gloss was NZ's answer to US soap "Dynasty", with the Carrington oil scions replaced by the wealthy Redferns and their Auckland magazine empire. It was a starting point for many actors who went on to many productions in New Zealand, Australia and around the world including Temuera Morrison, Miranda Harcourt, Peter Elliott, Lisa Chappell, Danielle Cormack and Kevin Smith. Many of them would go on to star in "Shortland Street", which has been New Zealand's most popular soap since its debut in 1992. It airs on TVNZ 2. Radio New Zealand began airing its first radio soap "You Me Now" in September 2010. It is available for podcast on its website. Relatively few daily soap operas have been produced on English Canadian television, with most Canadian stations and networks that carry soap operas airing those imported from the United States or the United Kingdom. Notable daily soaps that did exist included "Family Passions", "Scarlett Hill", "Strange Paradise", "Metropia", "Train 48" and the international co-production "Foreign Affairs". "Family Passions" was an hour-long program, as is typical of American daytime soaps; all of the others ran half-hour episodes. Unlike American or British soap operas, the most influential of which have run for years or even decades, even daily Canadian soap operas have run for a few seasons at most. Short-run soaps, including "49th & Main" and "North/South", have also aired. Many of these were produced in an effort to comply with Canadian content regulations, which require a percentage of programming on Canadian television to originate from Canada. Notable prime time soap operas in Canada have included "Riverdale", "House of Pride", "Paradise Falls", "Lance et Compte" ("He Shoots, He Scores"), "Loving Friends and Perfect Couples" and "The City". The "Degrassi" franchise of youth dramas also incorporated some elements of the soap opera format. On French-language television in Quebec, the "téléroman" has been a popular mainstay of network programming since the 1950s. Notable téléromans have included "Rue des Pignons", "Les Belles Histoires des pays d'en haut", "Diva", "La famille Plouffe", and the soap opera parody "Le Cœur a ses raisons". Unlike the season based production in most countries, most of Indian television fiction tends to be regular-broadcasting soap opera. These started in the 1980s, as more and more people began to purchase television sets. At the beginning of the 21st century, soap operas became an integral part of Indian culture. Indian soap operas mostly concentrate on the conflict between love and arranged marriages occurring in India, and many includes family melodrama. Indian soap operas have multilingual production. Many soap operas produced in India are also broadcast overseas in the UK, Canada, the United States, and some parts of Europe, South Africa, Australia and South East Asia. They are often mass-produced under large production banners, with companies like Balaji Telefilms running different language versions of the same serial on different television networks or channels. The Australian serial "The Restless Years" was remade in the Netherlands as "Goede tijden, slechte tijden" (which debuted in 1990) and in Germany as "Gute Zeiten, schlechte Zeiten" (which has aired since 1992): both titles translate to "good times, bad times". These remakes are still airing, but have long since diverged from the original Australian storylines. The two shows are the highest-rated soap operas in their respective countries. A later Australian serial, "Sons and Daughters", has inspired five remakes produced under license from the original producers and based, initially, on original story and character outlines. These are "Verbotene Liebe" (Germany, 1995–2015); "Skilda världar" (Sweden, 1996–2002); "Apagorevmeni agapi" (Greece, 1998); "Cuori Rubati" (Italy, 2002–2003) and "Zabranjena ljubav" (Croatia, 2004–2008). Both "The Restless Years" and "Sons and Daughters" were created and produced in Australia by the Reg Grundy Organisation. Another Australian soap opera reformatted for a European audience was "E Street" which ran on Network Ten in Australia from 1989 to 1993. Germany produced 37 episodes of "Westerdeich" ("Westside") in 1995 using scripts from 1989 episodes of "E Street". It was also remade in Belgium as "Wittekerke" ("Whitechurch") and ran from 1993 to 2008. The Norwegian soap opera "Hotel Cæsar" aired on TV 2 from 1998 to 2017, and is the longest-running television drama in Scandinavia. Popular foreign soaps in the country include "Days of Our Lives" (broadcast on TV6 (Norway), "The Bold and the Beautiful" (TNT (Norway) and "Home and Away" (TV 2), all of which are subtitled. Serials have included "Goede tijden, slechte tijden" (1990–present), "Onderweg naar Morgen" (1994–2010) and "Goudkust" (1996–2001). In 2016 "Goede tijden, slechte tijden" spin-off "Nieuwe Tijden" started airing. U.S. daytime serials "As The World Turns" and "The Bold and the Beautiful" have aired in the Netherlands; "As the World Turns" began airing in the country in 1990, with Dutch subtitles. In the 1980s, West German networks successfully added American daytime and primetime soap operas to their schedule before Das Erste introduced its first self-produced weekly soap with "Lindenstraße", which was seen as a German counterpart to "Coronation Street". Like in other countries, the soap opera met with negative reviews, but eventually proved critics wrong with nearly 13 million viewers tuning in each week. Even though the format proved successful, it was not until 1992 before "Gute Zeiten, schlechte Zeiten" became the first German daily soap opera. Early ratings were bad as were the reviews, but the RTL network was willing to give its first soap opera a chance; ratings would improve, climbing to 7 million viewers by 2002. Not long after "Gute Zeiten, schlechte Zeiten", Das Erste introduced "Marienhof", which aired twice a week. After successfully creating the first German daily soap, production company Grundy Ufa wanted to produce another soap for RTL. Like "GZSZ", the format was based on an Australian soap opera from Reg Watson. But RTL did not like the plot idea about separated twins who meet each other for the first time after 20 years and fall in love without knowing that they are related. The project was then taken to Das Erste, which commissioned the program, titled "Verbotene Liebe", which premiered on January 2, 1995. With the premiere of "Verbotene Liebe", the network turned "Marienhof" into a daily soap as well. In the meanwhile, RTL debuted the Grundy Ufa-produced "Unter uns" in late 1994. ZDF started a business venture with Canada and co-produced the short-lived series "Family Passions", starring actors such as Gordon Thomson, Roscoe Born, Dietmar Schönherr and a young Hayden Christensen. The daytime serial premiered on December 5, 1994, lasting 130 episodes. After its cancellation, the network debuted "Jede Menge Leben". Even after a crossover with three soaps, "Freunde fürs Leben", "Forsthaus Falkenau" and "Unser Lehrer Doktor Specht", the soap was canceled after 313 episodes. Sat.1 tried to get into the soap business as well, after successfully airing the Australian soap opera "Neighbours", which was dropped in 1995 due to the talk show phenomenon that took over most of the daytime schedules of German networks. The network first tried to tell a family saga with "So ist das Leben! Die Wagenfelds", before failing with "Geliebte Schwestern". RTL II made its own short-lived attempt with "Alle zusammen – jeder für sich". The teen soap opera "Schloss Einstein" debuted on September 4, 1998, focusing on the life of a group of teenagers at the fictional titular boarding school near Berlin. As of July 2014, the series has produced over 815 episodes during the course of 17 seasons, a milestone in German television programming, and was renewed for an 18th season to debut in 2015. In 1999, after the lasting success of "Gute Zeiten, schlechte Zeiten", "Marienhof", "Unter uns" and "Verbotene Liebe", ProSieben aired "Mallorca – Suche nach dem Paradies", set on the Spanish island with the same name. After nine months, the network canceled the program due to low viewership and high production costs. Even though ratings had improved, the show ended its run in a morning timeslot. The soap opera became something of a cult classic, as its 200-episode run was repeated several times on free-to-air and pay television. In 2006, "Alles was zählt" became the last successful daily soap to make its debut, airing as a lead-in to "Gute Zeiten, schlechte Zeiten" and also produced by Grundy Ufa. Since Germany started to produce its own telenovelas, all soap operas faced declines in ratings. "Unter uns" was in danger of cancellation in 2009, but escaped such a fate due to budget cuts imposed by the show's producers and the firing of original cast member Holger Franke, whose firing and the death of his character outraged fans, resulting in a ratings spike in early 2010. After "Unter uns" was saved, Das Erste planned to make changes to its soap lineup. "Marienhof" had to deal with multiple issues in its storytelling, as well as in producing a successful half-hour show. Several changes were made within months, however "Marienhof" was canceled in June 2011. "Verbotene Liebe" was in danger of being cancelled as well, but convinced the network to renew it with changes that it made in both 2010 and 2011; the soap was later expanded to 45 minutes after "Marienhof" was canceled, and the network tried to decide on whether to revamp its lineup. While "Gute Zeiten, schlechte Zeiten", "Unter uns" and "Alles was zählt" are currently the only daily soaps on the air after "Verbotene Liebe" has been cancelled and aired its last episode in June, 2015 due to low ratings, the telenovelas "Sturm der Liebe" and "Rote Rosen" are considered soaps by the press as well, thanks to the changing protagonists every season. In Belgium, the two major soap operas are "Thuis" ("Home") and "Familie" ("Family"). Soap operas have been very popular in Flanders, the Dutch-speaking part of Belgium. "Familie" debuted in late 1991, and with nearly 6,000 half-hour episodes, it has the highest episode total of any soap in Europe outside of the United Kingdom. The highest-rated soap opera is "Thuis", which has aired on "één" since late 1995. "Thuis" is often one of the five most-watched Belgian shows and regularly garners over one million viewers (with 6.3 million Flemmings in total). During the 1990s, foreign soap operas such as "Neighbours" and "The Bold and the Beautiful" were extremely popular, the latter having achieved a cult status in Belgium and airing in the middle of the decade during prime time. Both soaps still air today, along with other foreign soaps such as "Days of Our Lives", Australia's "Home and Away" and Germany's "Sturm der Liebe". Vitaya unsuccessful attempted to air the Dutch soap opera "Goede Tijden, Slechte Tijden" in 2010. Other foreign soaps that previously aired on Belgian television include "The Young and the Restless", "EastEnders" (both on VTM), "Port Charles" (at één, then known as TV1) and "Coronation Street" (on Vitaya). "Santa Barbara" aired during the 1990s on VTM for its entire run. The only teen soap opera on Belgian television was "Spring" ("Jump" in English), which aired on the youth-oriented Ketnet and produced over 600 15-minute episodes from late 2002 until 2009, when it was cancelled after a steady decline in ratings following the departures of many of its original characters. The most successful soap opera in Italy is the evening series "Un posto al sole" ("A Place Under the Sun"), which had aired on Rai 3 since 1996 (whose format is based on the Australian soap opera "Neighbours"). Several other Italian soaps have been produced such as "Ricominciare" ("Starting Over"), "Cuori rubati" ("Stolen Hearts"), "Vivere" ("Living"), "Sottocasa" ("Downstairs"), "Agrodolce" ("Bittersweet") and "Centovetrine" ("Hundred Shop Windows"). The most popular Italian prime-time soap opera, "Incantesimo" ("Enchantment"), which ran from 1998 to 2008, became a daytime soap opera for the final two years of its run, airing five days a week on Rai 1. The same happened with "Il paradiso delle signore" ("Woman’s Paradise"), a period drama, which ran from 2015 to 2017 in prime-time, and became a daytime period soap opera from 2018. In the early years of RTÉ, the network produced several dramas but had not come close to launching a long-running serial. RTÉ's first television soap was "Tolka Row", which was set in urban Dublin. For several years, both "Tolka Row" and "The Riordans" were produced by RTÉ; however, the urban soap was soon dropped in favor of the more popular rural soap opera "The Riordans", which premiered in 1965. Executives from Yorkshire Television visited during on-location shoots for "The Riordans" in the early 1970s and in 1972, debuted "Emmerdale Farm" (now "Emmerdale"), based on the successful format of the Irish soap opera. In the late 1970s, "The Riordans" was controversially dropped. The creator of that series would then go on to produce the second of his "Agri-soap" trilogy "Bracken", starring Gabriel Byrne, whose character had appeared in the last few seasons of "The Riordans". Bracken was soon replaced by the third "Agri-soap" "Glenroe", which ran until 2001. As RTÉ wanted a drama series for its Sunday night lineup rather than a soap opera, "On Home Ground" (2001–2002), "The Clinic" (2002–2009) and "RAW" (2010–2013) replaced the agri-soaps of the previous decades. In 1989, RTÉ decided to produce its first Dublin-based soap opera since the 1960s. "Fair City", which is set in the fictional city of Carrickstown, initially aired one night a week during the 1989-90 season, and similar to its rural soaps, much of the footage was filmed on location – in a suburb of Dublin City. In 1992, RTÉ made a major investment into the series by copying the houses used in the on-location shoots for an on-site set in RTÉ's Headquarters in Dublin 4. By the early 1990s, it was airing two nights a week for 35 weeks a year. With competition from the UK soap operas, RTÉ expanded "Fair City" to three nights a week for most of the year and one night a week during the summer in 1996, later expanding to four nights a week and two nights during the summer. Until the early 2000s, the series produced four episodes a week, airing all 52 weeks of the year. "Fair City" airs Sundays, Tuesdays and Thursdays at 8.00 p.m. GMT on RTÉ One; however, after rival network TV3 moved "Coronation Street" to Thursday night, the Wednesday night episode of "Fair City" began airing at 7:30 p.m. each week. TG4 produce the Irish language soap "Ros na Rún" ("Headland of the Secrets" or "Headland of the Sweethearts"); set in the fictional village of "Ros Na Rún", located outside Galway and near Spiddal, it centres on the domestic and professional lives of the town's residents. It is modeled on an average village in the West of Ireland, but with its own distinct personality – with a diverse population that share secrets, romances and friendships among other things. While the core community has remained the same, the look and feel of "Ros Na Rún" has changed and evolved over the years to incorporate the changing face of rural Ireland. It has an established a place not only in the hearts and minds of the Irish speaking public, but also the wider Irish audience. The program has dealt with many topics, including domestic violence, infidelity, theft, arson, abortion, homosexuality, adoption, murder, rape, drugs, teen pregnancy and paedophilia. It runs twice a week for 35 weeks of the year, currently airing Tuesday and Thursday nights. "Ros na Rún" is the single largest independent production commissioned in the history of Irish broadcasting. Prior to TG4's launch, it originally aired on RTÉ One in the early 1990s. Although Ireland has access to international soaps (such as "Coronation Street", "Emmerdale", "EastEnders", "Home and Away", "Hollyoaks" and "Neighbours"), "Fair City" continues to outperform them all, and is Ireland's most popular soap opera, with the show peaking at over 700,000 viewers. January 2015 "Red Rock" has broadcast on TV3. Red Rock airs twice a week on Wednesday and Thursday nights. The series is base in a fishing village in Dublin. The soaps centres around the local Garda station but also includes stories from the village. RTÉ Radio produced its first radio soap, "Kennedys of Castleross", which ran from April 13, 1955 to 1975. In 1979 RTÉ long running TV soap The Riordans moved to Radio until December 24, 1985. In the mid-1980s, RTÉ debuted a new radio soap, "Harbour Hotel", which ran until the mid-1990s. The network later ran two short-lived radio soaps, "Konvenience Korner" and "Riverrun", which were followed in 2004 by "Driftwood". RTÉ does not run any radio soaps, however RTÉ Radio 1 continues to air radio dramas as part of its nighttime schedule. In Greece, there have been several soap operas. An early serial was "Sti skia tou hrimatos" ("Money Shadows"), which ran from 1990 to 1991. September 1991 saw the debut of "Lampsi" ("the Shining"), from creator Nicos Foskolos. The series would become Greece's longest-running soap opera. After the success of "Lampsi" came the short lived "To galazio diamandi" ("Blue Diamond") and "Simphonia siopis" ("Omertà"). "Lampsi" was canceled in June 2005 due to declining ratings. It was replaced by "Erotas" ("Love"), a soap that ran from 2005 to 2008. After that series ended, ANT1 abandoned the soap opera genre and focused on comedy series and weekly dramas. Greece's second longest-running soap is "Kalimera Zoi" ("Goodmorning Life"), which ran from September 1993 until its cancellation in June 2006 due to low ratings. Mega Channel began producing soap operas in 1990 with the prime time serial "I Dipsa" ("The Thirst"), which ran for 102 episodes. Other daytime soaps have included "Paralliloi dromoi" (1992–1994) and its successor "Haravgi" ("Daylight", 1994–1995), both of which were cancelled due to low viewership; as well as the serials "Apagorevmeni Agapi" ("Forbidden Love"), which ran from 1998 to 2006; "Gia mia thesi ston Ilio" ("A Spot Under the Sun"), which ran from 1998 to 2002; "Filodoxies" ("Expectations"), which ran from 2002 to 2006; and "Vera Sto Deksi" ("Ring on the Right Hand"), which ran from 2004 to 2006 and proved to be a successful competitor to "Lampsi", causing that show's ratings to decline. "Ta Mistika Tis Edem" ("Edem Secrets"), which was created by the producers of "Vera Sto Deksi", debuted in 2008 and has eclipsed that show's success. Its ratings place it consistently among the three highest-rated daytime programs. IENED (which was renamed ERT2 in 1982) was responsible for the first Greek soap operas "I Kravgi Ton Likon" and "Megistanes". ERT also produced the long-running soap "O Simvoleografos". Since 2000 and with the introduction of private television, ERT produced additional daily soap operas, which included "Pathos" ("Passion"), "Erotika tis Edem" ("Loving in Eden") and "Ta ftera tou erota" ("The Wings of Love"). These failed to achieve high ratings and were canceled shortly after their premiere. Alpha produced "Kato apo tin Acropoli" ("Under the Acropolis"), which ran for 2½ years. The first daytime soap opera produced by a Cyprus channel was LOGOs TV's "Odos Den Ksehno" ("'Don't Forget' Street"), which ran from January to December 1996. It was followed by "To Serial", which also ran for one year from September 1997 to June 1998. CyBC created the third weekdaily soap, "Anemi Tou Pathous" ("Passion Winds"), running from January 2000 to June 2004, which was replaced by "I Platia" ("The Square") from September 2004 to July 2006. "Epikindini Zoni" ran from 2009 to 2010, and was cancelled after 120 episodes. "Vimata Stin Ammo" made its debut in September 2010. Sigma TV first commissioned the weekdaily comedic soap "Sto Para Pente", which aired from September 1998 to June 2004, and was the longest weekday show in Cyprus television history, before it was surpassed by "Se Fonto Kokkino", which ran from September 2008 to July 2012. Other Sigma TV weekday shows include "Akti Oniron" (which ran from 1999 to 2001), "Vourate Geitonoi" (which ran from 2001 to 2005, and was the most successful weekdaily series, achieving ratings shares of up to 70% of all television households in the country), "Oi Takkoi" (which ran from 2002 to 2005), "S' Agapo" (which ran from 2001 to 2002), "Vasiliki" (which ran from 2005 to 2006), "Vendetta" (which ran from September 2005 to December 2006), "30 kai Kati" (which ran from 2006 to 2007) and "Mila Mou" (which ran from September 2007 to January 2009). ANT1 Cyprus aired the soap "I Goitia Tis Amartias" in 2002, which was soon canceled. "Dikse Mou To Filo Sou" followed from 2006 to 2009, along with "Gia Tin Agapi Sou", which ran from 2008 to 2009 and itself was followed by "Panselinos", which has aired since 2009. The longest-running weekly show on Cyprus television is "Istories Tou Horkou" ("Villages Stories", which premiered on CyBC in March 1996 and ran until its cancellation in June 2006; it was revived in September 2010 but was cancelled again in March 2011 due to very low ratings), followed by "Manolis Ke Katina" ("Manolis and Katina", which ran from 1995 to 2004). The most controversial of these series was "To Kafenio" ("The Coffee Shop"), which premiered on CyBC on 1993 as a weekly series, before moving to MEGA Channel Cyprus six years later in 1999 as a weekdaily show and then moved to ANT1 Cyprus on 2000, which canceled the show one year later. There were plans to move the show back to CyBC as a weekly series in 2001, with the original cast, however this plan was never realised. The most successful weekly shows in Cyprus currently are ANT1's "Eleni I Porni" ("Eleni, The Whore"), which premiered in October 2010 and CyBC's "Stin Akri Tu Paradisou" ("At The Heaven's Edge"), which premiered in 2007. The most successful weekdaily soap was "Aigia Fuxia", which aired on ANT1 Cyprus from 2008 to 2010. The only daily Finnish soap opera so far is "Salatut elämät" ("Secret Lives"), which has achieved popularity in Finland since its 1999 debut on MTV3. It focuses on the lives of people along the imaginary Pihlajakatu street in Helsinki. The show has also spawned several Internet spin-off series and a film based on the show that was released in 2012. Other soap-like shows in Finland are YLE shows "Uusi päivä" (which has aired since 2009) and "Kotikatu" (which ran from 1995 to 2012), however these programs do not adhere to a five-episode-a-week schedule. With the advent of internet television and mobile phones, several soap operas have also been produced specifically for these platforms, including "", a spin-off of the established "EastEnders". For those produced only for the mobile phone, episodes may generally consist of about six or seven pictures and accompanying text. On September 13, 2011, TG4 launched a new 10-part online series titled, "Na Rúin" (an Internet spin-off of "Ros na Rún"). The miniseries took on the theme of a mystery; the viewer had to read Rachel and Lorcán's blogs as well as watch video diaries detailing each character's thoughts to solve the mystery of missing teenage character Ciara. In motion pictures, the 1982 comedy "Tootsie" has the lead character impersonating a woman in order to gain acting work on a long running television soap opera. Several scenes parody the production of soaps, their outrageous storylines and idiosyncratic stylistic elements. The 1991 comedy "Soapdish" stars Sally Field as an aging soap opera actress on the fictional series "The Sun Also Sets" who pines over her own neuroses and misfortunes, such as her live-in boyfriend who leaves her to go back to his wife, and the incidents of backstabbing and scheming behind the scenes, some of which are more interesting than the stories on the program. Another 1991 comedy, "Delirious", stars John Candy as a soap opera writer who, after a head injury, has a dream experience of being in his own creation. The dream experience is an increasingly outrageous exaggeration of soap opera plot elements. On television, several soap opera parodies have been produced:
https://en.wikipedia.org/wiki?curid=27009