text
stringlengths
10
951k
source
stringlengths
39
44
Athens Athens ( ; ; (pl.) ) is the capital and largest city in Greece. Athens dominates the Attica region and is one of the world's oldest cities, with its recorded history spanning over 3,400 years and its earliest human presence started somewhere between the 11th and 7th millennium BC. Classical Athens was a powerful city-state that emerged in conjunction with the seagoing development of the port of Piraeus. A center for the arts, learning and philosophy, home of Plato's Academy and Aristotle's Lyceum, it is widely referred to as the cradle of Western civilization and the birthplace of democracy, largely because of its cultural and political impact on the European continent, and in particular the Romans. In modern times, Athens is a large cosmopolitan metropolis and central to economic, financial, industrial, maritime, political and cultural life in Greece. Athens is a global city and one of the biggest economic centers in southeastern Europe. It has a large financial sector, and its port Piraeus is both the largest passenger port in Europe, and the second largest in the world. The Municipality of Athens (also City of Athens), which actually constitutes a small administrative unit of the entire city, had a population of 664,046 (in 2011) within its official limits, and a land area of . The Athens Urban Area (Greater Athens and Greater Piraeus) extends beyond its administrative municipal city limits, with a population of 3,090,508 (in 2011) over an area of . According to Eurostat in 2011, the functional urban area (FUA) of Athens was the 9th most populous FUA in the European Union (the 6th most populous capital city of the EU), with a population of 3.8 million people. Athens is also the southernmost capital on the European mainland and the warmest major city in Europe. The heritage of the classical era is still evident in the city, represented by ancient monuments and works of art, the most famous of all being the Parthenon, considered a key landmark of early Western civilization. The city also retains Roman and Byzantine monuments, as well as a smaller number of Ottoman monuments. Athens is home to two UNESCO World Heritage Sites, the Acropolis of Athens and the medieval Daphni Monastery. Landmarks of the modern era, dating back to the establishment of Athens as the capital of the independent Greek state in 1834, include the Hellenic Parliament and the so-called "architectural trilogy of Athens", consisting of the National Library of Greece, the National and Kapodistrian University of Athens and the Academy of Athens. Athens is also home to several museums and cultural institutions, such as the National Archeological Museum, featuring the world's largest collection of ancient Greek antiquities, the Acropolis Museum, the Museum of Cycladic Art, the Benaki Museum and the Byzantine and Christian Museum. Athens was the host city of the first modern-day Olympic Games in 1896, and 108 years later it hosted the 2004 Summer Olympics, making it one of the few cities to have hosted the Olympics more than once. In Ancient Greek, the name of the city was ("Athênai", in Classical Attic) a plural. In earlier Greek, such as Homeric Greek, the name had been current in the singular form though, as ("Athḗnē"). It was possibly rendered in the plural later on, like those of ("Thêbai") and ("Μukênai"). The root of the word is probably not of Greek or Indo-European origin, and is possibly a remnant of the Pre-Greek substrate of Attica. In antiquity, it was debated whether Athens took its name from its patron goddess Athena (Attic , "Athēnâ", Ionic , "Athḗnē", and Doric , "Athā́nā") or Athena took her name from the city. Modern scholars now generally agree that the goddess takes her name from the city, because the ending -"ene" is common in names of locations, but rare for personal names. During the medieval period, the name of the city was rendered once again in the singular as . However, after the establishment of the modern Greek state, and partly due to the conservatism of the written language, became again the official name of the city and remained so until the abandonment of Katharevousa in the 1970s, when Ἀθήνα, "Athína", became the official name. According to the ancient Athenian founding myth, Athena, the goddess of wisdom, competed against Poseidon, the God of the Seas, for patronage of the yet-unnamed city; they agreed that whoever gave the Athenians the better gift would become their patron and appointed Cecrops, the king of Athens, as the judge. According to the account given by Pseudo-Apollodorus, Poseidon struck the ground with his trident and a salt water spring welled up. In an alternative version of the myth from Vergil's "Georgics", Poseidon instead gave the Athenians the first horse. In both versions, Athena offered the Athenians the first domesticated olive tree. Cecrops accepted this gift and declared Athena the patron goddess of Athens. Eight different etymologies, now commonly rejected, were proposed during the 17th century. Christian Lobeck proposed as the root of the name the word ("áthos") or ("ánthos") meaning "flower", to denote Athens as the "flowering city". Ludwig von Döderlein proposed the stem of the verb , stem θη- ("tháō", "thē-", "to suck") to denote Athens as having fertile soil. In classical literature, the city was sometimes referred to as the City of the Violet Crown, first documented in Pindar's ἰοστέφανοι Ἀθᾶναι ("iostéphanoi Athânai"), or as ("tò kleinòn ásty", "the glorious city"). In medieval texts, variant names include Setines, Satine, and Astines, all derivations involving false splitting of prepositional phrases. Today the caption ("ī protévousa"), "the capital", has become somewhat common. King Alphonse X of Castile in his 'General Storia' (Global History) cites etymology of Athens as: 'Athinai', 'the one without death', death here meaning ignorance. Athenians were called cicada-wearers () because they used to wear pins of golden cicadas. A symbol of being autochthon (earth-born), because the legendary founder of Athens, Erechtheus was an autochthon or of being musicians, because the cicada is a "musician" insect. The oldest known human presence in Athens is the Cave of Schist, which has been dated to between the 11th and 7th millennia BC. Athens has been continuously inhabited for at least 5,000 years. By 1400 BC the settlement had become an important centre of the Mycenaean civilization and the Acropolis was the site of a major Mycenaean fortress, whose remains can be recognised from sections of the characteristic Cyclopean walls. Unlike other Mycenaean centers, such as Mycenae and Pylos, it is not known whether Athens suffered destruction in about 1200 BC, an event often attributed to a Dorian invasion, and the Athenians always maintained that they were pure Ionians with no Dorian element. However, Athens, like many other Bronze Age settlements, went into economic decline for around 150 years afterwards. Iron Age burials, in the Kerameikos and other locations, are often richly provided for and demonstrate that from 900 BC onwards Athens was one of the leading centres of trade and prosperity in the region. The leading position of Athens may well have resulted from its central location in the Greek world, its secure stronghold on the Acropolis and its access to the sea, which gave it a natural advantage over inland rivals such as Thebes and Sparta. By the 6th century BC, widespread social unrest led to the reforms of Solon. These would pave the way for the eventual introduction of democracy by Cleisthenes in 508 BC. Athens had by this time become a significant naval power with a large fleet, and helped the rebellion of the Ionian cities against Persian rule. In the ensuing Greco-Persian Wars Athens, together with Sparta, led the coalition of Greek states that would eventually repel the Persians, defeating them decisively at Marathon in 490 BC, and crucially at Salamis in 480 BC. However, this did not prevent Athens from being captured and sacked twice by the Persians within one year, after a heroic but ultimately failed resistance at Thermopylae by Spartans and other Greeks led by King Leonidas, after both Boeotia and Attica fell to the Persians. The decades that followed became known as the Golden Age of Athenian democracy, during which time Athens became the leading city of Ancient Greece, with its cultural achievements laying the foundations for Western civilization. The playwrights Aeschylus, Sophocles and Euripides flourished in Athens during this time, as did the historians Herodotus and Thucydides, the physician Hippocrates, and the philosopher Socrates. Guided by Pericles, who promoted the arts and fostered democracy, Athens embarked on an ambitious building program that saw the construction of the Acropolis of Athens (including the Parthenon), as well as empire-building via the Delian League. Originally intended as an association of Greek city-states to continue the fight against the Persians, the league soon turned into a vehicle for Athens's own imperial ambitions. The resulting tensions brought about the Peloponnesian War (431–404 BC), in which Athens was defeated by its rival Sparta. By the mid-4th century BC, the northern Greek kingdom of Macedon was becoming dominant in Athenian affairs. In 338 BC the armies of Philip II defeated an alliance of some of the Greek city-states including Athens and Thebes at the Battle of Chaeronea, effectively ending Athenian independence. Later, under Rome, Athens was given the status of a free city because of its widely admired schools. The Roman emperor Hadrian, in the 2nd century CE, ordered the construction of a library, a gymnasium, an aqueduct which is still in use, several temples and sanctuaries, a bridge and financed the completion of the Temple of Olympian Zeus. By the end of Late Antiquity, Athens had shrunk due to sacks by the Herulians, Visigoths, and Early Slavs which caused massive destruction in the city. In this era, the first Christian churches were built in Athens, and the Parthenon and other temples were converted into churches. Athens expanded its settlement in the second half of the Middle Byzantine Period, in the 9th to 10th centuries CE, and was relatively prosperous during the Crusades, benefiting from Italian trade. After the Fourth Crusade the Duchy of Athens was established. In 1458 it was conquered by the Ottoman Empire and entered a long period of decline. Following the Greek War of Independence and the establishment of the Greek Kingdom, Athens was chosen as the capital of the newly independent Greek state in 1834, largely because of historical and sentimental reasons. At the time, it was reduced to a town of about 4,000 people in a loose swarm of houses along the foot of the Acropolis. The first King of Greece, Otto of Bavaria, commissioned the architects Stamatios Kleanthis and Eduard Schaubert to design a modern city plan fit for the capital of a state. The first modern city plan consisted of a triangle defined by the Acropolis, the ancient cemetery of Kerameikos and the new palace of the Bavarian king (now housing the Greek Parliament), so as to highlight the continuity between modern and ancient Athens. Neoclassicism, the international style of this epoch, was the architectural style through which Bavarian, French and Greek architects such as Hansen, Klenze, Boulanger or Kaftantzoglou designed the first important public buildings of the new capital. In 1896, Athens hosted the first modern Olympic Games. During the 1920s a number of Greek refugees, expelled from Asia Minor after the Greco-Turkish War, swelled Athens's population; nevertheless it was most particularly following World War II, and from the 1950s and 1960s, that the population of the city exploded, and Athens experienced a gradual expansion. In the 1980s it became evident that smog from factories and an ever-increasing fleet of automobiles, as well as a lack of adequate free space due to congestion, had evolved into the city's most important challenge. A series of anti-pollution measures taken by the city's authorities in the 1990s, combined with a substantial improvement of the city's infrastructure (including the Attiki Odos motorway, the expansion of the Athens Metro, and the new Athens International Airport), considerably alleviated pollution and transformed Athens into a much more functional city. In 2004 Athens hosted the 2004 Summer Olympics. Athens sprawls across the central plain of Attica that is often referred to as the Athens Basin or the Attica Basin (). The basin is bounded by four large mountains: Mount Aigaleo to the west, Mount Parnitha to the north, Mount Pentelicus to the northeast and Mount Hymettus to the east. Beyond Mount Aegaleo lies the Thriasian plain, which forms an extension of the central plain to the west. The Saronic Gulf lies to the southwest. Mount Parnitha is the tallest of the four mountains (), and has been declared a national park. Athens is built around a number of hills. Lycabettus is one of the tallest hills of the city proper and provides a view of the entire Attica Basin. The meteorology of Athens is deemed to be one of the most complex in the world because its mountains cause a temperature inversion phenomenon which, along with the Greek Government's difficulties controlling industrial pollution, was responsible for the air pollution problems the city has faced. This issue is not unique to Athens; for instance, Los Angeles and Mexico City also suffer from similar atmospheric inversion problems. The Cephissus river, the Ilisos and the Eridanos stream are the historical rivers of Athens. By the late 1970s, the pollution of Athens had become so destructive that according to the then Greek Minister of Culture, Constantine Trypanis, "...the carved details on the five the caryatids of the Erechtheum had seriously degenerated, while the face of the horseman on the Parthenon's west side was all but obliterated." A series of measures taken by the authorities of the city throughout the 1990s resulted in the improvement of air quality; the appearance of smog (or "nefos" as the Athenians used to call it) has become less common. Measures taken by the Greek authorities throughout the 1990s have improved the quality of air over the Attica Basin. Nevertheless, air pollution still remains an issue for Athens, particularly during the hottest summer days. In late June 2007, the Attica region experienced a number of brush fires, including a blaze that burned a significant portion of a large forested national park in Mount Parnitha, considered critical to maintaining a better air quality in Athens all year round. Damage to the park has led to worries over a stalling in the improvement of air quality in the city. The major waste management efforts undertaken in the last decade (particularly the plant built on the small island of Psytalia) have greatly improved water quality in the Saronic Gulf, and the coastal waters of Athens are now accessible again to swimmers. In January 2007, Athens faced a waste management problem when its landfill near Ano Liosia, an Athenian suburb, reached capacity. The crisis eased by mid-January when authorities began taking the garbage to a temporary landfill. Athens ranks in the lowest percentage for the risk on frequency and severity of terrorist attacks according to the EU Global Terrorism Database (EIU 2007-2016 calculations). The city also ranked 35th in Digital Security, 21st on Health Security, 29th on Infrastructure Security and 41st on Personal Security globally in a 2017 The Economist Intelligence Unit report. It also ranks as a very safe city (39th globally out of 162 cities overall) on the ranking of the safest and most dangerous countries. A 2019 crime index from Numbeo places Athens at 130th position, rating safer than Tampa, Florida or Dublin, Ireland. According to a Mercer 2019 Quality of Living Survey, Athens ranks 89th on the Mercer Quality of Living Survey ranking. Athens has a hot-summer Mediterranean climate (Köppen climate classification: "Csa"). The dominant feature of Athens' climate is alternation between prolonged hot and dry summers and mild winters with moderate rainfall. With an average of of yearly precipitation, rainfall occurs largely between the months of October and April. July and August are the driest months when thunderstorms occur sparsely. Furthermore, some coastal areas of Athens, known as the Athens Riviera have a hot semi-arid climate ("BSh") according to the climate atlas published by the Hellenic National Meteorological Service (H.N.M.S). Owing to the rain shadow of the Pindus Mountains, annual precipitation of Athens is lower than most other parts of Greece, especially western Greece. As an example, Ioannina receives around per year, and Agrinio around per year. Daily average highs for July have been measured around in downtown Athens, but some parts of the city may be even hotter for the higher density of buildings, such as the center, in particular, western areas due to a combination of industrialization and a number of natural factors, knowledge of which has existed since the mid-19th century. Athens is affected by the urban heat island effect in some areas which is caused by human activity, altering its temperatures compared to the surrounding rural areas, and leaving detrimental effects on energy usage, expenditure for cooling, and health. The urban heat island of the city has also been found to be partially responsible for alterations of the climatological temperature time-series of specific Athens meteorological stations, because of its impact on the temperatures and the temperature trends recorded by some meteorological stations. On the other hand, specific meteorological stations, such as the National Garden station and Thiseio meteorological station, are less affected or do not experience the urban heat island. Athens holds the World Meteorological Organization record for the highest temperature ever recorded in Europe, at , which was recorded in the Elefsina and Tatoi suburbs of Athens on 10 July 1977. The Municipality of Athens, the City Centre of the Athens Urban Area, is divided into several districts: Omonoia, Syntagma, Exarcheia, Agios Nikolaos, Neapolis, Lykavittos, Lofos Strefi, Lofos Finopoulou, Lofos Filopappou, Pedion Areos, Metaxourgeio, Aghios Kostantinos, Larissa Station, Kerameikos, Psiri, Monastiraki, Gazi, Thission, Kapnikarea, Aghia Irini, Aerides, Anafiotika, Plaka, Acropolis, Pnyka, Makrygianni, Lofos Ardittou, Zappeion, Aghios Spyridon, Pangrati, Kolonaki, Dexameni, Evaggelismos, Gouva, Aghios Ioannis, Neos Kosmos, Koukaki, Kynosargous, Fix, Ano Petralona, Kato Petralona, Rouf, Votanikos, Profitis Daniil, Akadimia Platonos, Kolonos, Kolokynthou, Attikis Square, Lofos Skouze, Sepolia, Kypseli, Aghios Meletios, Nea Kypseli, Gyzi, Polygono, Ampelokipoi, Panormou-Gerokomeio, Pentagono, Ellinorosson, Nea Filothei, Ano Kypseli, Tourkovounia-Lofos Patatsou, Lofos Elikonos, Koliatsou, Thymarakia, Kato Patisia, Treis Gefyres, Aghios Eleftherios, Ano Patisia, Kypriadou, Menidi, Prompona, Aghios Panteleimonas, Pangrati, Goudi, Vyronas and Ilisia. Parnitha National Park is punctuated by well-marked paths, gorges, springs, torrents and caves dotting the protected area. Hiking and mountain-biking in all four mountains are popular outdoor activities for residents of the city. The National Garden of Athens was completed in 1840 and is a green refuge of 15.5 hectares in the centre of the Greek capital. It is to be found between the Parliament and Zappeion buildings, the latter of which maintains its own garden of seven hectares. Parts of the City Centre have been redeveloped under a masterplan called the "Unification of Archeological Sites of Athens", which has also gathered funding from the EU to help enhance the project. The landmark Dionysiou Areopagitou Street has been pedestrianised, forming a scenic route. The route starts from the Temple of Olympian Zeus at Vasilissis Olgas Avenue, continues under the southern slopes of the Acropolis near Plaka, and finishes just beyond the Temple of Hephaestus in Thiseio. The route in its entirety provides visitors with views of the Parthenon and the Agora (the meeting point of ancient Athenians), away from the busy City Centre. The hills of Athens also provide green space. Lycabettus, Philopappos hill and the area around it, including Pnyx and Ardettos hill, are planted with pines and other trees, with the character of a small forest rather than typical metropolitan parkland. Also to be found is the Pedion tou Areos ("Field of Mars") of 27.7 hectares, near the National Archaeological Museum. Athens' largest zoo is the Attica Zoological Park, a 20-hectare (49-acre) private zoo located in the suburb of Spata. The zoo is home to around 2000 animals representing 400 species, and is open 365 days a year. Smaller zoos exist within public gardens or parks, such as the zoo within the National Garden of Athens. The Athens Metropolitan Area consists of 58 densely populated municipalities, sprawling around the Municipality of Athens (the City Centre) in virtually all directions. For the Athenians, all the urban municipalities surrounding the City Centre are called suburbs. According to their geographic location in relation to the City of Athens, the suburbs are divided into four zones; the northern suburbs (including Agios Stefanos, Dionysos, Ekali, Nea Erythraia, Kifissia, Kryoneri, Maroussi, Pefki, Lykovrysi, Metamorfosi, Nea Ionia, Nea Filadelfeia, Irakleio, Vrilissia, Melissia, Penteli, Chalandri, Agia Paraskevi, Gerakas, Pallini, Galatsi, Psychiko and Filothei); the southern suburbs (including Alimos, Nea Smyrni, Moschato, Tavros, Agios Ioannis Rentis, Kallithea, Piraeus, Agios Dimitrios, Palaio Faliro, Elliniko, Glyfada, Lagonisi, Saronida, Argyroupoli, Ilioupoli, Varkiza, Voula, Vari and Vouliagmeni); the eastern suburbs (including Zografou, Dafni, Vyronas, Kaisariani, Cholargos and Papagou); and the western suburbs (including Peristeri, Ilion, Egaleo, Koridallos, Agia Varvara, Keratsini, Perama, Nikaia, Drapetsona, Chaidari, Petroupoli, Agioi Anargyroi, Ano Liosia, Aspropyrgos, Eleusina, Acharnes and Kamatero). The Athens city coastline, extending from the major commercial port of Piraeus to the southernmost suburb of Varkiza for some , is also connected to the City Centre by tram. In the northern suburb of Maroussi, the upgraded main Olympic Complex (known by its Greek acronym OAKA) dominates the skyline. The area has been redeveloped according to a design by the Spanish architect Santiago Calatrava, with steel arches, landscaped gardens, fountains, futuristic glass, and a landmark new blue glass roof which was added to the main stadium. A second Olympic complex, next to the sea at the beach of Palaio Faliro, also features modern stadia, shops and an elevated esplanade. Work is underway to transform the grounds of the old Athens Airport – named Elliniko – in the southern suburbs, into one of the largest landscaped parks in Europe, to be named the Hellenikon Metropolitan Park. Many of the southern suburbs (such as Alimos, Palaio Faliro, Elliniko, Glyfada, Voula, Vouliagmeni and Varkiza) known as the Athens Riviera, host a number of sandy beaches, most of which are operated by the Greek National Tourism Organisation and require an entrance fee. Casinos operate on both Mount Parnitha, some from downtown Athens (accessible by car or cable car), and the nearby town of Loutraki (accessible by car via the Athens – Corinth National Highway, or the suburban rail service Proastiakos). The large City Centre () of the Greek capital falls directly within the Municipality of Athens or Athens Municipality ()—also City of Athens. Athens Municipality is the largest in population size in Greece. Piraeus also forms a significant city centre on its own, within the Athens Urban Area and it is the second largest in population size within it, with Peristeri, Kallithea and Kypseli following. The Athens Urban Area (), also known as Urban Area of the Capital (), today consists of 40 municipalities, 35 of which make up what is referred to as the Greater Athens municipalities, located within 4 regional units (North Athens, West Athens, Central Athens, South Athens); and a further 5, which make up the Greater Piraeus municipalities, located within the regional unit of Piraeus as mentioned above. The densely built up urban area of the Greek capital sprawls across throughout the Attica Basin and has a total population of 3,074,160 (in 2011). The Athens Municipality forms the core and center of Greater Athens, which in its turn consists of the Athens Municipality and 34 more municipalities, divided in four regional units (Central, North, South and West Athens), accounting for 2,641,511 people (in 2011) within an area of . Until 2010, these four regional units made up the abolished Athens Prefecture. The municipality of Piraeus, the historic Athenian port, with its 4 suburban municipalities make up the regional unit of Piraeus, which in turn forms Greater Piraeus. Greater Athens and Greater Piraeus with part of East and West Attica regional units combined make up the continuous Athens Urban Area, also called the "Urban Area of the Capital" or simply "Athens" (the most common use of the term), spanning over , with a population of 3,090,508 people as of 2011. The Athens Urban Area is considered to form the city of Athens as a whole, despite its administrative divisions, which is the largest in Greece and one of the most populated urban areas in Europe. The Athens Metropolitan Area (), also commonly known in English as Greater Athens, spans within the Attica region and includes a total of 58 municipalities, which are organized in seven regional units (those outlined above, along with East Attica and West Attica), having reached a population of 3,737,550 based on the preliminary results of the 2011 census. Athens and Piraeus municipalities serve as the two metropolitan centres of the Athens Metropolitan Area. There are also some inter-municipal centres serving specific areas. For example, Kifissia and Glyfada serve as inter-municipal centres for northern and southern suburbs respectively. The Municipality of Athens has an official population of 664,046 people. The four regional units that make up what is referred to as Greater Athens have a combined population of 2,640,701. They together with the regional unit of Piraeus (Greater Piraeus) make up the dense Athens Urban Area which reaches a total population of 3,090,508 inhabitants (in 2011). As Eurostat the FUA of Athens had in 2013 3,828,434 inhabitants, being apparently decreasing compared with the pre-economic crisis date of 2009 (4,164,175) The municipality (Center) of Athens is the most populous in Greece, with a population of 664,046 people (in 2011) and an area of , forming the core of the Athens Urban Area within the Attica Basin. The incumbent Mayor of Athens is Kostas Bakoyannis of New Democracy. The municipality is divided into seven municipal districts which are mainly used for administrative purposes. As of the 2011 census, the population for each of the seven municipal districts of Athens is as follows: For the Athenians the most popular way of dividing the downtown is through its neighbourhoods such as Pagkrati, Ambelokipi, Goudi, Exarcheia, Patissia, Ilissia, Petralona, Plaka, Anafiotika, Koukaki and Kypseli, each with its own distinct history and characteristics. The Athens Metropolitan Area, with an area of and inhabited by 3,753,783 people in 2011, consists of the Athens Urban Area with the addition of the towns and villages of East and West Attica, which surround the dense urban area of the Greek capital. It actually sprawls over the whole peninsula of Attica, which is the best part of the region of Attica, excluding the islands. Mycenean Athens in 1600–1100 BC could have reached the size of Tiryns; that would put the population at the range of 10,000–15,000. During the Greek Dark Ages the population of Athens was around 4,000 people. In 700 BC the population grew to 10,000. In 500 BC the area probably contained 200,000 people. During the classical period the city's population is estimated from 150,000–350,000 and up to 610,000 according to Thucydides. When Demetrius of Phalerum conducted a population census in 317 BC the population was 21,000 free citizens, plus 10,000 resident aliens and 400,000 slaves. This suggests a total population of 431,000. This figure is highly suspect because of the lopsided number of slaves and does not include free women and children and resident foreigners: an estimated based on Thucydides is: 40,000 male citizens, 100,000 family members, 70,000 metics (resident foreigners) and 150,000-400,000 slaves. However the numbers would include all of Athenian territory which consisted almost entirely of modern Attica region expect the territory of the city-state of Megaris and the island section. The ancient site of the main city is centred on the rocky hill of the acropolis. In the whole of Athenian territory they existed many towns. Acharnae, Afidnes, Cytherus, Colonus, Corydallus, Cropia, Decelea, Euonymos, Vravron among others was important towns in Athenian countryside. The new port of Piraeus was a prototype harbour with the infrastructure and housing located in the site between modern passenger section of the port (named Kantharos in ancient times) and the Pasalimani harbour (named Zea in ancient times). The old one Phaliro was in the site of modern Palaio Faliro and gradually declined after the construction of the new prototype port but remained as a minor port and important settlement with historic significance in late Classical times. The rapid expansion of the modern city, which continues to this day, was initiated in the 1950s and 1960s, because of Greece's transition from an agricultural to an industrial nation. The expansion is now particularly toward the East and North East (a tendency greatly related to the new Eleftherios Venizelos International Airport and the Attiki Odos, the freeway that cuts across Attica). By this process Athens has engulfed many former suburbs and villages in Attica, and continues to do so. The table below shows the historical population of Athens in recent times. Athens became the capital of Greece in 1834, following Nafplion, which was the provisional capital from 1829. The municipality (City) of Athens is also the capital of the Attica region. The term "Athens" can refer either to the Municipality of Athens, to Greater Athens, or to the entire Athens Urban Area. Athens is twinned with: Athens is the financial capital of Greece. According to data from 2014, Athens as a metropolitan economic area produced 130 billion US-dollars as GDP in PPP, which consists nearly a half of the production for the whole country. In the list with the strongest economic metropoles of the world Athens was ranked that year 102nd, while the GDP per capita for the same year was 32,000 US-dollars. Athens is one of the major economic centres in south-eastern Europe and is considered as a regional economic power in Europe generally. The port of Piraeus, where big investments by COSCO have already been delivered during the recent decade, the completion of the new Cargo Centre in Thriasion, the expansion of the Athens Metro and the Athens Tram, as well as the projected metropolitan park in Elliniko and other economic projects are the economic landmarks of the upcoming years. Important Greek companies such as Hellenic Aerospace Industry, Mytilineos Holdings, Titan Cement, Hellenic Petroleum, Papadopoulos E.J., Folli Follie, Jumbo S.A., OPAP, and COSMOTE have their headquarters in the metropolitan area of Athens. Multinational companies such as Ericsson, Sony, Siemens, Motorola, Samsung, Microsoft, Novartis, Mondelez, Coca-Cola, etc. have their regional research and development headquarters also there. The banking sector is represented by National Bank of Greece, Alpha Bank, Eurobank, and Piraeus Bank, while the Bank of Greece is also situated in the City Centre. The Athens Stock Exchange, the only in Greece, has been severely hit by the Greek government-debt crisis and the decision of the government to proceed into capital controls during summer 2015. As a whole the economy of Athens and Greece has been severely hit with today's data showing a change from long recession to growth of 1.4% in 2017. Tourism is also a great contributor for the economy of the city, which is considered as one of the top destinations in Europe for city-break tourism and is also the gateway for excursions to the islands or the mainland. Greece attracted 26.5 million visitors in 2015, 30.1 million visitors in 2017 and over 33 million in 2018, making Greece one of the most visited countries in Europe and the world, and contributing 18% to the nation's Gross Domestic Product. Athens welcomed more than 5 million tourists in 2018 and 1,4 million of them were "city-breakers" (in 2013 the city-breakers were only 220.000). Athens is serviced by a variety of transportation means, forming the largest mass transit system of Greece. The Athens Mass Transit System consists of a large bus fleet, a trolleybus fleet that mainly serves Athens's city center, the city's Metro, a Proastiakos service and a tram network, connecting the southern suburbs to the City Centre. OSY () (Odikes Sygkoinonies S.A.) which is subsidiary company of OASA (Athens urban transport organisation), is the main operator of buses and trolleybusses in Athens. Its network consists of about 300 bus lines and 22 trolleybus lines which span the Athens Metropolitan Area, with a fleet of 1,839 buses and 366 trolleybuses. Of those 1,839 buses 416 run on compressed natural gas, making up the largest fleet of natural gas-powered buses in Europe and all trolleybusses are equipped to enable them to run on diesel in case of power failure. International links are provided by a number of private companies. National and regional bus links are provided by KTEL from two InterCity Bus Terminals, Kifissos Bus Terminal A and Liosion Bus Terminal B, both located in the north-western part of the city. "Kifissos" provides connections towards Peloponnese, North Greece, West Greece and some Ionian Islands, whereas "Liosion" is used for most of Central Greece. The Athens Metro is operated by STASY S.A () (Statheres Sygkoinonies S.A) which is a subsidiary company of OASA (Athens urban transport organisation) and provides public transport throughout the Athens Urban Area. While its main purpose is transport, it also houses Greek artifacts found during construction of the system. The Athens Metro has an operating staff of 387 and runs three metro lines, namely the line 1 (Green Line), line 2 (Red Line) and line 3 (Blue Line) lines, of which the first was constructed in 1869, and the other two largely during the 1990s, with the initial new sections opened in January 2000. Line 1 mostly runs at ground level and the other two (Line 2 & 3) routes run entirely underground. A fleet of 42 trains, using 252 carriages, operates on the network, with a daily occupancy of 1,353,000 passengers. The "line 1" (Green Line) serves 24 stations, and is the oldest line of the Athens metro network. It runs from Piraeus station to Kifissia station and covers a distance of 25.6-kilometre (15.9 mi). There are transfer connections with the Blue Line 3 at Monastiraki station and with the Red Line 2 at Omonoia and Attiki stations. The "line 2" (Red Line) runs from Anthoupoli station to Elliniko station and covers a distance of . The line connects the western suburbs of Athens with the southeast suburbs, passing through the center of Athens. The Red Line has transfer connections with the Green Line 1 at Attiki and Omonoia stations. There are also transfer connections with the Blue Line 3 at Syntagma station and with the tram at Syntagma, Syngrou Fix and Neos Kosmos stations. The "line 3" (Blue Line) runs from Agia Marina and Egaleo stations, through the central Monastiraki and Syntagma stations to Doukissis Plakentias avenue in the northeastern suburb of Halandri, covering a distance of . It then ascends to ground level and continues to Eleftherios Venizelos International Airport, using the Suburban Railway infrastructure, extending its total length to . The spring 2007 extension from Monastiraki westwards to Egaleo connected some of the main night life hubs of the city, namely those of Gazi (Kerameikos station) with Psirri (Monastiraki station) and the City Centre (Syntagma station). Extensions are under construction to the west southwest suburbs of Athens, as far as the port of Piraeus. The new stations will be Agia Barvara, Koridallos, Nikaia, Maniatika, Piraeus and Dimotiko Theatro. The completed extension will be ready in 2022 (the first three stations will open in summer 2020), connecting the biggest port of Greece, Piraeus Port, with the biggest airport of Greece the Athens International Airport. The Athens commuter rail service, referred to as the "Proastiakós", connects Eleftherios Venizelos International Airport to the city of Kiato, west of Athens, via Larissa station, the city's central rail station and the port of Piraeus. The length of Athens's commuter rail network extends to , and is expected to stretch to by 2010. Athens tram is operated by STASY S.A (Statheres Sygkoinonies S.A) which is a subsidiary company of OASA (Athens urban transport organisation). It has a fleet of 35 Sirio type vehicles which serve 48 stations, employ 345 people with an average daily occupancy of 65,000 passengers. The tram network spans a total length of and covers ten Athenian suburbs. The network runs from Syntagma Square to the southwestern suburb of Palaio Faliro, where the line splits in two branches; the first runs along the Athens coastline toward the southern suburb of Voula, while the other heads toward Neo Faliro. The network covers the majority of the Athens coastline. Further extensions are planned towards the major commercial port of Piraeus. The expansion to Piraeus will include 12 new stations, increase the overall length of tram route by , and increase the overall transportation network. Athens is served by the Athens International Airport (ATH), located near the town of Spata, in the eastern Messoghia plain, some east of center of Athens. The airport, awarded the "European Airport of the Year 2004" Award, is intended as an expandable hub for air travel in southeastern Europe and was constructed in 51 months, costing 2.2 billion euros. It employs a staff of 14,000. The airport is served by the Metro, the suburban rail, buses to Piraeus port, Athens' City Centre, Liosion and Kifisos Intercity bus stations and Elliniko metro's line 2 southern terminal, and also taxis. The airport accommodates 65 landings and take-offs per hour, with its 24-passenger boarding bridges, 144 check-in counters and broader main terminal; and a commercial area of which includes cafés, duty-free shops, and a small museum. In 2018, the airport handled 24,135,736 a huge increase over the last 4 years. In 2014, the airport handled 15,196,369 passengers, an increase of 21.2% over the previous year of 2013. Of those 15,196,369 passengers, 5,267,593 passed through the airport for domestic flights, and 9,970,006 passengers travelled through for international flights. Beyond the dimensions of its passenger capacity, ATH handled 205,294 total flights in 2007, or approximately 562 flights per day. Athens is the hub of the country's national railway system (OSE), connecting the capital with major cities across Greece and abroad (Istanbul, Sofia, Belgrade and Bucharest). The Port of Piraeus is the largest port in Greece and one of the largest in Europe. It is the largest container port in East Mediterranean Sea Basin. It is also the busiest passenger port in Europe and one of the largest passenger ports in the world. Connects Athens to the numerous Greek islands of the Aegean Sea, with ferries departing, while also serving the cruise ships that arrive. Rafina and Lavrio act as alternative ports of Athens, connects the city with numerous Greek islands of the Aegean Sea, Evia and Cesme in Turkey, while also serving the cruise ships that arrive. Two main motorways of Greece begin in Athens, namely the A1/E75, heading north towards Greece's second largest city, Thessaloniki; and the border crossing of Evzones and the A8/E94 heading west, towards Greece's third largest city, Patras, which incorporated the GR-8A. Before their completion much of the road traffic used the GR-1 and the GR-8. Athens' Metropolitan Area is served by the motorway network of the Attiki Odos toll-motorway (code: A6). Its main section extends from the western industrial suburb of Elefsina to Athens International Airport; while two beltways, namely the Aigaleo Beltway (A65) and the Hymettus Beltway (A64) serve parts of western and eastern Athens respectively. The span of the Attiki Odos in all its length is , making it the largest metropolitan motorway network in all of Greece. Located on Panepistimiou Street, the old campus of the University of Athens, the National Library, and the Athens Academy form the "Athens Trilogy" built in the mid-19th century. Most of the university's workings have been moved to a much larger, modern campus located in the eastern suburb of Zografou. The second higher education institution in the city is the Athens Polytechnic School, found in Patission Street. This was the location where on 17 November 1973, more than 13 students were killed and hundreds injured inside the university during the Athens Polytechnic uprising, against the military junta that ruled the nation from 21 April 1967 until 23 July 1974. Other universities that lie within Athens are the Athens University of Economics and Business, the Panteion University, the Agricultural University of Athens and the University of Piraeus. There are overall ten state-supported Institutions of Higher (or Tertiary) education located in the Athens Urban Area, these are by chronological order: Athens School of Fine Arts (1837), National Technical University of Athens (1837), National and Kapodistrian University of Athens (1837), Agricultural University of Athens (1920), Athens University of Economics and Business (1920), Panteion University of Social and Political Sciences (1927), University of Piraeus (1938), Harokopio University of Athens (1990), School of Pedagogical and Technological Education (2002), University of West Attica (2018). There are also several other private "colleges", as they called formally in Greece, as the establishment of private universities is prohibited by the constitution. Many of them are accredited by a foreign state or university such as the American College of Greece and the Athens Campus of the University of Indianapolis. The city is a world centre of archaeological research. Along with national institutions, such as the Athens University and the Archaeological Society, there are multiple archaeological Museums including the National Archaeological Museum, the Cycladic Museum, the Epigraphic Museum, the Byzantine & Christian Museum, as well as museums at the ancient Agora, Acropolis, Kerameikos, and the Kerameikos Archaeological Museum. The city is also home to the Demokritos laboratory for Archaeometry, alongside regional and national archaeological authorities that form part of the Greek Department of Culture. Athens hosts 17 Foreign Archaeological Institutes which promote and facilitate research by scholars from their home countries. As a result, Athens has more than a dozen archaeological libraries and three specialized archaeological laboratories, and is the venue of several hundred specialized lectures, conferences and seminars, as well as dozens of archaeological exhibitions, each year. At any given time, hundreds of international scholars and researchers in all disciplines of archaeology are to be found in the city. Athens incorporates architectural styles ranging from Greco-Roman and Neoclassical to modern times. They are often to be found in the same areas, as Athens is not marked by a uniformity of architectural style. A visitor will quickly notice the absence of tall buildings: Athens has very strict height restriction laws in order to ensure the Acropolis hill is visible throughout the city. For the greatest part of the 19th century Neoclassicism dominated Athens, as well as some deviations from it such as Eclecticism, especially in the early 20th century. Thus, the Old Royal Palace was the first important public building to be built, between 1836 and 1843. Later in the mid and late 19th century, Theophil Freiherr von Hansen and Ernst Ziller took part in the construction of many neoclassical buildings such as the Athens Academy and the Zappeion Hall. Ziller also designed many private mansions in the centre of Athens which gradually became public, usually through donations, such as Schliemann's Iliou Melathron. Beginning in the 1920s, modern architecture including Bauhaus and Art Deco began to exert an influence on almost all Greek architects, and buildings both public and private were constructed in accordance with these styles. Localities with a great number of such buildings include Kolonaki, and some areas of the centre of the city; neighbourhoods developed in this period include Kypseli. In the 1950s and 1960s during the extension and development of Athens, other modern movements such as the International style played an important role. The centre of Athens was largely rebuilt, leading to the demolition of a number of neoclassical buildings. The architects of this era employed materials such as glass, marble and aluminium, and many blended modern and classical elements. After World War II, internationally known architects to have designed and built in the city included Walter Gropius, with his design for the US Embassy, and, among others, Eero Saarinen, in his postwar design for the east terminal of the Ellinikon Airport. All over the city can be found several statues or busts. Apart from the neoclassicals by Leonidas Drosis at the Academy of Athens (Plato, Socrates, Apollo, Athena), other notable include the statue of Theseus by Georgios Fytalis at Thiseion, of philhellenes like Lord Byron, George Canning and William Gladstone, the equestrian statue of Theodoros Kolokotronis by Lazaros Sochos in front of the Old Parliament, statues of Ioannis Kapodistrias, Rigas Feraios and Adamantios Korais at the University, of Evangelos Zappas and Konstantinos Zappas at Zappeion, of Ioannis Varvakis at the National Garden, the "woodbreaker" by Dimitrios Filippotis, the equestrian statue of Alexandros Papagos at Papagou district and various busts of fighters of Greek independence at the Pedion tou Areos. A significant landmark is also the Tomb of the Unknown Soldier in Syntagma. Athens' most important museums include: Athens has been a destination for travellers since antiquity. Over the past decade, the city's infrastructure and social amenities have improved, in part because of its successful bid to stage the 2004 Olympic Games. The Greek Government, aided by the EU, has funded major infrastructure projects such as the state-of-the-art Eleftherios Venizelos International Airport, the expansion of the Athens Metro system, and the new Attiki Odos Motorway. Athens was voted as the third best European city to visit in 2015 by European Best Destination. More than 240,000 people voted. Athens is home to 148 theatrical stages, more than any other city in the world, including the ancient Odeon of Herodes Atticus, home to the Athens Festival, which runs from May to October each year. In addition to a large number of multiplexes, Athens plays host to open air garden cinemas. The city also supports music venues, including the Athens Concert Hall ("Megaro Moussikis"), which attracts world class artists. The Athens Planetarium, located in Andrea Syngrou Avenue, in Palaio Faliro is one of the largest and best equipped digital planetaria in the world. The Stavros Niarchos Foundation Cultural Center, inaugurated in 2016, will house the National Library of Greece and the Greek National Opera. The most successful songs during the period 1870–1930 were the so-called Athenian serenades (Αθηναϊκές καντάδες), based on the Heptanesean kantádhes (καντάδες 'serenades'; sing.: καντάδα) and the songs performed on stage (επιθεωρησιακά τραγούδια 'theatrical revue songs') in revues, musical comedies, operettas and nocturnes that were dominating Athens' theatre scene. Notable composers of operettas or nocturnes were Kostas Giannidis, Dionysios Lavrangas, Nikos Hatziapostolou, while Theophrastos Sakellaridis' "The Godson" remains probably the most popular operetta. Despite the fact that the Athenian songs were not autonomous artistic creations (in contrast with the serenades) and despite their original connection with mainly dramatic forms of Art, they eventually became hits as independent songs. Notable actors of Greek operettas, who made also a series of melodies and songs popular at that time, include Orestis Makris, Kalouta sisters, Vasilis Avlonitis, Afroditi Laoutari, Eleni Papadaki, Marika Nezer, Marika Krevata and others. After 1930, wavering among American and European musical influences as well as the Greek musical tradition. Greek composers begin to write music using the tunes of the tango, waltz, swing, foxtrot, some times combined with melodies in the style of Athenian serenades' repertory. Nikos Gounaris was probably the most renowned composer and singer of the time. In 1923, after the population exchange between Greece and Turkey, many ethnic Greeks from Asia Minor fled to Athens as a result of the Greco-Turkish War. They settled in poor neighborhoods and brought with them Rebetiko music, making it popular also in Greece, which became later the base for the Laïko music. Other forms of song popular today in Greece are elafrolaika, entechno, dimotika, and skyladika. Greece's most notable, and internationally famous, composers of Greek song, mainly of the entechno form, are Manos Hadjidakis and Mikis Theodorakis. Both composers have achieved fame abroad for their composition of film scores. Athens has a long tradition in sports and sporting events, serving as home to the most important clubs in Greek sport and housing a large number of sports facilities. The city has also been host to sports events of international importance. Athens has hosted the Summer Olympic Games twice, in 1896 and 2004. The 2004 Summer Olympics required the development of the Athens Olympic Stadium, which has since gained a reputation as one of the most beautiful stadiums in the world, and one of its most interesting modern monuments. The biggest stadium in the country, it hosted two finals of the UEFA Champions League, in 1994 and 2007. Athens' other major stadium, located in the Piraeus area, is the Karaiskakis Stadium, a sports and entertainment complex, host of the 1971 UEFA Cup Winners' Cup Final. Athens has hosted the EuroLeague final three times, the first in 1985 and second in 1993, both at the Peace and Friendship Stadium, most known as SEF, a large indoor arena, and the third time in 2007 at the Olympic Indoor Hall. Events in other sports such as athletics, volleyball, water polo etc., have been hosted in the capital's venues. Athens is home to three European multi-sport clubs: Olympiacos, Panathinaikos, AEK Athens. In football, Olympiacos have dominated the domestic competitions, Panathinaikos made it to the 1971 European Cup Final, while AEK Athens is the other member of the big three. These clubs also have basketball teams; Panathinaikos and Olympiacos are among the top powers in European basketball, having won the Euroleague six times and three respectively, whilst AEK Athens was the first Greek team to win a European trophy in any team sport. Other notable clubs within Athens are Athinaikos, Panionios, Atromitos, Apollon, Panellinios, Ethnikos Piraeus, Maroussi BCE and Peristeri B.C.. Athenian clubs have also had domestic and international success in other sports. The Athens area encompasses a variety of terrain, notably hills and mountains rising around the city, and the capital is the only major city in Europe to be bisected by a mountain range. Four mountain ranges extend into city boundaries and thousands of miles of trails criss-cross the city and neighbouring areas, providing exercise and wilderness access on foot and bike. Beyond Athens and across the prefecture of Attica, outdoor activities include skiing, rock climbing, hang gliding and windsurfing. Numerous outdoor clubs serve these sports, including the Athens Chapter of the Sierra Club, which leads over 4,000 outings annually in the area. Beside the above clubs, inside the boundaries of Athens Municipality there are some more clubs with presence in national divisions or notable action for short periods. Some of them are PAO Rouf (Rouf) with earlier presence in Gamma Ethniki, Petralona F.C.() (Petralona), football club founded in 1963, with earlier presence in Beta Ethniki, Attikos F.C.() (Kolonos), football club founded in 1919 with short presence in Gamma Ethniki, (Kypseli), football club founded in 1938 with short presence in Gamma Ethniki, Gyziakos (Gyzi), basketball club founded in 1937 with short presence in Beta Ethniki basketball and Aetos B.C.() (Agios Panteleimonas), basketball club founded in 1992 with earlier presence in A2 Ethniki Basketball. Another important Athenian sport club is the Athens Tennis Club founded in 1895 with important offer for the Greek tennis. The revival of the modern Olympic Games was brought forth in 1896, by Frenchman Pierre de Coubertin. Thanks to his efforts, Athens was awarded the first modern Olympic Games. In 1896, the city had a population of 123,000 and the event helped boost the city's international profile. Of the venues used for these Olympics, the Kallimarmaro Stadium, and Zappeion were most crucial. The Kallimarmaro is a replica of the ancient Athenian stadiums, and the only major stadium (in its capacity of 60,000) to be made entirely of white marble from Mount Penteli, the same material used for construction of the Parthenon. The 1906 Summer Olympics, or the 1906 Intercalated games, were held in Athens. The intercalated competitions were intermediate games to the internationally organized Olympics, and were meant to be organized in Greece every four years, between the main Olympics. This idea later lost support from the IOC and these games were discontinued. Athens was awarded the 2004 Summer Olympics on 5 September 1997 in Lausanne, Switzerland, after having lost a previous bid to host the 1996 Summer Olympics, to Atlanta, United States. It was to be the second time Athens would host the games, following the inaugural event of 1896. After an unsuccessful bid in 1990, the 1997 bid was radically improved, including an appeal to Greece's Olympic history. In the last round of voting, Athens defeated Rome with 66 votes to 41. Prior to this round, the cities of Buenos Aires, Stockholm and Cape Town had been eliminated from competition, having received fewer votes. During the first three years of preparations, the International Olympic Committee had expressed concern over the speed of construction progress for some of the new Olympic venues. In 2000 the Organising Committee's president was replaced by Gianna Angelopoulos-Daskalaki, who was the president of the original Bidding Committee in 1997. From that point forward, preparations continued at a highly accelerated, almost frenzied pace. Although the heavy cost was criticized, estimated at $1.5 billion, Athens was transformed into a more functional city that enjoys modern technology both in transportation and in modern urban development. Some of the finest sporting venues in the world were created in the city, all of which were fully ready for the games. The games welcomed over 10,000 athletes from all 202 countries. The 2004 Games were judged a success, as both security and organization worked well, and only a few visitors reported minor problems mainly concerning accommodation issues. The 2004 Olympic Games were described as "Unforgettable, dream Games", by IOC President Jacques Rogge for their return to the birthplace of the Olympics, and for meeting the challenges of holding the Olympic Games. The only observable problem was a somewhat sparse attendance of some early events. Eventually, however, a total of more than 3.5 million tickets were sold, which was higher than any other Olympics with the exception of Sydney (more than 5 million tickets were sold there in 2000). In 2008 it was reported that most of the Olympic venues had fallen into disrepair: according to those reports, 21 of the 22 facilities built for the games had either been left abandoned or are in a state of dereliction, with several squatter camps having sprung up around certain facilities, and a number of venues afflicted by vandalism, graffiti or strewn with rubbish. These claims, however, are disputed and likely to be inaccurate, as most of the facilities used for the Athens Olympics are either in use or in the process of being converted for post-Olympics use. The Greek Government has created a corporation, Olympic Properties SA, which is overseeing the post-Olympics management, development and conversion of these facilities, some of which will be sold off (or have already been sold off) to the private sector, while other facilities are still in use just as during the Olympics, or have been converted for commercial use or modified for other sports. Concerts and theatrical shows, such as those by the troupe Cirque du Soleil, have recently been held in the complex.
https://en.wikipedia.org/wiki?curid=1216
Anguilla Anguilla ( ) is a British overseas territory in the Caribbean. It is one of the most northerly of the Leeward Islands in the Lesser Antilles, lying east of Puerto Rico and the Virgin Islands and directly north of Saint Martin. The territory consists of the main island of Anguilla, approximately long by wide at its widest point, together with a number of much smaller islands and cays with no permanent population. The territory's capital is The Valley. The total land area of the territory is , with a population of approximately 17,400 (July 2018 est.). The name Anguilla is from the Italian "" meaning "eel" (in turn from the Latin "anguilla", diminutive of "anguis", snake) in reference to the island's shape. It is believed by most sources to have been named by Christopher Columbus. For similar reasons, it was also known as "Snake" or "Snake Island". Anguilla was first settled by Indigenous Amerindian peoples who migrated from South America. The earliest Native American artefacts found on Anguilla have been dated to around 1300 ; remains of settlements date from  600. The native Arawak name for the island was "Malliouhana". Precisely when Anguilla was first seen by Europeans is uncertain: some sources claim that Columbus sighted the island during his second voyage in 1493, while others state that the first European explorer was the French Huguenot nobleman and merchant René Goulaine de Laudonnière in 1564. The Dutch West India Company established a fort on the island in 1631. However, the Company later withdrew after its fort was destroyed by the Spanish in 1633. Traditional accounts state that Anguilla was first colonised by English settlers from Saint Kitts beginning in 1650. The settlers focused on planting tobacco, and to a lesser extent cotton. The French temporarily took over the island in 1666 but returned it to English control under the terms of the Treaty of Breda the next year. A Major John Scott who visited in September 1667, wrote of leaving the island "in good condition" and noted that in July 1668, "200 or 300 people fled thither in time of war". The French attacked again in 1688, 1745 and 1798, causing much destruction but failing to capture the island. It is likely that the early European settlers brought enslaved Africans with them. Historians confirm that African slaves lived in the region in the early 17th century, such as slaves from Senegal living on St Kitts in the mid 1600s. By 1672 a slave depot existed on the island of Nevis, serving the Leeward Islands. While the time of African arrival in Anguilla is difficult to place precisely, archival evidence indicates a substantial African presence of at least 100 enslaved people by 1683; these seem to have come from Central Africa as well as West Africa. The slaves were forced to work on the sugar plantations which had begun to replace tobacco as Anguilla's main crop. Over time the African slaves and their descendants came to vastly outnumber the white settlers. The African slave trade was eventually terminated within the British Empire in 1807, and slavery outlawed completely in 1834. Many planters subsequently sold up or left the island. During the early colonial period, Anguilla was administered by the British through Antigua; in 1825, it was placed under the administrative control of nearby Saint Kitts. Anguilla was federated with St Kitts and Nevis in 1882, against the wishes of many Anguillians. Economic stagnation, and the severe effects of several droughts in the 1890s and later the Great Depression of the 1930s led many Anguillians to emigrate for better prospects elsewhere. Full adult suffrage was introduced to Anguilla in 1952. After a brief period as part of the West Indies Federation (1958–62), the island of Anguilla became part of the associated state of Saint Kitts-Nevis-Anguilla with full internal autonomy in 1967. However many Anguillians had no wish to be a part of this union, and resented the dominance of St Kitts within it. On 30 May 1967 Anguillians forcibly ejected the St Kitts police force from the island and declared their separation from St Kitts following a referendum. The events, led by Atlin Harrigan and Ronald Webster amongst others, became known as the Anguillian Revolution; its goal was not independence per se, but rather independence from Saint Kitts and Nevis and a return to being a British colony. With negotiations failing to break the deadlock, a second referendum confirming Anguillians' desire for separation from St Kitts was held and the Republic of Anguilla was declared unilaterally, with Ronald Webster as President. Efforts by British envoy William Whitlock failed to break the impasse and 300 British troops were subsequently sent in March 1969. British authority was restored, and confirmed by the Anguilla Act of July 1971. In 1980, Anguilla was finally allowed to formally secede from Saint Kitts and Nevis and become a separate British Crown colony (now a British overseas territory). Since then, Anguilla has been politically stable, and has seen a large growth in its tourism and offshore financing sectors. Anguilla is a flat, low-lying island of coral and limestone in the Caribbean Sea, measuring some 16 miles (26 km) long and 3.5 miles (6 km) in width. It lies to the east of Puerto Rico and the Virgin Islands and directly north of Saint Martin, separated from that island by the Anguilla Channel. The soil is generally thin and poor, supporting scrub, tropical and forest vegetation. The terrain is generally low-lying, with the highest terrain located in the vicinity of The Valley; Crocus Hill, Anguilla's highest peak at , lies in the western regions of the town. Anguilla is noted for its ecologically important coral reefs and beaches. Apart from the main island of Anguilla itself, the territory includes a number of other smaller islands and cays, mostly tiny and uninhabited: Anguilla has a volcanic origin and has been submerged repeatedly from climate change. Northeastern trade winds keep this tropical island relatively cool and dry. Average annual temperature is . July–October is its hottest period, December–February, its coolest. Rainfall averages annually, although the figures vary from season to season and year to year. The island is subject to both sudden tropical storms and hurricanes, which occur in the period from July to November. The island suffered damage in 1995 from Hurricane Luis and severe flooding from Hurricane Lenny. Anguilla is an internally self-governing overseas territory of the United Kingdom. Its politics take place in a framework of a parliamentary representative democratic dependency, whereby the Premier is the head of government, and of a pluriform multi-party system. The United Nations Committee on Decolonization includes Anguilla on the United Nations list of Non-Self-Governing Territories. The territory's constitution is the Anguilla Constitutional Order 1 April 1982 (amended 1990). Executive power is exercised by the government, with legislative power being vested in both the government and the House of Assembly. The judiciary is independent of the executive and the legislature. As a dependency of the UK, the UK is responsible for Anguilla's military defence, although there are no active garrisons or armed forces present. Anguilla has a small marine police force, comprising around 32 personnel, which operates one VT Halmatic M160-class 52-foot fast patrol boat. The majority of residents (90.08%) are black, most of whom are the descendants of slaves transported from Africa. Minorities include whites at 3.74% and people of mixed race at 4.65% (figures from 2001 census). 72% of the population is Anguillian while 28% is non-Anguillian (2001 census). Of the non-Anguillian population, many are citizens of the United States, United Kingdom, St Kitts & Nevis, the Dominican Republic, Jamaica and Nigeria. 2006 and 2007 saw an influx of large numbers of Chinese, Indian and Mexican workers, brought in as labour for major tourist developments due to the local population not being large enough to support the labour requirements. Christian churches did not have a consistent or strong presence during the initial period of English colonisation; spiritual and religious practices of Europeans and Africans tended to reflect their regional origins. As early as 1813, Christian ministers formally ministered to enslaved Africans and promoted literacy among converts. The Wesleyan (Methodist) Missionary Society of England built churches and schools from 1817. According to the 2001 census, Christianity is Anguilla's predominant religion, with 29% of the population practising Anglicanism and another 23.9% are Methodist. Other churches on the island include Seventh-day Adventist, Baptist, Roman Catholic (served by the Diocese of Saint John's–Basseterre, with the See at Saint John on Antigua and Barbuda) and a small community of Jehovah's Witnesses (0.7%). Between 1992 and 2001, the number of followers of the Church of God and Pentecostals increased considerably. There are at least 15 churches on the island. Although a minority on the island, Anguilla is an important location to followers of Rastafarian religion as the birthplace of Robert Athlyi Rogers, author of the "Holy Piby" which had a strong influence on Rastafarian and other Africa-centre belief systems. More recently, a Muslim cultural centre has opened on the island. Today most people in Anguilla speak a British-influenced variety of standard English. Other languages are also spoken on the island, including varieties of Spanish, Chinese and the languages of other immigrant communities. However, the most common language other than Standard English is the island's own English-lexifier Creole language (not to be confused with Antillean Creole ('French Creole'), spoken in French islands such as Martinique and Guadeloupe). It is referred to locally by terms such as "dialect" (pronounced "dialek"), Anguilla Talk or "Anguillian". It has its main roots in early varieties of English and West African languages, and is similar to the dialects spoken in English-speaking islands throughout the Eastern Caribbean in terms of its structural features. Linguists who are interested in the origins of Anguillian and other Caribbean Creoles point out that some of its grammatical features can be traced to African languages while others can be traced to European languages. Three areas have been identified as significant for the identification of the linguistic origins of those forced migrants who arrived before 1710: the Gold Coast, the Slave Coast and the Windward Coast. Sociohistorical information from Anguilla's archives suggest that Africans and Europeans formed two distinct, but perhaps overlapping speech communities in the early phases of the island's colonisation. "Anguillian" is believed to have emerged as the language of the masses as time passed, slavery was abolished and locals began to see themselves as "belonging" to Anguillian society. There are six government primary schools, one government secondary school (Albena Lake Hodge Comprehensive School), and two private schools. There is a single library, the Edison L. Hughes Education & Library Complex of the Anguilla Public Library. A branch of the Saint James School of Medicine was established in 2011 in Anguilla. It is a private, for-profit medical school headquartered in Park Ridge, Illinois. There is a University of the West Indies Open campus site in the island. The island's cultural history begins with the native Taino, Arawak and Carib. Their artefacts have been found around the island, telling of life before European settlers arrived. The Anguilla National Trust (ANT) was established in 1988 and opened its offices in 1993 charged with the responsibility of preserving the heritage of the island, including its cultural heritage. As throughout the Caribbean, holidays are a cultural fixture. Anguilla's most important holidays are of historic as much as cultural importance – particularly the anniversary of the emancipation (previously August Monday in the Park), celebrated as the Summer Festival, or Carnival. British festivities, such as the Queen's Birthday, are also celebrated. Anguillian cuisine is influenced by native Caribbean, African, Spanish, French and English cuisines. Seafood is abundant, including prawns, shrimp, crab, spiny lobster, conch, mahi-mahi, red snapper, marlin and grouper. Salt cod is a staple food eaten on its own and used in stews, casseroles and soups. Livestock is limited due to the small size of the island and people there use poultry, pork, goat and mutton, along with imported beef. Goat is the most commonly eaten meat, used in a variety of dishes. The official national food of Anguilla is pigeon peas and rice. A significant amount of the island's produce is imported due to limited land suitable for agriculture production; much of the soil is sandy and infertile. Among the agriculture produced in Anguilla includes tomatoes, peppers, limes and other citrus fruits, onion, garlic, squash, pigeon peas and callaloo. Starch staple foods include imported rice and other foods that are imported or locally grown, including yams, sweet potatoes and breadfruit. The Anguilla National Trust has programmes encouraging Anguillian writers and the preservation of the island's history. In 2015, "Where I See The Sun – Contemporary Poetry in Anguilla" A New Anthology by Lasana M. Sekou was published by House of Nehesi Publishers. Among the forty three poets in the collection are Rita Celestine-Carty, Bankie Banx, John T. Harrigan, Patricia J. Adams, Fabian Fahie, Dr. Oluwakemi Linda Banks, and Reuel Ben Lewi. Various Caribbean musical genres are popular on the island, such as reggae, soca and calypso. Boat racing has deep roots in Anguillian culture and is the national sport. There are regular sailing regattas on national holidays, such as Carnival, which are contested by locally built and designed boats. These boats have names and have sponsors that print their logo on their sails. As in many other former British colonies, cricket is also a popular sport. Anguilla is the home of Omari Banks, who played for the West Indies Cricket Team, while Cardigan Connor played first-class cricket for English county side Hampshire and was 'chef de mission' (team manager) for Anguilla's Commonwealth Games team in 2002. Other noted players include Chesney Hughes, who plays for Derbyshire in England. Rugby union is represented in Anguilla by the Anguilla Eels RFC, who were formed in April 2006. The Eels have been finalists in the St. Martin tournament in November 2006 and semi-finalists in 2007, 2008, 2009 and Champions in 2010. The Eels were formed in 2006 by Scottish club national second row Martin Welsh, Club Sponsor and President of the AERFC Ms. Jacquie Ruan, and Canadian standout Scrumhalf Mark Harris (Toronto Scottish RFC). Anguilla is the birthplace of sprinter Zharnel Hughes who has represented Great Britain since 2015, and England at the 2018 Commonwealth Games. He won the 100 metres at the 2018 European Athletics Championships, the 4 x 100 metres at the same championships, and the 4 x 100 metres for England at the 2018 Commonwealth Games. Shara Proctor, British Long Jump Silver Medalist in World Championships in Beijing first represented Anguilla in the event until 2010 when she began to represent Great Britain and England. Under the Anguillian Flag she achieved several medals in the NACAC games. Keith Connor, triple jumper, is also an Anguillian. He represented Great Britain and England and achieved several international titles including Commonwealth and European Games gold medals and an Olympic bronze medal. Keith later became Head Coach of Australia Athletics. Anguilla has habitat for the Cuban tree frogs ("Osteopilus septentrionalis"). The red-footed tortoise ("Chelonoidis carbonaria") is a species of tortoise found here, which originally came from South America. Hurricanes led to over-water dispersal for the green iguanas ("Iguana iguana") to colonise Anguilla. All three animals are introduced. Five species of bats are known in the literature from Anguilla – the threatened insular single leaf bat ("Monophyllus plethodon"), the Antillean fruit-eating bat ("Brachyphylla cavernarum"), the Jamaican fruit bat ("Artibeus jamaicensis"), the Mexican funnel-eared bat ("Natalus stramineus"), and the velvety free-tailed bat ("Molossus molossus"). Anguilla's thin arid soil being largely unsuitable for agriculture, the island has few land-based natural resources. Its main industries are tourism, offshore incorporation and management, offshore banking, captive insurance and fishing. Anguilla's currency is the East Caribbean dollar, though the US dollar is also widely accepted. The exchange rate is fixed to the US dollar at US$1 = EC$2.70. The economy, and especially the tourism sector, suffered a setback in late 1995 due to the effects of Hurricane Luis in September. Hotels were hit particularly hard but a recovery occurred the following year. Another economic setback occurred during the aftermath of Hurricane Lenny in 2000. Before the 2008 worldwide crisis, the economy of Anguilla was growing strongly, especially the tourism sector which was driving major new developments in partnerships with multi-national companies. Anguilla's tourism industry received a major boost when it was selected to host the World Travel Awards in December 2014. Known as "the Oscars of the travel industry", the awards ceremony was held at the CuisinArt Resort and Spa and was hosted by Vivica A. Fox. Anguilla was voted the World's Leading Luxury Island Destination from a short list of top-tier candidates such as St. Barts, the Maldives and Mauritius. Anguilla's financial system comprises seven banks, two money services businesses, more than 40 company managers, more than 50 insurers, 12 brokers, more than 250 captive intermediaries, more than 50 mutual funds and eight trust companies. Anguilla has become a popular tax haven, having no capital gains, estate, profit, sales, or corporate taxes. In April 2011, faced with a mounting deficit, it introduced a 3% "Interim Stabilisation Levy", Anguilla's first form of income tax. Anguilla also has a 0.75% property tax. Anguilla aims to obtain 15% of its energy from solar power to become less reliant on expensive imported diesel. The Climate & Development Knowledge Network is helping the government gather the information it needs to change the territory's legislation, so that it can integrate renewables into its grid. Barbados has also made good progress in switching to renewables, but many other Small Island Developing States are still at the early stages of planning how to integrate renewable energy into their grids. "For a small island we're very far ahead," said Beth Barry, Coordinator of the Anguilla Renewable Energy Office. "We've got an Energy Policy and a draft Climate Change policy and have been focussing efforts on the question of sustainable energy supply for several years now. As a result, we have a lot of information we can share with other islands." Anguilla is served by Clayton J. Lloyd International Airport (prior to 4 July 2010 known as Wallblake Airport). The primary runway at the airport is in length and can accommodate moderate-sized aircraft. Services connect to various other Caribbean islands via regional carrier LIAT, local charter airlines and others. Although there are no direct scheduled flights to or from continental America or Europe, Tradewind Aviation and Cape Air provide scheduled air service to San Juan, Puerto Rico. The airport can handle large narrow-body jets such as the Boeing 727, Boeing 737 and Boeing 757. Aside from taxis, there is no public transport on the island. Cars drive on the left. There are regular ferries from Saint Martin to Anguilla. It is a 20-minute crossing from Marigot, St. Martin to Blowing Point, Anguilla. Ferries commence service from 7:00 am. There is also a charter service, from Blowing Point, Anguilla to Princess Juliana Airport to make travel easier. This way of travel is the most common method of transport between Anguilla and St. Martin.
https://en.wikipedia.org/wiki?curid=1217
Telecommunications in Anguilla This article is about communications systems in Anguilla. Telephones - main lines in use: 6,200 (2002) Telephones - mobile cellular: 1,800 (2002) Telephone system: "Domestic:" Modern internal telephone system "International:" EAST CARIBBEAN FIBRE SYSTEM ECFS (cable system) " microwave radio relay to island of Saint Martin (Guadeloupe and Netherlands Antilles) Mobile Phone Operators: FLOW (Anguilla) Ltd. - GSM and UMTS 850 and 1900 MHz, LTE 700 MHz with Island-wide coverage Digicel (Anguilla) Ltd. - GSM and UMTS 850 to 1900 MHz , LTE 700 MHz Mobiles: ? (2007) Radio broadcast stations: AM 3, FM 7, shortwave 0 (2007) Radios: 3,000 (1997) Television broadcast stations: 1 (1997) Televisions: 1,000 (1997) Internet country code: .ai (Top level domain) Internet Service Providers (ISPs): 2 (FLOW - , Digicel Anguilla - ) Internet hosts: 269 (2012) Internet: users: 12,377 (2018)
https://en.wikipedia.org/wiki?curid=1223
Ashmore and Cartier Islands The Territory of Ashmore and Cartier Islands is an uninhabited external territory of Australia consisting of four low-lying tropical islands in two separate reefs, and the territorial sea generated by the islands. The territory is located in the Indian Ocean situated on the edge of the continental shelf, about off the northwest coast of Australia and south of the Indonesian island of Rote. Ashmore Reef is called "Pulau Pasir" by Indonesians and "Nusa Solokaek" in the Rotenese language. Both names have the meaning "sand island". The Territory comprises Ashmore Reef, which includes West, Middle, and East Islands, and two lagoons, and Cartier Reef, which includes Cartier Island. Ashmore Reef covers approximately and Cartier Reef , both measurements extending to the limits of the reefs. West, Middle, and East Islands have a combined land area variously reported as , and . Cartier Island has a reported land area of . According to Australian literature, Cartier Island was discovered by Captain Nash in 1800, and named after his ship "Cartier". Ashmore Island was discovered by Captain Samuel Ashmore in 1811 from his ship and named after him. Ashmore Island was annexed by the United Kingdom in 1878, as was Cartier Island in 1909. A British order-in-council dated 23 July 1931 stated that Ashmore and Cartier Islands would be placed under the authority of the Commonwealth of Australia when Australia passed legislation to accept them. The Commonwealth's resulting "Ashmore and Cartier Islands Acceptance Act 1933" came into operation on 10 May 1934, when the islands formally became a territory. The act authorised the Governor of Western Australia to make ordinances for the territory. In July 1938 the territory was annexed to the Northern Territory, then also administered by the Commonwealth, whose laws, ordinances and regulations applied to the Territory. When self-government was granted to the Northern Territory on 1 July 1978, administration of Ashmore and Cartier Islands was retained by the Commonwealth. In 1983 the territory was declared a nature reserve under the "National Parks and Wildlife Conservation Act 1975", now replaced by the "Environment Protection and Biodiversity Conservation Act 1999". After the islands became a first point of contact with the Australian migration zone, in September 2001, the Australian government excised the Ashmore and Cartier Islands from the Australian migration zone. Ashmore has been regularly visited and fished by Indonesian fishermen since the early eighteenth century. A 1974 Memorandum of Understanding (MOU) between Australia and Indonesia sets out arrangements by which traditional fishers can access resources in Australia's territorial sea in the region. This allows traditional Indonesian fishermen to access parts of Ashmore for shelter, freshwater and to visit grave sites. The area, known as the MOU Box, contains the Ashmore and Cartier Islands Territory. Today, the Territory is administered from Canberra by the Department of Infrastructure, Regional Development and Cities, which is also responsible for the administration of the territories of Christmas Island, Cocos (Keeling) Islands, the Coral Sea Islands, Jervis Bay Territory and Norfolk Island. The Attorney-General's Department had been responsible for the administration of Australian territories until the 2010 federal election. In that year the responsibility for Australian territories was transferred to the then Department of Regional Australia, Local Government, Arts and Sport, and from 18 September 2013 the Department of Infrastructure and Regional Development has administered Australian territories. Defence of Ashmore and Cartier Islands is the responsibility of Australia, with periodic visits by the Royal Australian Navy, Royal Australian Air Force and Australian Customs and Border Protection Service. Nearby Hibernia Reef, northeast of Ashmore Reef, is not part of the Territory, but belongs to Western Australia. It has no permanently dry land area, although large parts of the reef become exposed during low tide. The Ashmore Reef Marine Park and Cartier Island Marine Park are both classed as strict nature reserves (IUCN Ia) and protect biodiverse areas of significant and international importance, as well as cultural heritage. There is no economic activity in the Territory, Ashmore and Cartier Islands being uninhabited. Cartier Island is an unvegetated sand island. Access to Cartier Island is prohibited because of the risk of unexploded ordnances. There are no ports or harbours, only offshore anchorage. The customs vessel ACV "Ashmore Guardian" is stationed off the reef for up to 330 days per year. The islands are also visited by seasonal caretakers and occasional scientific researchers. The area has been a traditional fishing ground of Indonesian fishermen for centuries, and continues. In the 1850s, American whalers operated in the region. Mining of phosphate deposits took place on Ashmore Island in the latter half of the 19th century. Today, all the wells in the Territory are infected with cholera or contaminated and undrinkable. Petroleum extraction activities take place at the Jabiru and Challis oil fields, which are adjacent to the Territory, and which are administered by the Northern Territory Department of Mines and Energy on behalf of the Commonwealth. As Ashmore Reef is the closest point of Australian territory to Indonesia, it was a popular target for people smugglers transporting asylum seekers en route to Australia. Once they had landed on Ashmore Island, asylum seekers could claim to have entered Australian migration zone and request to be processed as refugees. The use of Ashmore Island for this purpose created great notoriety during late 2001, when refugee arrivals became a major political issue in Australia. The Australian Government argued that as Australia was not the country of first asylum for these "boat people", Australia did not have a responsibility to accept them. A number of things were done to discourage the use of the Territory for this purpose, such as attempting to have the people smugglers arrested in Indonesia; the so-called Pacific Solution of processing them in third countries; the boarding and forced turnaround of the boats by Australian military forces; and finally excising the Territory and many other small islands from the Australian migration zone. Two boatloads of asylum seekers were each detained for several days in the lagoon at Ashmore Island after failed attempts by the Royal Australian Navy to turn them back to Indonesia in October 2001.
https://en.wikipedia.org/wiki?curid=1227
Acoustic theory Acoustic theory is a scientific field that relates to the description of sound waves. It derives from fluid dynamics. See acoustics for the engineering approach. Propagation of sound waves in a fluid (such as water) can be modeled by an equation of continuity (conservation of mass) and an equation of motion (conservation of momentum) . With some simplifications, in particular constant density, they can be given as follows: where formula_2 is the acoustic pressure and formula_3 is the flow velocity vector, formula_4 is the vector of spatial coordinates formula_5, formula_6 is the time, formula_7 is the static mass density of the medium and formula_8 is the bulk modulus of the medium. The bulk modulus can be expressed in terms of the density and the speed of sound in the medium (formula_9) as If the flow velocity field is irrotational, formula_11, then the acoustic wave equation is a combination of these two sets of balance equations and can be expressed as where we have used the vector Laplacian, formula_13 The acoustic wave equation (and the mass and momentum balance equations) are often expressed in terms of a scalar potential formula_14 where formula_15. In that case the acoustic wave equation is written as and the momentum balance and mass balance are expressed as The derivations of the above equations for waves in an acoustic medium are given below. The equations for the conservation of linear momentum for a fluid medium are where formula_19 is the body force per unit mass, formula_20 is the pressure, and formula_21 is the deviatoric stress. If formula_22 is the Cauchy stress, then where formula_24 is the rank-2 identity tensor. We make several assumptions to derive the momentum balance equation for an acoustic medium. These assumptions and the resulting forms of the momentum equations are outlined below. In acoustics, the fluid medium is assumed to be Newtonian. For a Newtonian fluid, the deviatoric stress tensor is related to the flow velocity by where formula_26 is the shear viscosity and formula_27 is the bulk viscosity. Therefore, the divergence of formula_21 is given by Using the identity formula_30, we have The equations for the conservation of momentum may then be written as For most acoustics problems we assume that the flow is irrotational, that is, the vorticity is zero. In that case and the momentum equation reduces to Another frequently made assumption is that effect of body forces on the fluid medium is negligible. The momentum equation then further simplifies to Additionally, if we assume that there are no viscous forces in the medium (the bulk and shear viscosities are zero), the momentum equation takes the form An important simplifying assumption for acoustic waves is that the amplitude of the disturbance of the field quantities is small. This assumption leads to the linear or small signal acoustic wave equation. Then we can express the variables as the sum of the (time averaged) mean field (formula_37) that varies in space and a small fluctuating field (formula_38) that varies in space and time. That is and Then the momentum equation can be expressed as Since the fluctuations are assumed to be small, products of the fluctuation terms can be neglected (to first order) and we have Next we assume that the medium is homogeneous; in the sense that the time averaged variables formula_43 and formula_44 have zero gradients, i.e., The momentum equation then becomes At this stage we assume that the medium is at rest, which implies that the mean flow velocity is zero, i.e., formula_47. Then the balance of momentum reduces to Dropping the tildes and using formula_49, we get the commonly used form of the acoustic momentum equation The equation for the conservation of mass in a fluid volume (without any mass sources or sinks) is given by where formula_52 is the mass density of the fluid and formula_53 is the flow velocity. The equation for the conservation of mass for an acoustic medium can also be derived in a manner similar to that used for the conservation of momentum. From the assumption of small disturbances we have and Then the mass balance equation can be written as If we neglect higher than first order terms in the fluctuations, the mass balance equation becomes Next we assume that the medium is homogeneous, i.e., Then the mass balance equation takes the form At this stage we assume that the medium is at rest, i.e., formula_47. Then the mass balance equation can be expressed as To close the system of equations we need an equation of state for the pressure. To do that we assume that the medium is an ideal gas and all acoustic waves compress the medium in an adiabatic and reversible manner. The equation of state can then be expressed in the form of the differential equation: where formula_63 is the specific heat at constant pressure, formula_64 is the specific heat at constant volume, and formula_65 is the wave speed. The value of formula_66 is 1.4 if the acoustic medium is air. For small disturbances where formula_9 is the speed of sound in the medium. Therefore, The balance of mass can then be written as Dropping the tildes and defining formula_71 gives us the commonly used expression for the balance of mass in an acoustic medium: If we use a cylindrical coordinate system formula_73 with basis vectors formula_74, then the gradient of formula_20 and the divergence of formula_76 are given by where the flow velocity has been expressed as formula_78. The equations for the conservation of momentum may then be written as In terms of components, these three equations for the conservation of momentum in cylindrical coordinates are The equation for the conservation of mass can similarly be written in cylindrical coordinates as The acoustic equations for the conservation of momentum and the conservation of mass are often expressed in time harmonic form (at fixed frequency). In that case, the pressures and the flow velocity are assumed to be time harmonic functions of the form where formula_83 is the frequency. Substitution of these expressions into the governing equations in cylindrical coordinates gives us the fixed frequency form of the conservation of momentum and the fixed frequency form of the conservation of mass In the special case where the field quantities are independent of the z-coordinate we can eliminate formula_86 to get Assuming that the solution of this equation can be written as we can write the partial differential equation as The left hand side is not a function of formula_90 while the right hand side is not a function of formula_91. Hence, where formula_93 is a constant. Using the substitution we have The equation on the left is the Bessel equation, which has the general solution where formula_97 is the cylindrical Bessel function of the first kind and formula_98 are undetermined constants. The equation on the right has the general solution where formula_100 are undetermined constants. Then the solution of the acoustic wave equation is Boundary conditions are needed at this stage to determine formula_102 and the other undetermined constants.
https://en.wikipedia.org/wiki?curid=1234
Alexander Mackenzie (politician) Alexander Mackenzie, (January 28, 1822April 17, 1892) was a Scottish-Canadian politician who served as the second prime minister of Canada, in office from 1873 to 1878. Mackenzie was born in Logierait, Perthshire, Scotland. He left school at the age of 13, following his father's death to help his widowed mother, and trained as a stonemason. Mackenzie immigrated to Canada when he was 19, settling in what became Ontario. His masonry business prospered, allowing him to pursue other interests – such as the editorship of a pro-Reformist of a newspaper called the" Lambton Shield". Mackenzie was elected to the Legislative Assembly of the Province of Canada in 1861, as a supporter of George Brown. In 1867, Mackenzie was elected to the new House of Commons of Canada for the Liberal Party. He became leader of the party (thus Leader of the Opposition) in mid-1873, and a few months later succeeded John A. Macdonald as prime minister, following Macdonald's resignation in the aftermath of the Pacific Scandal. Mackenzie and the Liberals won a clear majority at the 1874 election. He was popular among the general public for his humble background and apparent democratic tendencies. As prime minister, Mackenzie continued the nation-building programme that had been begun by his predecessor. His government established the Supreme Court of Canada and Royal Military College of Canada, and created the District of Keewatin to better administer Canada's newly acquired western territories. However, it made little progress on the transcontinental railway, and struggled to deal with the aftermath of the Panic of 1873. At the 1878 election, Mackenzie's government suffered a landslide defeat. He remained leader of the Liberal Party for another two years, and continued on as a Member of Parliament until his death, due to a stroke. Mackenzie was born on 28 January 1822 in Logierait, Perthshire, Scotland, the son of Mary Stewart (Fleming) and Alexander Mackenzie, Sr (born 1884), who were married in 1817. The site of his birthplace is known as Clais-'n-deoir "The Hollow of the Weeping", where families said their goodbyes as the convicted were led to nearby Gallows Hill. The house in which he was born was built by his father and is still standing in 2019. He was the third of 10 boys, seven of whom survived infancy. Alexander Mackenzie, Sr., was a carpenter and ship's joiner who had to move around frequently for work after the end of the Napoleonic Wars in 1815. Mackenzie's father died on 7 March 1836 and at the age of 13, Alexander Mackenzie, Jr., was thus forced to end his formal education to help support his family. He apprenticed as a stonemason and met his future wife, Helen Neil, in Irvine, where her father was also a stonemason. The Neils were Baptist and shortly thereafter, Mackenzie converted from Presbyterianism to Baptist beliefs. Together with the Neils, he immigrated to Canada in 1842 to seek a better life. Mackenzie's faith was to link him to the increasingly influential temperance cause, particularly strong in Canada West where he lived, a constituency of which he was to represent in the Parliament of Canada. The Neils and Mackenzie settled in Kingston, Ontario. The limestone in the area proved too hard for his stonemason tools, and not having money to buy new tools, Mackenzie took a job as a labourer constructing a building on Princess Street. The contractor on the job claimed financial difficulty, so Mackenzie accepted a promissory note for summer wages. The note later proved to be worthless. Subsequently, Mackenzie won a contract building a bomb-proof arch at Fort Henry. He later became a foreman on the construction of Kingston's four Martello Towers - Murney Tower, Fort Frederick, Cathcart Tower, and Shoal Tower. He was also a foreman on the construction of the Welland Canal and the Lachine Canal. While working on the Beauharnois Canal, a one-ton stone fell and crushed one of his legs. He recovered, but never regained the strength in that leg. While in Kingston, Mackenzie became a vocal opponent of religious and political entitlement and corruption in government. Mackenzie married Helen Neil (1826–52) in 1845 and with her had three children, with only one girl, Mary, surviving infancy. Helen and he moved to Sarnia, Ontario (known as Canada West) in 1847 and Mary was born in 1848. They were soon joined from Scotland by the rest of Mackenzie's brothers and his mother. He began working as a general contractor, earning a reputation for being a hard-working, honest man, as well as having a working man's view on fiscal policy. Mackenzie helped construct many courthouses and jails across southern Ontario. A number of these still stand today, including the Sandwich Courthouse and Jail now known as the Mackenzie Hall Cultural Centre in Windsor, Ontario, and the Kent County Courthouse and Jail in Chatham, Ontario. He even bid, unsuccessfully, on the construction of the Parliament buildings in Ottawa in 1859. Helen died in 1852, finally succumbing to the effects of excessive doses of mercury-based calomel used to treat a fever while in Kingston. In 1853, he married Jane Sym (1825–93). Mackenzie involved himself in politics almost from the moment he arrived in Canada. He fought passionately for equality and the elimination of all forms of class distinction. In 1851, he became the secretary for the Reform Party for Lambton. After convincing him to run in Kent/Lambton, Mackenzie campaigned relentlessly for George Brown, owner of the Reformist paper "The Globe" in the 1851 election, helping Brown to win his first seat in the Legislative Assembly. Mackenzie and Brown remained the closest of friends and colleagues for the rest of their lives. In 1852, Mackenzie became editor of another reformist paper, the "Lambton Shield". As editor, Mackenzie was perhaps a little too vocal, leading the paper to a lawsuit for libel against the local conservative candidate. Because a key witness claimed Cabinet Confidence and would not testify, the paper lost the suit and was forced to fold due to financial hardship. After his brother, Hope Mackenzie, declined to run, Alexander was petitioned to run and won his first seat in the Legislative Assembly as a supporter of George Brown in 1861. When Brown resigned from the Great Coalition in 1865 over reciprocity negotiations with the United States, Mackenzie was invited to replace him as the president of the council. Wary of Macdonald's motivations and true to his principles, Mackenzie declined. He entered the House of Commons of Canada in 1867, representing the Lambton, Ontario, riding. No cohesive national Liberal Party of Canada existed at the time and with Brown not winning his seat, no official leader emerged. Mackenzie did not believe he was the best qualified for the position, and although he resisted offers of the position, he nevertheless sat as the "de facto" leader of the Official Opposition. When the Macdonald government fell due to the Pacific Scandal in 1873, the Governor General, Lord Dufferin, called upon Mackenzie, who had been chosen as the leader of the Liberal Party a few months earlier, to form a new government. Mackenzie formed a government and asked the Governor General to call an election for January 1874. The Liberals won, having garnered 53.8% of the popular vote. The voter support of 53.8% remains the record in Canada for all federal elections. Mackenzie remained prime minister until the 1878 election when Macdonald's Conservatives returned to power with a majority government. For a man of Mackenzie's humble origins to attain such a position was unusual in an age which generally offered such opportunity only to the privileged. Lord Dufferin expressed early misgivings about a stonemason taking over government, but on meeting Mackenzie, Dufferin revised his opinions: Mackenzie served concurrently as Minister of Public Works and oversaw the completion of the Parliament buildings. While drawing up the plans for the West Block, he included a circular staircase leading directly from his office to the outside of the building, which allowed him to escape the patronage-seekers waiting for him in his ante-chamber. Proving Dufferin's reflections on his character to be true, Mackenzie disliked intensely the patronage inherent in politics. Nevertheless, he found it a necessary evil to maintain party unity and ensure the loyalty of his fellow Liberals. In keeping with his democratic ideals, Mackenzie refused the offer of a knighthood three times, and was thus the only one of Canada's first eight Prime Ministers not to be knighted. He also declined appointment to the UK Privy Council and hence does not bear the title "Right Honourable". His pride in his working class origins never left him. Once, while touring Fort Henry as prime minister, he asked the soldier accompanying him if he knew the thickness of the wall beside them. The embarrassed escort confessed that he didn't and Mackenzie replied, "I do. It is five feet, ten inches. I know, because I built it myself!" As Prime Minister, Alexander Mackenzie strove to reform and simplify the machinery of government, achieving a remarkable record of reform legislation. He introduced the secret ballot; advised the creation of the Supreme Court of Canada; the establishment of the Royal Military College of Canada in Kingston in 1874; and the creation of the Office of the Auditor General in 1878. He completed the Intercolonial Railway, but struggled to progress on the national railway due to a worldwide economic depression, almost coming to blows with the then Governor General Lord Dufferin over imperial interference. Mackenzie stood up for the rights of Canada as a nation and fought for the supremacy of Parliament and honesty in government. Above all else, he was known and loved for his honesty and integrity. However, his term was marked by economic depression that had grown out of the Panic of 1873, which Mackenzie's government was unable to alleviate. In 1874, Mackenzie negotiated a new free trade agreement with the United States, eliminating the high protective tariffs on Canadian goods in US markets. However, this action did not bolster the economy, and construction of the CPR slowed drastically due to lack of funding. In 1876, the Conservative opposition announced a National Policy of protective tariffs, which resonated with voters. When an election was held at the conclusion of Mackenzie's five-year term, the Conservatives were swept back into office in a landslide victory. Mackenzie chose the following jurists to be appointed as justices of the Supreme Court of Canada by the Governor General: After his government's defeat, Mackenzie remained Leader of the Opposition for another two years, until 1880. He was soon struck with a mysterious ailment that sapped his strength and all but took his voice. Sitting in silence, he nevertheless remained an undefeated MP until his death in 1892 from a stroke that resulted from hitting his head during a fall. He died in Toronto and was buried in Lakeview Cemetery in Sarnia, Ontario. Mackenzie's first biography in 1892 referred to him as Canada's Stainless Statesman. He was a devout Baptist and teetotaller who found refuge in, and drew strength from, his family, friends, and faith. He was also a loyal friend and an incorrigible prankster (stuffed chimney on young in-laws; rolled boulder down Thunder Cape towards friend A. McKellar; burned Tory campaign placards in hotel woodstove early in morning). Unpretentious and down to earth, his public official austerity was in striking contrast to private compassion and giving nature. He was the soul of honour and integrity, a proud man who sought no recognition or personal enrichment and accepted gifts reluctantly. He preferred to follow than lead (unreferenced - many times he refused leadership offers) and often found duty outweighed heavy burden of office. He was uncompromising on his principles, perhaps too much so. An historian at the time said, “He was, and ever will remain, the Sir Galahad of Canadian politics.” Very proud of his Scottish heritage, he was forever a Scot “"Nemo me impune lacessit"” (no one attacks me with impunity). The Upper Canada rebellion leader W.L. Mackenzie referred to him, “He is every whit a self-made, self-educated man. Has large mental capacity and indomitable energy.” Canada's Governor General, Lord Dufferin, said of him, he is “as pure as crystal, and as true as steel, with lots of common sense.” A close friend, Chief Justice Sir Louis Davies, said he was “the best debater the House of Commons has ever known.” A friend and colleague in Cabinet who went on to become prime minister of Canada, Sir Wilfrid Laurier, said he was “one of the truest and strongest characters to be met within Canadian history. He was endowed with a warm heart and a copious and rich fancy, though veiled by a somewhat reticent exterior, and he was of friends the most tender and true.” Another friend and colleague, who went on to become premier of Ontario, Sir George Ross, said, “Mackenzie was "sui generis" a debater. His humorous sallies blistered like a blast from a flaming smelter. His sterling honesty is a great heritage, and will keep his memory green to all future generations.” At his eulogy, Rev. Dr. Thomas compared him to the Duke of Wellington, who “stood four square, to all the winds that blow.” Newspaper around the world and in Canada had this to say about him. "The London Times" – the untiring energy, the business-like accuracy, the keen perception and reliable judgment, and above all the inflexible integrity, which marked his private life, he carried without abatement of one jot into his public career. "The Westminster Review" – a man, who although, through failing health and failing voice, he had virtually passed out of public life, yet retained to the last the affectionate veneration of the Canadian people as no other man of the time can be said to have done. The "Charlottetown Patriot" – in all that constitutes the real man, the honest statesman, the true patriot, the warm friend, and sincere Christian, he had few equals. Possessed of a clear intellect, a retentive memory, and a ready command of appropriate words, he was one of the most logical and powerful speakers we have ever heard. The "St. John Telegraph" – he was loved by the people and his political opponents were compelled to respect him even above their own chosen leader. As a statesman, he has had few equals. The "Montreal Star" – it is one of the very foremost architects of the Canadian nationality that we mourn. In the dark days of ’73, Canadians were in a state of panic, distrusting the stability of their newly-built Dominion; no one can tell what would have happened had not the stalwart form of Alexander Mackenzie lifted itself above the screaming, vociferating and denying mass of politicians, and all Canada felt at once, there was a man who could be trusted. The "Toronto Globe" – he was a man who loved the people and fought for their rights against privilege and monopoly in every form. The "Philadelphia Record" – Like Caesar, who twice refused a knightly crown, Alexander Mackenzie refused knighthood three times. Unlike Caesar, he owed his political overthrow to his incorruptible honesty and unswerving integrity. In their 1999 study of the Prime Ministers of Canada, which included the results of a survey of Canadian historians, J. L. Granatstein and Norman Hillmer found that Mackenzie was in 11th place just after John Sparrow David Thompson. The following are named in honour of Alexander Mackenzie:
https://en.wikipedia.org/wiki?curid=1235
Christology In Christianity, Christology (from Greek Χριστός "Khristós" and , "-logia"), translated literally from Greek as "the study of Christ", is a branch of theology that concerns Jesus. Different denominations have different opinions on questions like whether Jesus was human, divine, or both, and as a messiah what his role would be in the freeing of the Jewish people from foreign rulers or in the prophesied Kingdom of God, and in the salvation from what would otherwise be the consequences of sin. The earliest Christian writings gave several titles to Jesus, such as Son of Man, Son of God, Messiah, and Kyrios, which were all derived from the Hebrew scriptures. These terms centered around two opposing themes, namely "Jesus as a preexistent figure who becomes human and then returns to God", versus adoptionism - that Jesus was human who was "adopted" by God at his baptism, crucificion, or resurrection. From the second to the fifth centuries, the relation of what Christians there considered to be the human and divine nature of Christ was a major focus of debates in the early church and at the first seven ecumenical councils. The Council of Chalcedon in 451 issued a formulation of the hypostatic union of the two natures of Christ, one human and one divine, "united with neither confusion nor division". Most of the major branches of Western Christianity and Eastern Orthodoxy subscribe to this formulation, while many branches of Oriental Orthodox Churches reject it, subscribing to miaphysitism. Christology (from Greek Χριστός "Khristós" and , "-logia"), literally "the understanding of Christ," is the study of the nature (person) and work (role in salvation) of Jesus Christ. It studies Jesus Christ's humanity and divinity, and the relation between these two aspects; and the role he plays in salvation. "Ontological Christology" analyzes the nature or being of Jesus Christ. "Functional Christology" analyzes the works of Jesus Christ, while "soteriological Christology" analyzes the "salvific" standpoints of Christology. Several approaches can be distinguished within Christology. The term "Christology from above" or "high Christology" refers to approaches that include aspects of divinity, such as Lord and Son of God, and the idea of the pre-existence of Christ as the "Logos" (the Word), as expressed in the . These approaches interpret the works of Christ in terms of his divinity. According to Pannenberg, Christology from above "was far more common in the ancient Church, beginning with Ignatius of Antioch and the second century Apologists." The term "Christology from below" or "low Christology" refers to approaches that begin with the human aspects and the ministry of Jesus (including the miracles, parables, etc.) and move towards his divinity and the mystery of incarnation. A basic christological teaching is that the person of Jesus Christ is both human and divine. The human and divine natures of Jesus Christ apparently ("prosopic") form a duality, as they coexist within one person ("hypostasis"). There are no direct discussions in the New Testament regarding the dual nature of the Person of Christ as both divine and human, and since the early days of Christianity, theologians have debated various approaches to the understanding of these natures, at times resulting in ecumenical councils, and schisms. Some historical christological doctrines gained broad support. We show them here with simplified summaries; see the linked articles for details. Influential Christologies which were broadly condemned as heretical are: Various church councils, mainly in the 4th and 5th centuries, resolved most of these controversies, making the doctrine of the Trinity orthodox in nearly all branches of Christianity. Among them, only the Dyophysite doctrine was recognized as true and not heretical, belonging to the Christian orthodoxy and deposit of faith. In Christian theology, atonement is the method by which human beings can be reconciled to God through Christ's sacrificial suffering and death. Atonement is the forgiving or pardoning of sin in general and original sin in particular through the suffering, death and resurrection of Jesus, enabling the reconciliation between God and his creation. Due to the influence of Gustaf Aulèn's (1879-1978) "Christus Victor" (1931), the various theories or paradigma's of atonement are often grouped as "classical paradigm," "objective paradigm," and the "subjective paradigm": Other theories are the "embracement theory" and the "shared atonement" theory. The earliest christological reflections were shaped by both the Jewish background of the earliest Christians, and by the Greek world of the eastern Mediterranean in which they operated. The earliest Christian writings give several titles to Jesus, such as Son of Man, Son of God, Messiah, and Kyrios, which were all derived from the Hebrew scriptures. According to Matt Stefon and Hans J. Hillerbrand, Historically in the Alexandrian school of thought (fashioned on the Gospel of John), Jesus Christ is the eternal "Logos" who already possesses unity with the Father before the act of Incarnation. In contrast, the Antiochian school viewed Christ as a single, unified human person apart from his relationship to the divine. The notion of pre-existence is deeply rooted in Jewish thought, and can be found in apocalyptic thought and among the rabbis of Paul's time, but Paul was most influenced by Jewish-Hellenistic wisdom literature, where "'Wisdom' is extolled as something existing before the world and already working in creation. According to Witherington, Paul "subscribed to the christological notion that Christ existed prior to taking on human flesh [,] founding the story of Christ [...] on the story of divine Wisdom." The title Kyrios for Jesus is central to the development of New Testament Christology. In the Septuagint it translates the Tetragrammaton, the holy and unpronounceable Name of God. As such, it closely links Jesus with God - in the same way a verse such as Matthew 28:19, "The Name (singular) of the Father, the Son, and the Holy Ghost". "Kyrios" is also conjectured to be the Greek translation of Aramaic "Mari", which in everyday Aramaic usage was a very respectful form of polite address, which means more than just "Teacher" and was somewhat similar to Rabbi. While the term "Mari" expressed the relationship between Jesus and his disciples during his life, the Greek "Kyrios" came to represent his lordship over the world. The early Christians placed "Kyrios" at the center of their understanding, and from that center attempted to understand the other issues related to the Christian mysteries. The question of the deity of Christ in the New Testament is inherently related to the "Kyrios" title of Jesus used in the early Christian writings and its implications for the absolute lordship of Jesus. In early Christian belief, the concept of "Kyrios" included the pre-existence of Christ, for they believed if Christ is one with God, he must have been united with God from the very beginning. Two fundamentally different Christologies developed in the early Church, namely a "low" or adoptionist Christology, and a "high" or "incarnation Christology." The chronology of the development of these early Christologies is a matter of debate within contemporary scholarship. The "low Christology" or "adoptionist Christology" is the belief "that God exalted Jesus to be his Son by raising him from the dead," thereby raising him to "divine status." According to the "evolutionary model" c.q. "evolutionary theories," the christological understanding of Christ developed over time, as witnessed in the Gospels, with the earliest Christians believing that Jesus was a human who was exalted, c.q. adopted as God's Son, when he was resurrected. Later beliefs shifted the exaltation to his baptism, birth, and subsequently to the idea of his pre-existence, as witnessed in the Gospel of John. This "evolutionary model" was proposed by proponents of the "Religionsgeschichtliche Schule", especially Wilhelm Boussets influential "Kyrios Christos" (1913). This evolutionary model was very influential, and the "low Christology" has long been regarded as the oldest Christology. The other early Christology is "high Christology," which is "the view that Jesus was a pre-existent divine being who became a human, did the Father’s will on earth, and then was taken back up into heaven whence he had originally come," and from where he appeared on earth. According to Bousset, this "high Christology" developed at the time of Paul's writing, under the influence of Gentile Christians, who brought their pagan Hellenistic traditions to the early Christian communities, introducing divine honours to Jesus. According to Casey and Dunn, this "high Christology" developed after the time of Paul, at the end of the first century CE when the Gospel according to John was written. Since the 1970s, these late datings for the development of a "high Christology" have been contested, and a majority of scholars argue that this "High Christology" existed already before the writings of Paul. According to the "New "Religionsgeschichtliche Schule"," c.q. "Early High Christology Club," which includes Martin Hengel, Larry Hurtado, N. T. Wright, and Richard Bauckham, this "incarnation Christology" or "high Christology" did not evolve over a longer time, but was a "big bang" of ideas which were already present at the start of Christianity, and took further shape in the first few decades of the church, as witnessed in the writings of Paul. Some 'Early High Christology' proponents scholars argue that this "High Christology" may go back to Jesus himself. There is a controversy regarding whether Jesus himself claimed to be divine. In "Honest to God", then-Bishop of Woolwich John A. T. Robinson, questioned the idea. John Hick, writing in 1993, mentioned changes in New Testament studies, citing "broad agreement" that scholars do not today support the view that Jesus claimed to be God, quoting as examples Michael Ramsey (1980), C. F. D. Moule (1977), James Dunn (1980), Brian Hebblethwaite (1985) and David Brown (1985). Larry Hurtado, who argues that the followers of Jesus within a very short period developed an exceedingly high level of devotional reverence to Jesus, at the same time rejects the view that Jesus made a claim to messiahship or divinity to his disciples during his life as "naive and ahistorical". According to Gerd Lüdemann, the broad consensus among modern New Testament scholars is that the proclamation of the divinity of Jesus was a development within the earliest Christian communities. N. T. Wright points out that arguments over the claims of Jesus regarding divinity have been passed over by more recent scholarship, which sees a more complex understanding of the idea of God in first century Judaism. But, Andrew Loke argues that if Jesus did not claim and show himself to be truly divine and rise from the dead, the earliest Christian leaders who were devout ancient monotheistic Jews would have regarded Jesus as merely a teacher or a prophet, but not as truly divine, which they did. The study of the various Christologies of the Apostolic Age is based on early Christian documents. The oldest Christian sources are the writings of Paul. The central Christology of Paul conveys the notion of Christ's pre-existence and the identification of Christ as "Kyrios". Both notions already existed before him in the early Christian communities, and Paul deepened them and used them for preaching in the Hellenistic communities. What exactly Paul believed about the nature of Jesus cannot be determined decisively. In Philippians 2, Paul states that Jesus was preexistent and came to Earth "by taking the form of a servant, being made in human likeness." This sounds like an incarnation Christology. In Romans 1:4, however, Paul states that Jesus "was declared with power to be the Son of God by his resurrection from the dead," which sounds like an adoptionistic Christology, where Jesus was a human being who was "adopted" after his death. Different views would be debated for centuries by Christians and finally settled on the idea that he was both fully human and fully divine by the middle of the 5th century in the Council of Ephesus. Paul's thoughts on Jesus' teachings, versus his nature and being, is more defined, in that Paul believed Jesus was sent as an atonement for the sins of everyone. The Pauline epistles use "Kyrios" to identify Jesus almost 230 times, and express the theme that the true mark of a Christian is the confession of Jesus as the true Lord. Paul viewed the superiority of the Christian revelation over all other divine manifestations as a consequence of the fact that Christ is the Son of God. The Pauline epistles also advanced the "cosmic Christology" later developed in the fourth gospel, elaborating the cosmic implications of Jesus' existence as the Son of God, as in : "Therefore, if anyone is in Christ, he is a new creation. The old has passed away; behold, the new has come." Also, in : "He is the image of the invisible God, the firstborn of all creation." The synoptic Gospels date from after the writings of Paul. They provide episodes from the life of Jesus and some of his works, but the authors of the New Testament show little interest in an absolute chronology of Jesus or in synchronizing the episodes of his life, and as in , the Gospels do not claim to be an exhaustive list of his works. Christologies that can be gleaned from the three Synoptic Gospels generally emphasize the humanity of Jesus, his sayings, his parables, and his miracles. The Gospel of John provides a different perspective that focuses on his divinity. The first 14 verses of the Gospel of John are devoted to the divinity of Jesus as the "Logos", usually translated as "Word", along with his pre-existence, and they emphasize the cosmic significance of Christ, e.g. : "All things were made through him, and without him was not any thing made that was made." In the context of these verses, the Word made flesh is identical with the Word who was in the beginning with God, being exegetically equated with Jesus. Following the Apostolic Age, from the second century onwards, a number of controversies developed about how the human and divine are related within the person of Jesus. As of the second century, a number of different and opposing approaches developed among various groups. In contrast to prevailing monoprosopic views on the Person of Christ, alternative dyoprosopic notions were also promoted by some theologians, but such views were rejected by the ecumenical councils. For example, Arianism did not endorse divinity, Ebionism argued Jesus was an ordinary mortal, while Gnosticism held docetic views which argued Christ was a spiritual being who only appeared to have a physical body. The resulting tensions led to schisms within the church in the second and third centuries, and ecumenical councils were convened in the fourth and fifth centuries to deal with the issues. Although some of the debates may seem to various modern students to be over a theological iota, they took place in controversial political circumstances, reflecting the relations of temporal powers and divine authority, and certainly resulted in schisms, among others that separated the Church of the East from the Church of the Roman Empire. In 325, the First Council of Nicaea defined the persons of the Godhead and their relationship with one another, decisions which were ratified at the First Council of Constantinople in 381. The language used was that the one God exists in three persons (Father, Son, and Holy Spirit); in particular, it was affirmed that the Son was "homoousios" (of the same being) as the Father. The Nicene Creed declared the full divinity and full humanity of Jesus. After the First Council of Nicaea in 325 the Logos and the second Person of the Trinity were being used interchangeably. In 431, the First Council of Ephesus was initially called to address the views of Nestorius on Mariology, but the problems soon extended to Christology, and schisms followed. The 431 council was called because in defense of his loyal priest Anastasius, Nestorius had denied the "Theotokos" title for Mary and later contradicted Proclus during a sermon in Constantinople. Pope Celestine I (who was already upset with Nestorius due to other matters) wrote about this to Cyril of Alexandria, who orchestrated the council. During the council, Nestorius defended his position by arguing there must be two persons of Christ, one human, the other divine, and Mary had given birth only to a human, hence could not be called the "Theotokos", i.e. "the one who gives birth to God". The debate about the single or dual nature of Christ ensued in Ephesus. The First Council of Ephesus debated miaphysitism (two natures united as one after the hypostatic union) versus dyophysitism (coexisting natures after the hypostatic union) versus monophysitism (only one nature) versus Nestorianism (two hypostases). From the christological viewpoint, the council adopted "Mia Physis (But being made one κατὰ φύσιν)" - Council of Ephesus, Epistle of Cyril to Nestorius, i.e. One Nature of the Word of God Incarnate (μία φύσις τοῦ θεοῦ λόγου σεσαρκωμένη mía phýsis toû theoû lógou sesarkōménē). In 451, the Council of Chalcedon affirmed dyophysitism. The Oriental Orthodox rejected this and subsequent councils and continued to consider themselves as "miaphysite" according to the faith put forth at the Councils of Nicaea and Ephesus. The council also confirmed the "Theotokos" title and excommunicated Nestorius. The 451 Council of Chalcedon was highly influential, and marked a key turning point in the christological debates. It is the last council which many Anglicans and most Protestants consider ecumenical. The Council of Chalcedon fully promulgated the Western dyophysite understanding put forth by Pope Leo I of Rome of the "hypostatic union", the proposition that Christ has one human nature "[physis]" and one divine nature "[physis]", each distinct and complete, and united with neither confusion nor division. Most of the major branches of Western Christianity (Roman Catholicism, Anglicanism, Lutheranism, and Reformed) and Eastern Orthodoxy subscribe to the Chalcedonian Christological formulation, while many branches of Oriental Orthodox Churches (Syrian Orthodoxy, Assyrian Church, Coptic Orthodoxy, Ethiopian Orthodoxy, and Armenian Apostolicism) reject it. Although the Chalcedonian Creed did not put an end to all christological debate, it did clarify the terms used and became a point of reference for many future Christologies. But it also broke apart the church of the Eastern Roman Empire in the fifth century, and unquestionably established the primacy of Rome in the East over those who accepted the Council of Chalcedon. This was reaffirmed in 519, when the Eastern Chalcedonians accepted the Formula of Hormisdas, anathematizing all of their own Eastern Chalcedonian hierarchy, who died out of communion with Rome from 482–519. The Second Council of Constantinople in 553 interpreted the decrees of Chalcedon, and further explained the relationship of the two natures of Jesus. It also condemned the alleged teachings of Origen on the pre-existence of the soul, and other topics. The Third Council of Constantinople in 681 declared that Christ has two wills of his two natures, human and divine, contrary to the teachings of the Monothelites, with the divine will having precedence, leading and guiding the human will. The Second Council of Nicaea was called under the Empress Regent Irene of Athens in 787, known as the second of Nicaea. It supports the veneration of icons while forbidding their worship. It is often referred to as "The Triumph of Orthodoxy". The term "monastic Christology" has been used to describe spiritual approaches developed by Anselm of Canterbury, Peter Abelard and Bernard of Clairvaux. The Franciscan piety of the 12th and 13th centuries led to "popular Christology". Systematic approaches by theologians, such as Thomas Aquinas, are called "scholastic Christology". In the 13th century, Saint Thomas Aquinas provided the first systematic Christology that consistently resolved a number of the existing issues. In his Christology from above, Aquinas also championed the principle of perfection of Christ's human attributes. The Middle Ages also witnessed the emergence of the "tender image of Jesus" as a friend and a living source of love and comfort, rather than just the "Kyrios" image. John Calvin maintained there was no human element in the Person of Christ which could be separated from the Person of The Word. Calvin also emphasized the importance of the "Work of Christ" in any attempt at understanding the Person of Christ and cautioned against ignoring the Works of Jesus during his ministry. The 19th century saw the rise of Liberal Protestant theology, which questioned the dogmatic foundations of Christianity, and approached the Bible with critical-historical tools. The divinity of Jesus was problematized, and replaced with an emphasis on the ethical aspects of his teachings. Catholic theologian Karl Rahner sees the purpose of modern Christology as to formulate the Christian belief that "God became man and that God-made-man is the individual Jesus Christ" in a manner that this statement can be understood consistently, without the confusions of past debates and mythologies. Rahner pointed out the coincidence between the Person of Christ and the Word of God, referring to and which state whoever is ashamed of the words of Jesus is ashamed of the Lord himself. Hans von Balthasar argued the union of the human and divine natures of Christ was achieved not by the "absorption" of human attributes, but by their "assumption". Thus, in his view, the divine nature of Christ was not affected by the human attributes and remained forever divine. The Nativity of Jesus impacted the christological issues about his Person from the earliest days of Christianity. Luke's Christology centers on the dialectics of the dual natures of the earthly and heavenly manifestations of existence of the Christ, while Matthew's Christology focuses on the mission of Jesus and his role as the savior. The salvific emphasis of later impacted the theological issues and the devotions to Holy Name of Jesus. The accounts of the crucifixion and subsequent resurrection of Jesus provides a rich background for christological analysis, from the canonical Gospels to the Pauline Epistles. A central element in the christology presented in the Acts of the Apostles is the affirmation of the belief that the death of Jesus by crucifixion happened "with the foreknowledge of God, according to a definite plan". In this view, as in , the cross is not viewed as a scandal, for the crucifixion of Jesus "at the hands of the lawless" is viewed as the fulfilment of the plan of God. Paul's Christology has a specific focus on the death and resurrection of Jesus. For Paul, the crucifixion of Jesus is directly related to his resurrection and the term "the cross of Christ" used in Galatians 6:12 may be viewed as his abbreviation of the message of the gospels. For Paul, the crucifixion of Jesus was not an isolated event in history, but a cosmic event with significant eschatological consequences, as in Cor 2:8. In the Pauline view, Jesus, obedient to the point of death (Phil 2:8), died "at the right time" (Rom 4:25) based on the plan of God. For Paul, the "power of the cross" is not separable from the resurrection of Jesus. The threefold office (Latin "munus triplex") of Jesus Christ is a Christian doctrine based upon the teachings of the Old Testament. It was described by Eusebius and more fully developed by John Calvin. It states that Jesus Christ performed three functions (or "offices") in his earthly ministry – those of prophet (Deuteronomy 18:14–22), priest (Psalm 110:1-4), and king (Psalm 2). In the Old Testament, the appointment of someone to any of these three positions could be indicated by anointing him or her by pouring oil over the head. Thus, the term messiah, meaning "anointed one", is associated with the concept of the threefold office. While the office of king is that most frequently associated with the Messiah, the role of Jesus as priest is also prominent in the New Testament, being most fully explained in chapters 7 to 10 of the Book of Hebrews. Some Christians, notably Roman Catholics, view Mariology as a key component of Christology. In this view, not only is Mariology a logical and necessary consequence of Christology, but without it, Christology is incomplete, since the figure of Mary contributes to a fuller understanding of who Christ is and what he did. Protestants have criticized Mariology because many of its assertions lack any biblical foundation. Strong Protestant reaction against Roman Catholic Marian devotion and teaching has been a significant issue for ecumenical dialogue. Joseph Cardinal Ratzinger (later Pope Benedict XVI) expressed this sentiment about Roman Catholic Mariology when in two separate occasions he stated, "The appearance of a truly Marian awareness serves as the touchstone indicating whether or not the christological substance is fully present" and "It is necessary to go back to Mary, if we want to return to the truth about Jesus Christ."
https://en.wikipedia.org/wiki?curid=7355
Complaint In legal terminology, a complaint is any formal legal document that sets out the facts and legal reasons (see: cause of action) that the filing party or parties (the plaintiff(s)) believes are sufficient to support a claim against the party or parties against whom the claim is brought (the defendant(s)) that entitles the plaintiff(s) to a remedy (either money damages or injunctive relief). For example, the Federal Rules of Civil Procedure (FRCP) that govern civil litigation in United States courts provide that a civil action is commenced with the filing or service of a pleading called a complaint. Civil court rules in states that have incorporated the Federal Rules of Civil Procedure use the same term for the same pleading. In Civil Law, a “complaint” is the first formal action taken to officially begin a lawsuit. This written document contains the allegations against the defense, the specific laws violated, the facts that led to the dispute, and any demands made by the plaintiff to restore justice. In some jurisdictions, specific types of criminal cases may also be commenced by the filing of a complaint, also sometimes called a criminal complaint or felony complaint. Most criminal cases are prosecuted in the name of the governmental authority that promulgates criminal statutes and enforces the police power of the state with the goal of seeking criminal sanctions, such as the State (also sometimes called the People) or Crown (in Commonwealth realms). In the United States, the complaint is often associated with misdemeanor criminal charges presented by the prosecutor without the grand jury process. In most U.S. jurisdictions, the charging instrument presented to and authorized by a grand jury is referred to as an indictment. Virtually every U.S. state has some forms available on the web for most common complaints for lawyers and self-representing litigants; if a petitioner cannot find an appropriate form in their state, they often can modify a form from another state to fit his or her request. Several United States federal courts publish general guidelines for the petitioners and Civil Rights complaint forms. A complaint generally has the following structural elements: After the complaint has been filed with the court, it has to be properly served to the opposite parties, but usually petitioners are not allowed to serve the complaint personally. The court also can issue a summons – an official summary document which the plaintiff needs to have served together with the complaint. The defendants have limited time to respond, depending on the State or Federal rules. A defendant's failure to answer a complaint can result in a default judgment in favor of the petitioner. For example, in United States federal courts, any person who is at least 18 years old and not a party may serve a summons and complaint in a civil case. The defendant must submit an answer within 21 days after being served with the summons and complaint, or request a waiver, according to FRCP Rule 12. After the civil complaint has been served to the defendants, the plaintiff must, as soon as practicable initiate a conference between the parties to plan for the rest of the discovery process and then the parties should submit a proposed discovery plan to the judge within 14 days after the conference. In many U.S. jurisdictions, a complaint submitted to a court must be accompanied by a Case Information Statement, which sets forth specific key information about the case and the lawyers representing the parties. This allows the judge to make determinations about which deadlines to set for different phases of the case, as it moves through the court system. There are also freely accessible web search engines to assist parties in finding court decisions that can be cited in the complaint as an example or analogy to resolve similar questions of law. Google Scholar is the biggest database of full text state and federal courts decisions that can be accessed without charge. These web search engines often allow one to select specific state courts to search. Federal courts created the Public Access to Court Electronic Records (PACER) system to obtain case and docket information from the United States district courts, United States courts of appeals, and United States bankruptcy courts. The system is managed by the Administrative Office of the United States Courts; it allows lawyers and self-represented clients to obtain documents entered in the case much faster than regular mail. In addition to Federal Rules of Civil Procedure, many of the U.S. district courts have developed their own requirements included in Local Rules for filing with the Court. Local Rules can set up a limit on the number of pages, establish deadlines for motions and responses, explain whether it is acceptable to combine a motion petition with a response, specify if a judge needs an additional copy of the documents (called "judge’s copy"), etc. Local Rules can define page layout elements like: margins, text font/size, distance between lines, mandatory footer text, page numbering, and provide directions on how the pages need to be bound together – i.e. acceptable fasteners, number and location of fastening holes, etc. If the filed motion does not comply with the Local Rules then the judge can choose to strike the motion completely, or order the party to re-file its motion, or grant a special exception to the Local Rules. According to Federal Rules of Civil Procedure (FRCP) , sensitive text like Social Security number, Taxpayer Identification Number, birthday, bank accounts and children’s names, should be redacted from the filings made with the court and accompanying exhibits, (however, exhibits normally do not need to be attached to the original complaint, but should be presented to Court after the discovery). The redacted text can be erased with black-out or white-out, and the page should have an indication that it was redacted - most often by stamping word "redacted" on the bottom. Alternately, the filing party may ask the court’s permission to file some exhibits completely under seal. A minor's name of the petitions should be replaced with initials. A person making a redacted filing can file an unredacted copy under seal, or the Court can choose to order later that an additional filing be made under seal without redaction. Copies of both redacted and unredacted documents filed with court should be provided to the other parties in the case. Some courts also require that an additional electronic courtesy copy be emailed to the other parties. Before filing the complaint, it is important for plaintiff(s) to remember that Federal courts can impose liability for the prevailing party's attorney fees to the losing party, if the judge considers the case frivolous or for purposes of harassment, even when the case was voluntarily dismissed. In the case of "Fox v. Vice", the U.S. Supreme Court held that reasonable attorneys' fees could be awarded to the defendant under 42 U.S.C. Sec. 1988, but only for costs that the defendant would not have incurred "but for the frivolous claims." Even when there is no actual trial or judgment, if there is only pre-trial motion practice such as motions to dismiss, attorney fee shifting still can be awarded under FRCP Rule 11 when the opposing party files a Motion for Sanctions and the court issue an order identifying the sanctioned conduct and the basis for the sanction. The losing party has a right to appeal any order for sanctions in the higher court. In the state courts, however, each party is generally responsible only for its own attorney fees, with certain exceptions.
https://en.wikipedia.org/wiki?curid=7357
Casimir III the Great Casimir III the Great (; 30 April 1310 – 5 November 1370) reigned as the King of Poland from 1333 to 1370. He was the third son of Władysław I the Elbow-high and Jadwiga of Kalisz, and the last Polish king from the Piast dynasty. Casimir inherited a kingdom weakened by war and made it prosperous and wealthy. He reformed the Polish army and doubled the size of the kingdom. He reformed the judicial system and introduced a legal code, gaining the title "the Polish Justinian". Casimir built extensively and founded the University of Kraków, the oldest Polish university. He also confirmed privileges and protections previously granted to Jews and encouraged them to settle in Poland in great numbers. Casimir left no lawful male heir to his throne, producing only daughters. When he died in 1370 from an injury received while hunting, his nephew, King Louis I of Hungary, succeeded him as king of Poland in personal union with Hungary. He was born on 30 April 1310 in Kowal, Kuyavia. He had two brothers who died in infancy and three sisters: Kunegunda, Elżbieta, and Jadwiga. When Casimir attained the throne in 1333, his position was in danger, as his neighbours did not recognise his title and instead called him "king of Kraków". The kingdom was depopulated and exhausted by war, and the economy was ruined. In 1335, in the Treaty of Trentschin, Casimir was forced to relinquish his claims to Silesia "in perpetuity". Casimir rebuilt and his kingdom became prosperous and wealthy, with great prospects for the future. He waged many victorious wars and doubled the size of the kingdom, mostly through addition of lands in modern-day Ukraine (then called the Duchy of Halych). Casimir built extensively during his reign, ordering the construction of over 40 castles, including many castles along the Trail of the Eagle's Nests, and he reformed the Polish army. At the Sejm in Wiślica, on 11 March 1347, Casimir introduced reforms to the Polish judicial system and sanctioned civil and criminal codes for Great and Lesser Poland, earning the title "the Polish Justinian". He founded the University of Kraków, the oldest Polish University, and he organized a meeting of kings in Kraków in 1364 at which he exhibited the wealth of the Polish kingdom. Casimir is the only king in Polish history to both receive and retain the title of "Great", as Bolesław I is more commonly known as "the Brave". In 1355, in Buda, Casimir designated his nephew Louis I of Hungary as his successor should he produce no male heir, just as his father had with Charles I of Hungary to gain help against Bohemia. In exchange Casimir gained a favourable Hungarian attitude, needed in disputes with the hostile Teutonic Order and the Kingdom of Bohemia. At the time Casimir was 45 years old, and so producing a son did not seem unreasonable (he already had a few children). Casimir left no legal son, however, begetting five daughters instead. He tried to adopt his grandson, Casimir IV, Duke of Pomerania, in his last will. The child had been born to his second daughter, Elisabeth, Duchess of Pomerania, in 1351. This part of the testament was invalidated by Louis I of Hungary, however, who had traveled to Kraków quickly after Casimir died (in 1370) and bribed the nobles with future privileges. Casimir III also had a son-in-law, Louis VI of Bavaria, Margrave and Prince-elector of Brandenburg, who was considered a possible successor, but he was deemed ineligible as his wife, Casimir's daughter Cunigunde, had died in 1357 without issue. Thus King Louis I of Hungary became successor in Poland. Louis was proclaimed king upon Casimir's death in 1370, though Casimir's sister Elisabeth (Louis's mother) held much of the real power until her death in 1380. Casimir was facetiously named "the Peasants' King". He introduced the codes of law of Greater and Lesser Poland as an attempt to end the overwhelming superiority of the nobility. During his reign all three major classes — the nobility, priesthood, and bourgeoisie — were more or less counterbalanced, allowing Casimir to strengthen his monarchic position. He was known for siding with the weak when the law did not protect them from nobles and clergymen. He reportedly even supported a peasant whose house had been demolished by his own mistress, after she had ordered it to be pulled down because it disturbed her enjoyment of the beautiful landscape. On 9 October 1334, Casimir confirmed the privileges granted to Jews in 1264 by Bolesław V the Chaste. Under penalty of death, he prohibited the kidnapping of Jewish children for the purpose of enforced Christian baptism, and he inflicted heavy punishment for the desecration of Jewish cemeteries. While Jews had lived in Poland since before his reign, Casimir allowed them to settle in Poland in great numbers and protected them as "people of the king". Casimir's legendary Jewish mistress Esterka remains unconfirmed by direct historical evidence. Casimir III was married four times: On 30 April or 16 October 1325, Casimir married Aldona of Lithuania. She was also known as Anna, possibly a baptismal name. She was a daughter of Grand Duke Gediminas of Lithuania and Jewna. They had two children: Aldona died on 26 May 1339. Casimir remained a widower for two years. On 29 September 1341, Casimir married his second wife, Adelaide of Hesse. She was a daughter of Henry II, Landgrave of Hesse, and Elizabeth of Meissen. They had no children. Casimir started living separately from Adelaide soon after the marriage. Their loveless marriage lasted until 1356, when he declared himself divorced. After Casimir "divorced" Adelaide he married his mistress Christina Rokiczana, the widow of Miklusz Rokiczani, a wealthy merchant. Her own origins are unknown. Following the death of her first husband she had entered the court of Bohemia in Prague as a lady-in-waiting. Casimir brought her with him from Prague and convinced the abbot of the Benedictine abbey of Tyniec to marry them. The marriage was held in a secret ceremony but soon became known. Queen Adelaide renounced it as bigamous and returned to Hesse. Casimir continued living with Christine despite complaints by Pope Innocent VI on behalf of Queen Adelaide. This marriage lasted until 1363–64 when Casimir again declared himself divorced. They had no children. In about 1365, Casimir married his fourth wife Hedwig of Żagań. She was a daughter of Henry V of Iron, Duke of Żagań and Anna of Mazovia. They had three children: As Adelheid was still alive (and possibly Christina as well), the marriage to Hedwig was also considered bigamous. Because of this, the legitimacy of his three young daughters was disputed. Casimir managed to have Anna and Kunigunde legitimated by Pope Urban V on 5 December 1369. Jadwiga the younger was legitimated by Pope Gregory XI on 11 October 1371 (after Casimir's death). Casimir's full title was: "Casimir by the grace of God king of Poland and Russia (Ruthenia), lord and heir of the land of Kraków, Sandomierz, Sieradz, Łęczyca, Kuyavia, Pomerania (Pomerelia)". The title in Latin was: "Kazimirus, Dei gratia rex Polonie et Russie, nec non Cracovie, Sandomirie, Siradie, Lancicie, Cuiavie, et Pomeranieque Terrarum et Ducatuum Dominus et Heres."
https://en.wikipedia.org/wiki?curid=7362
Complexity Complexity characterises the behaviour of a system or model whose components interact in multiple ways and follow local rules, meaning there is no reasonable higher instruction to define the various possible interactions. The term is generally used to characterize something with many parts where those parts interact with each other in multiple ways, culminating in a higher order of emergence greater than the sum of its parts. The study of these complex linkages at various scales is the main goal of complex systems theory. Science takes a number of approaches to characterizing complexity; Zayed "et al." reflect many of these. Neil Johnson states that "even among scientists, there is no unique definition of complexity – and the scientific notion has traditionally been conveyed using particular examples..." Ultimately Johnson adopts the definition of "complexity science" as "the study of the phenomena which emerge from a collection of interacting objects". Definitions of complexity often depend on the concept of a "system" – a set of parts or elements that have relationships among them differentiated from relationships with other elements outside the relational regime. Many definitions tend to postulate or assume that complexity expresses a condition of numerous elements in a system and numerous forms of relationships among the elements. However, what one sees as complex and what one sees as simple is relative and changes with time. Warren Weaver posited in 1948 two forms of complexity: disorganized complexity, and organized complexity. Phenomena of 'disorganized complexity' are treated using probability theory and statistical mechanics, while 'organized complexity' deals with phenomena that escape such approaches and confront "dealing simultaneously with a sizable number of factors which are interrelated into an organic whole". Weaver's 1948 paper has influenced subsequent thinking about complexity. The approaches that embody concepts of systems, multiple elements, multiple relational regimes, and state spaces might be summarized as implying that complexity arises from the number of distinguishable relational regimes (and their associated state spaces) in a defined system. Some definitions relate to the algorithmic basis for the expression of a complex phenomenon or model or mathematical expression, as later set out herein. One of the problems in addressing complexity issues has been formalizing the intuitive conceptual distinction between the large number of variances in relationships extant in random collections, and the sometimes large, but smaller, number of relationships between elements in systems where constraints (related to correlation of otherwise independent elements) simultaneously reduce the variations from element independence and create distinguishable regimes of more-uniform, or correlated, relationships, or interactions. Weaver perceived and addressed this problem, in at least a preliminary way, in drawing a distinction between "disorganized complexity" and "organized complexity". In Weaver's view, disorganized complexity results from the particular system having a very large number of parts, say millions of parts, or many more. Though the interactions of the parts in a "disorganized complexity" situation can be seen as largely random, the properties of the system as a whole can be understood by using probability and statistical methods. A prime example of disorganized complexity is a gas in a container, with the gas molecules as the parts. Some would suggest that a system of disorganized complexity may be compared with the (relative) simplicity of planetary orbits – the latter can be predicted by applying Newton's laws of motion. Of course, most real-world systems, including planetary orbits, eventually become theoretically unpredictable even using Newtonian dynamics; as discovered by modern chaos theory. Organized complexity, in Weaver's view, resides in nothing else than the non-random, or correlated, interaction between the parts. These correlated relationships create a differentiated structure that can, as a system, interact with other systems. The coordinated system manifests properties not carried or dictated by individual parts. The organized aspect of this form of complexity vis-a-vis to other systems than the subject system can be said to "emerge," without any "guiding hand". The number of parts does not have to be very large for a particular system to have emergent properties. A system of organized complexity may be understood in its properties (behavior among the properties) through modeling and simulation, particularly modeling and simulation with computers. An example of organized complexity is a city neighborhood as a living mechanism, with the neighborhood people among the system's parts. There are generally rules which can be invoked to explain the origin of complexity in a given system. The source of disorganized complexity is the large number of parts in the system of interest, and the lack of correlation between elements in the system. In the case of self-organizing living systems, usefully organized complexity comes from beneficially mutated organisms being selected to survive by their environment for their differential reproductive ability or at least success over inanimate matter or less organized complex organisms. See e.g. Robert Ulanowicz's treatment of ecosystems. Complexity of an object or system is a relative property. For instance, for many functions (problems), such a computational complexity as time of computation is smaller when multitape Turing machines are used than when Turing machines with one tape are used. Random Access Machines allow one to even more decrease time complexity (Greenlaw and Hoover 1998: 226), while inductive Turing machines can decrease even the complexity class of a function, language or set (Burgin 2005). This shows that tools of activity can be an important factor of complexity. In several scientific fields, "complexity" has a precise meaning: Other fields introduce less precisely defined notions of complexity: Complexity has always been a part of our environment, and therefore many scientific fields have dealt with complex systems and phenomena. From one perspective, that which is somehow complex – displaying variation without being random – is most worthy of interest given the rewards found in the depths of exploration. The use of the term complex is often confused with the term complicated. In today's systems, this is the difference between myriad connecting "stovepipes" and effective "integrated" solutions. This means that complex is the opposite of independent, while complicated is the opposite of simple. While this has led some fields to come up with specific definitions of complexity, there is a more recent movement to regroup observations from different fields to study complexity in itself, whether it appears in anthills, human brains, or stock markets, social systems. One such interdisciplinary group of fields is relational order theories. The behavior of a complex system is often said to be due to emergence and self-organization. Chaos theory has investigated the sensitivity of systems to variations in initial conditions as one cause of complex behaviour. Recent developments around artificial life, evolutionary computation and genetic algorithms have led to an increasing emphasis on complexity and complex adaptive systems. In social science, the study on the emergence of macro-properties from the micro-properties, also known as macro-micro view in sociology. The topic is commonly recognized as social complexity that is often related to the use of computer simulation in social science, i.e.: computational sociology. Systems theory has long been concerned with the study of complex systems (in recent times, "complexity theory" and "complex systems" have also been used as names of the field). These systems are present in the research of a variety disciplines, including biology, economics, social studies and technology. Recently, complexity has become a natural domain of interest of real world socio-cognitive systems and emerging systemics research. Complex systems tend to be high-dimensional, non-linear, and difficult to model. In specific circumstances, they may exhibit low-dimensional behaviour. In information theory, algorithmic information theory is concerned with the complexity of strings of data. Complex strings are harder to compress. While intuition tells us that this may depend on the codec used to compress a string (a codec could be theoretically created in any arbitrary language, including one in which the very small command "X" could cause the computer to output a very complicated string like "18995316"), any two Turing-complete languages can be implemented in each other, meaning that the length of two encodings in different languages will vary by at most the length of the "translation" language – which will end up being negligible for sufficiently large data strings. These algorithmic measures of complexity tend to assign high values to random noise. However, those studying complex systems would not consider randomness as complexity. Information entropy is also sometimes used in information theory as indicative of complexity, but entropy is also high for randomness. Information fluctuation complexity, fluctuations of information about entropy, does not consider randomness to be complex and has been useful in many applications. Recent work in machine learning has examined the complexity of the data as it affects the performance of supervised classification algorithms. Ho and Basu present a set of complexity measures for binary classification problems. The complexity measures broadly cover: Instance hardness is a bottom-up approach that first seeks to identify instances that are likely to be misclassified (or, in other words, which instances are the most complex). The characteristics of the instances that are likely to be misclassified are then measured based on the output from a set of hardness measures. The hardness measures are based on several supervised learning techniques such as measuring the number of disagreeing neighbors or the likelihood of the assigned class label given the input features. The information provided by the complexity measures has been examined for use in meta learning to determine for which data sets filtering (or removing suspected noisy instances from the training set) is the most beneficial and could be expanded to other areas. A recent study based on molecular simulations and compliance constants describes molecular recognition as a phenomenon of organisation. Even for small molecules like carbohydrates, the recognition process can not be predicted or designed even assuming that each individual hydrogen bond's strength is exactly known. Computational complexity theory is the study of the complexity of problems – that is, the difficulty of solving them. Problems can be classified by complexity class according to the time it takes for an algorithm – usually a computer program – to solve them as a function of the problem size. Some problems are difficult to solve, while others are easy. For example, some difficult problems need algorithms that take an exponential amount of time in terms of the size of the problem to solve. Take the travelling salesman problem, for example. It can be solved in time formula_1 (where "n" is the size of the network to visit – the number of cities the travelling salesman must visit exactly once). As the size of the network of cities grows, the time needed to find the route grows (more than) exponentially. Even though a problem may be computationally solvable in principle, in actual practice it may not be that simple. These problems might require large amounts of time or an inordinate amount of space. Computational complexity may be approached from many different aspects. Computational complexity can be investigated on the basis of time, memory or other resources used to solve the problem. Time and space are two of the most important and popular considerations when problems of complexity are analyzed. There exist a certain class of problems that although they are solvable in principle they require so much time or space that it is not practical to attempt to solve them. These problems are called intractable. There is another form of complexity called hierarchical complexity. It is orthogonal to the forms of complexity discussed so far, which are called horizontal complexity.
https://en.wikipedia.org/wiki?curid=7363
Chastity Chastity is a virtue related to temperance, which is one of the seven Christian virtues, and it is defined as refraining from unreasonable sexual conduct or romantic relationships. Chastity is usually defined within the moral standards and guidelines of a culture, civilization or religion. The term is closely associated (and can be used interchangeably) with sexual abstinence, especially in the context of premarital and extramarital sex. The words "chaste" and "chastity" stem from the Latin adjective ("pure"). The words entered the English language around the middle of the 13th century. At that time, they meant slightly different things. "Chaste" meant "virtuous or pure from unlawful sexual intercourse" (referring to extramarital sex), while "chastity" meant "virginity". Not until the late 16th century did the two words come to have the same basic meaning as a related adjective and noun. Thomas Aquinas links ("chastity") to the Latin verb ("chastise, reprimand, correct"), with a reference to Aristotle's Nicomachean Ethics: "Chastity takes its name from the fact that reason "chastises" concupiscence, which, like a child, needs curbing, as the Philosopher states". For many Jews, Muslims, and Christians, acts of sexual nature are restricted to marriage. For unmarried persons, chastity is identified with sexual abstinence. Sexual acts outside or apart from marriage, such as adultery, fornication, and prostitution, are considered immoral due to lust. In many Christian traditions, chastity is synonymous with sexual purity. Chastity means not having any sexual relation outside marriage. It also means fidelity to husband or wife during marriage. In Catholic morality, chastity is placed opposite the deadly sin of lust, and is classified as one of seven virtues. The moderation of sexual desires is required to be virtuous. Reason, will, and desire can harmoniously work together to do what is good. In marriage, the spouses commit to a lifelong relationship that excludes sexual intimacy with other persons. A third form of chastity, often called "vidual chastity", is expected of a woman for a period after her husband dies. For example, Anglican Bishop Jeremy Taylor defined 5 rules in "Holy Living" (1650), including abstaining from marrying "so long as she is with child by her former husband" and "within the year of mourning". The particular ethical system may not prescribe each of these. For example, Roman Catholics view sex within marriage as chaste, but prohibit the use of artificial contraception as an offense against chastity, seeing contraception as unnatural, contrary to God's will and design of human sexuality. Many Anglican communities allow for artificial contraception, seeing the restriction of family size by artificial contraception as possibly not contrary to God's will. A stricter view is held by the Shakers, who prohibit marriage (and sexual intercourse under any circumstances) as a violation of chastity. The Catholic Church has set up various rules regarding clerical celibacy, while most Protestant communities allow clergy to marry. Celibacy is required of monastics—monks, nuns and friars—even in a rare system of double cloisters, in which husbands could enter the (men's) monastery while their wives entered a (women's) sister monastery. In 306, the Synod of Elvira proscribed clergy from marrying. This was unevenly enforced until the Second Lateran Council in 1139. Then it became church policy. It was not uniformly enforced among the clergy until 200 years later. Eastern Catholic priests are permitted to marry, provided they do so before ordination and outside monastic life. "Vows of chastity" can also be taken by laypersons, either as part of an organised religious life (such as Roman Catholic Beguines and Beghards in the past) or on an individual basis: as a voluntary act of devotion, or as part of an ascetic lifestyle (often devoted to contemplation), or both. Some Protestant religious communities, such as the Bruderhof, take vows of chastity as part of the church membership process. The voluntary aspect has led it to being included among the main counsels of perfection. Chastity is a central and pivotal concept in Roman Catholic praxis. Chastity's importance in traditional Roman Catholic teaching stems from the fact that it is regarded as essential in maintaining and cultivating the unity of body with spirit and thus the integrity of the human being. It is also regarded as fundamental to the practise of the Catholic life because it involves an "apprenticeship in self-mastery". By attaining mastery over one's passions, reason, will, and desire can harmoniously work together to do what is good. In The Church of Jesus Christ of Latter-day Saints (LDS, not evangelical church), chastity is very important, quoting: "Physical intimacy between husband and wife is a beautiful and sacred part of God's plan for His children. It is an expression of love within marriage and allows husband and wife to participate in the creation of life. God has commanded that this sacred power be expressed only between a man and a woman who are legally married. The law of chastity applies to both men and women. It includes strict abstinence from sexual relations before marriage and complete fidelity and loyalty to one's spouse after marriage. "The law of chastity requires that sexual relations be reserved for marriage between a man and a woman. "In addition to reserving sexual intimacy for marriage, we obey the law of chastity by controlling our thoughts, words, and actions. Jesus Christ taught,"Ye have heard that it was said by them of old time, Thou shalt not commit adultery: but I say unto you, That whosoever looketh on a woman to lust after her hath committed adultery with her already in his heart" (Matthew 5:27–28)." LDS teaching also includes that sexual expression within marriage is an important dimension of spousal bonding apart from, but not necessarily avoiding its procreative result. Chastity is mandatory in Islam. Sex outside legitimacy is prohibited, for both men and women whether married or unmarried. The most famous personal example of chastity in the Quran is Virgin Mary (Mariam) : "And [mention] the one who guarded her chastity, so We blew into her [garment] through Our angel [Gabriel], and We made her and her son a sign for the worlds." (21 :91) "And she took, in seclusion from them, a screen. Then We sent to her Our Angel, and he represented himself to her as a well-proportioned man.She said, "Indeed, I seek refuge in the Most Merciful from you, [so leave me], if you should be fearing of Allah ."He said, "I am only the messenger of your Lord to give you [news of] a pure boy."She said, "How can I have a boy while no man has touched me and I have not been unchaste?"" (19:17-20) Extramarital sex is forbidden and the Quran says: "And do not approach unlawful sexual intercourse. Indeed, it is ever an immorality and is evil as a way." (17:32) "And those who do not invoke with Allah another deity or kill the soul which Allah has forbidden [to be killed], except by right, and do not commit unlawful sexual intercourse. And whoever should do that will meet a penalty.Multiplied for him is the punishment on the Day of Resurrection, and he will abide therein humiliated -Except for those who repent, believe and do righteous work. For them Allah will replace their evil deeds with good. And ever is Allah Forgiving and Merciful." (25:68-70) The injunctions and forbiddings in Islam apply equally to men and women. The legal punishment for adultery is equal for men and women. Social hypocrisy in many societies over history had led to a double standard when considering sin committed by men versus sin committed by women. Society tended to be more lenient and permissive towards men forgiving men for sins not forgivable when women do them. In a list of commendable deeds the Quran says: "Indeed, the Muslim men and Muslim women, the believing men and believing women, the obedient men and obedient women, the truthful men and truthful women, the patient men and patient women, the humble men and humble women, the charitable men and charitable women, the fasting men and fasting women, the men who guard their private parts and the women who do so, and the men who remember Allah often and the women who do so - for them Allah has prepared forgiveness and a great reward." (33:35) Because the sex desire is usually attained before a man is financially capable of marriage, the love to God and mindfulness of Him should be sufficient motive for chastity: "But let them who find not [the means for] marriage abstain [from sexual relations] until Allah enriches them from His bounty. And those who seek a contract [for eventual emancipation] from among whom your right hands possess - then make a contract with them if you know there is within them goodness and give them from the wealth of Allah which He has given you. And do not compel your slave girls to prostitution, if they desire chastity, to seek [thereby] the temporary interests of worldly life. And if someone should compel them, then indeed, Allah is [to them], after their compulsion, Forgiving and Merciful." (24:33) The prophet's prescription to the youth was: Those of you who own the means should marry for this should keep their eyes uncraving and their chastity secure. Those who don't, may practise fasting for it curbs desire. " (Ibn Massoud) Chastity is an attitude and a way of life. In Islam it is both a personal and a social value. A Muslim society should not condone relations entailing or conducive to sexual license. Social patterns and practices calculated to inflame sexual desire are frowned upon by Islam, such incitements to immorality including permissive ideologies, titillating works of art and the failure to inculcate sound moral principles in the young. At the heart of such a view of human sexuality lies the conviction that the notion of personal freedom should never be misconstrued as the freedom to flout God's laws by overstepping the bounds which, in His infinite wisdom, He has set upon the relations of the sexes. Chastity is highly prized in the Bahá'í Faith. Similar to other Abrahamic religions, Bahá'í teachings call for the restriction of sexual activity to that between a wife and husband in Bahá'í marriage, and discourage members from using pornography or engaging in sexually explicit recreational activities. The concept of chastity is extended to include avoidance of alcohol and mind-altering drugs, profanity, and gaudy or immodest attire. Hinduism's view on premarital sex is rooted in its concept of Ashrama (stage) or the stages of life. The first of these stages, known as "Brahmacharya," roughly translates as chastity. Celibacy and chastity are considered the appropriate behavior for both male and female students during this stage, which precedes the stage of the married householder (Grihastha). Sanyasis and Hindu monks or Sadhus are also celibate as part of their ascetic discipline. In Sikhism, premarital or extra-marital sex is strictly forbidden. However, it is encouraged to marry and live as a family unit to provide and nurture children for the perpetual benefit of creation (as opposed to sannyasa or living as a monk, which was, and remains, a common spiritual practice in India). A Sikh is encouraged not to live as a recluse, beggar, monk, nun, celibate, or in any similar vein. Celibacy is a must for all Jain monks and nuns. Chastity (Bhramacharya) is one of the five major vows of Jainism. The general Jain code of ethics requires that one do no harm to any living being in thought, action, or word. Adultery is clearly a violation of a moral agreement with one's spouse, and therefore forbidden, and fornication too is seen as a violation of the state of chastity. The teachings of Buddhism include the Noble Eightfold Path, comprising a division called right action. Under the Five Precepts ethical code, Upāsaka and Upāsikā lay followers should abstain from sexual misconduct, while Bhikkhu and Bhikkhuni monastics should practice strict chastity. The Five Precepts of the Daoist religion include No Sexual Misconduct, which is interpreted as prohibiting extramarital sex for lay practitioners and marriage or sexual intercourse for monks and nuns.
https://en.wikipedia.org/wiki?curid=7366
Cosmic microwave background The cosmic microwave background (CMB, CMBR), in Big Bang cosmology, is electromagnetic radiation as a remnant from an early stage of the universe, also known as "relic radiation". The CMB is faint cosmic background radiation filling all space. It is an important source of data on the early universe because it is the oldest electromagnetic radiation in the universe, dating to the epoch of recombination. With a traditional optical telescope, the space between stars and galaxies (the "background") is completely dark. However, a sufficiently sensitive radio telescope shows a faint background noise, or glow, almost isotropic, that is not associated with any star, galaxy, or other object. This glow is strongest in the microwave region of the radio spectrum. The accidental discovery of the CMB in 1964 by American radio astronomers Arno Penzias and Robert Wilson was the culmination of work initiated in the 1940s, and earned the discoverers the 1978 Nobel Prize in Physics. CMB is landmark evidence of the Big Bang origin of the universe. When the universe was young, before the formation of stars and planets, it was denser, much hotter, and filled with a uniform glow from a white-hot fog of hydrogen plasma. As the universe expanded, both the plasma and the radiation filling it grew cooler. When the universe cooled enough, protons and electrons combined to form neutral hydrogen atoms. Unlike the uncombined protons and electrons, these newly conceived atoms could not scatter the thermal radiation by Thomson scattering, and so the universe became transparent instead of being an opaque fog. Cosmologists refer to the time period when neutral atoms first formed as the "recombination epoch", and the event shortly afterwards when photons started to travel freely through space rather than constantly being scattered by electrons and protons in plasma is referred to as photon decoupling. The photons that existed at the time of photon decoupling have been propagating ever since, though growing fainter and less energetic, since the expansion of space causes their wavelength to increase over time (and wavelength is inversely proportional to energy according to Planck's relation). This is the source of the alternative term "relic radiation". The "surface of last scattering" refers to the set of points in space at the right distance from us so that we are now receiving photons originally emitted from those points at the time of photon decoupling. Precise measurements of the CMB are critical to cosmology, since any proposed model of the universe must explain this radiation. The CMB has a thermal black body spectrum at a temperature of . The spectral radiance dEν/dν peaks at 160.23 GHz, in the microwave range of frequencies, corresponding to a photon energy of about 6.626 ⋅ 10−4 eV. Alternatively, if spectral radiance is defined as dEλ/dλ, then the peak wavelength is 1.063 mm (282 GHz, 1.168 ⋅ 10−3 eV photons). The glow is very nearly uniform in all directions, but the tiny residual variations show a very specific pattern, the same as that expected of a fairly uniformly distributed hot gas that has expanded to the current size of the universe. In particular, the spectral radiance at different angles of observation in the sky contains small anisotropies, or irregularities, which vary with the size of the region examined. They have been measured in detail, and match what would be expected if small thermal variations, generated by quantum fluctuations of matter in a very tiny space, had expanded to the size of the observable universe we see today. This is a very active field of study, with scientists seeking both better data (for example, the Planck spacecraft) and better interpretations of the initial conditions of expansion. Although many different processes might produce the general form of a black body spectrum, no model other than the Big Bang has yet explained the fluctuations. As a result, most cosmologists consider the Big Bang model of the universe to be the best explanation for the CMB. The high degree of uniformity throughout the observable universe and its faint but measured anisotropy lend strong support for the Big Bang model in general and the ΛCDM ("Lambda Cold Dark Matter") model in particular. Moreover, the fluctuations are coherent on angular scales that are larger than the apparent cosmological horizon at recombination. Either such coherence is acausally fine-tuned, or cosmic inflation occurred. The cosmic microwave background radiation is an emission of uniform, black body thermal energy coming from all parts of the sky. The radiation is isotropic to roughly one part in 100,000: the root mean square variations are only 18 µK, after subtracting out a dipole anisotropy from the Doppler shift of the background radiation. The latter is caused by the peculiar velocity of the Sun relative to the comoving cosmic rest frame as it moves at some 369.82 ± 0.11 km/s towards the constellation Leo (galactic longitude 264.021 ± 0.011, galactic latitude 48.253 ± 0.005). The CMB dipole as well as aberration at higher multipoles have been measured, consistent with galactic motion. In the Big Bang model for the formation of the universe, inflationary cosmology predicts that after about 10−37 seconds the nascent universe underwent exponential growth that smoothed out nearly all irregularities. The remaining irregularities were caused by quantum fluctuations in the inflaton field that caused the inflation event. Long before the formation of stars and planets, the early universe was smaller, much hotter and, starting 10−6 seconds after the Big Bang, filled with a uniform glow from its white-hot fog of interacting plasma of photons, electrons, and baryons. As the universe expanded, adiabatic cooling caused the energy density of the plasma to decrease until it became favorable for electrons to combine with protons, forming hydrogen atoms. This recombination event happened when the temperature was around 3000 K or when the universe was approximately 379,000 years old. As photons did not interact with these electrically neutral atoms, the former began to travel freely through space, resulting in the decoupling of matter and radiation. The color temperature of the ensemble of decoupled photons has continued to diminish ever since; now down to , it will continue to drop as the universe expands. The intensity of the radiation also corresponds to black-body radiation at 2.726 K because red-shifted black-body radiation is just like black-body radiation at a lower temperature. According to the Big Bang model, the radiation from the sky we measure today comes from a spherical surface called "the surface of last scattering". This represents the set of locations in space at which the decoupling event is estimated to have occurred and at a point in time such that the photons from that distance have just reached observers. Most of the radiation energy in the universe is in the cosmic microwave background, making up a fraction of roughly of the total density of the universe. Two of the greatest successes of the Big Bang theory are its prediction of the almost perfect black body spectrum and its detailed prediction of the anisotropies in the cosmic microwave background. The CMB spectrum has become the most precisely measured black body spectrum in nature. Density of energy for CMB is () or (400–500 photons/cm3). The cosmic microwave background was first predicted in 1948 by Ralph Alpher and Robert Herman. Alpher and Herman were able to estimate the temperature of the cosmic microwave background to be 5 K, though two years later they re-estimated it at 28 K. This high estimate was due to a mis-estimate of the Hubble constant by Alfred Behr, which could not be replicated and was later abandoned for the earlier estimate. Although there were several previous estimates of the temperature of space, these suffered from two flaws. First, they were measurements of the temperature of space and did not suggest that space was filled with a thermal Planck spectrum. Next, they depend on our being at a special spot at the edge of the Milky Way galaxy and they did not suggest the radiation is isotropic. The estimates would yield very different predictions if Earth happened to be located elsewhere in the universe.
https://en.wikipedia.org/wiki?curid=7376
Comparative law Comparative law is the study of differences and similarities between the law (legal systems) of different countries. More specifically, it involves the study of the different legal "systems" (or "families") in existence in the world, including the common law, the civil law, socialist law, Canon law, Jewish Law, Islamic law, Hindu law, and Chinese law. It includes the description and analysis of foreign legal systems, even where no explicit comparison is undertaken. The importance of comparative law has increased enormously in the present age of internationalism, economic globalization, and democratization. The origins of modern Comparative Law can be traced back to Gottfried Wilhelm Leibniz in 1667 in his Latin-language book "Nova Methodus Discendae Docendaeque Iurisprudentiae" (New Methods of Studying and Teaching Jurisprudence). Chapter 7 (Presentation of Law as the Project for all Nations, Lands and Times) introduces the idea of classifying Legal Systems into several families. Notably, a few years later, Leibniz introduced an idea of Language families. Although every Legal System is unique, Comparative Law through studies of their similarities and differences allows for classification of Legal Systems, wherein Law Families is the basic level of the classification. The main differences between Law Families are found in the source(s) of Law, the role of court precedents, the origin and development of the Legal System. Montesquieu is generally regarded as an early founding figure of comparative law. His comparative approach is obvious in the following excerpt from Chapter III of Book I of his masterpiece, "De l'esprit des lois" (1748; first translated by Thomas Nugent, 1750): Also, in Chapter XI (entitled 'How to compare two different Systems of Laws') of Book XXIX, discussing the French and English systems for punishment of false witnesses, he advises that "to determine which of those systems is most agreeable to reason, we must take them each as a whole and compare them in their entirety." Yet another place where Montesquieu's comparative approach is evident is the following, from Chapter XIII of Book XXIX: The modern founding figure of comparative and anthropological jurisprudence was Sir Henry Maine, a British jurist and legal historian. In his 1861 work "Ancient Law: Its Connection with the Early History of Society, and Its Relation to Modern Ideas", he set out his views on the development of legal institutions in primitive societies and engaged in a comparative discussion of Eastern and Western legal traditions. This work placed comparative law in its historical context and was widely read and influential. The first university course on the subject was established at the University of Oxford in 1869, with Maine taking up the position of professor. Comparative law in the US was brought by a legal scholar fleeing persecution in Germany, Rudolf Schlesinger. Schlesinger eventually became professor of comparative law at Cornell Law School helping to spread the discipline throughout the US. Comparative law is an academic discipline that involves the study of legal systems, including their constitutive elements and how they differ, and how their elements combine into a system. Several disciplines have developed as separate branches of comparative law, including comparative constitutional law, comparative administrative law, comparative civil law (in the sense of the law of torts, contracts, property and obligations), comparative commercial law (in the sense of business organisations and trade), and comparative criminal law. Studies of these specific areas may be viewed as micro- or macro-comparative legal analysis, i.e. detailed comparisons of two countries, or broad-ranging studies of several countries. Comparative civil law studies, for instance, show how the law of private relations is organised, interpreted and used in different systems or countries. The purposes of comparative law are: Comparative law is different from general jurisprudence (i.e. legal theory) and from public and private international law. However, it helps inform all of these areas of normativity. For example, comparative law can help international legal institutions, such as those of the United Nations System, in analyzing the laws of different countries regarding their treaty obligations. Comparative law would be applicable to private international law when developing an approach to interpretation in a conflicts analysis. Comparative law may contribute to legal theory by creating categories and concepts of general application. Comparative law may also provide insights into the question of legal transplants, i.e. the transplanting of law and legal institutions from one system to another. The notion of legal transplants was coined by Alan Watson, one of the world's renowned legal scholars specializing in comparative law. Also, the usefulness of comparative law for sociology of law and law and economics (and vice versa) is very large. The comparative study of the various legal systems may show how different legal regulations for the same problem function in practice. Conversely, sociology of law and law & economics may help comparative law answer questions, such as: Arminjon, Nolde, and Wolff believed that, for purposes of classifying the (then) contemporary legal systems of the world, it was required that those systems "per se" get studied, irrespective of external factors, such as geographical ones. They proposed the classification of legal system into seven groups, or so-called 'families', in particular the René David proposed the classification of legal systems, according to the different ideology inspiring each one, into five groups or families: Especially with respect to the aggregating by David of the Civil and Common laws into a single family, David argued that the antithesis between the Common law and Civil law systems, is of a technical rather than of an ideological nature. Of a different kind is, for instance, the antithesis between, say, Italian and American laws, and of a different kind than between the Soviet, Muslim, Hindu, or Chinese laws. According to David, the Civil law legal systems included those countries where legal science was formulated according to Roman law, whereas Common law countries are those dominated by judge-made law. The characteristics that he believed uniquely differentiate the Western legal family from the other four are: Konrad Zweigert and Hein Kötz propose a different, multidimensional methodology for categorizing laws, i.e. for ordering families of laws. They maintain that, to determine such families, five criteria should be taken into account, in particular: the historical background, the characteristic way of thought, the different institutions, the recognized sources of law, and the dominant ideology. Using the aforementioned criteria, they classify the legal systems of the world into six families: Up to the second German edition of their introduction to comparative law, Zweigert and Kötz also used to mention Soviet or socialist law as another family of laws.
https://en.wikipedia.org/wiki?curid=7378
Cyberspace Cyberspace is a concept describing a widespread, interconnected digital technology. The term entered the popular culture from science fiction and the arts but is now used by technology strategists, security professionals, government, military and industry leaders and entrepreneurs to describe the domain of the global technology environment. Others consider cyberspace to be just a notional environment in which communication over computer networks occurs. The word became popular in the 1990s when the uses of the Internet, networking, and digital communication were all growing dramatically and the term "cyberspace" was able to represent the many new ideas and phenomena that were emerging. As a social experience, individuals can interact, exchange ideas, share information, provide social support, conduct business, direct actions, create artistic media, play games, engage in political discussion, and so on, using this global network. They are sometimes referred to as "cybernauts". The term "cyberspace" has become a conventional means to describe anything associated with the Internet and the diverse Internet culture. The United States government recognizes the interconnected information technology and the interdependent network of information technology infrastructures operating across this medium as part of the US national critical infrastructure. Amongst individuals on cyberspace, there is believed to be a code of shared rules and ethics mutually beneficial for all to follow, referred to as cyberethics. Many view the right to privacy as most important to a functional code of cyberethics. Such moral responsibilities go hand in hand when working online with global networks, specifically, when opinions are involved with online social experiences. According to Chip Morningstar and F. Randall Farmer, cyberspace is defined more by the social interactions involved rather than its technical implementation. In their view, the computational medium in cyberspace is an augmentation of the communication channel between real people; the core characteristic of cyberspace is that it offers an environment that consists of many participants with the ability to affect and influence each other. They derive this concept from the observation that people seek richness, complexity, and depth within a virtual world. The term "cyberspace" first appeared in the visual arts in the late 1960s, when Danish artist Susanne Ussing (1940-1998) and her partner architect Carsten Hoff (b. 1934) constituted themselves as Atelier Cyberspace. Under this name the two made a series of installations and images entitled "sensory spaces" that were based on the principle of open systems adaptable to various influences, such as human movement and the behaviour of new materials. Atelier Cyberspace worked at a time when the Internet did not exist and computers were more or less off-limit to artists and creative engagement. In a 2015-interview with Scandinavian art magazine Kunstkritikk, Carsten Hoff recollects, that although Atelier Cyberspace did try to implement computers, they had no interest in the virtual space as such: And in the same interview Hoff continues: The works of Atelier Cyberspace were originally shown at a number of Copenhagen venues and have later been exhibited at The National Gallery of Denmark in Copenhagen as part of the exhibition "What’s Happening?" The term "cyberspace" first appeared in fiction in the 1980s in the work of cyberpunk science fiction author William Gibson, first in his 1982 short story "Burning Chrome" and later in his 1984 novel "Neuromancer". In the next few years, the word became prominently identified with online computer networks. The portion of "Neuromancer" cited in this respect is usually the following: Now widely used, the term has since been criticized by Gibson, who commented on the origin of the term in the 2000 documentary "No Maps for These Territories": Don Slater uses a metaphor to define cyberspace, describing the "sense of a social setting that exists purely within a space of representation and communication ... it exists entirely within a computer space, distributed across increasingly complex and fluid networks." The term "Cyberspace" started to become a de facto synonym for the Internet, and later the World Wide Web, during the 1990s, especially in academic circles and activist communities. Author Bruce Sterling, who popularized this meaning, credits John Perry Barlow as the first to use it to refer to "the present-day nexus of computer and telecommunications networks". Barlow describes it thus in his essay to announce the formation of the Electronic Frontier Foundation (note the spatial metaphor) in June 1990: As Barlow, and the EFF, continued public education efforts to promote the idea of "digital rights", the term was increasingly used during the Internet boom of the late 1990s. Although the present-day, loose use of the term "cyberspace" no longer implies or suggests immersion in a virtual reality, current technology allows the integration of a number of capabilities (sensors, signals, connections, transmissions, processors, and controllers) sufficient to generate a virtual interactive experience that is accessible regardless of a geographic location. It is for these reasons cyberspace has been described as the ultimate tax haven. In 1989, Autodesk, an American multinational corporation that focuses on 2D and 3D design software, developed a virtual design system called Cyberspace. Although several definitions of cyberspace can be found both in scientific literature and in official governmental sources, there is no fully agreed official definition yet. According to F. D. Kramer there are 28 different definitions of the term cyberspace. See in particular the following links: "Cyberpower and National Security: Policy Recommendations for a Strategic Framework," in Cyberpower and National Security, FD Kramer, S. Starr, L.K. Wentz (ed.), National Defense University Press, Washington (DC) 2009; see also Mayer, M., Chiarugi, I., De Scalzi, N., https://www.academia.edu/14336129/International_Politics_in_the_Digital_Age. The most recent draft definition is the following: The Joint Chiefs of Staff of the United States Department of Defense define cyberspace as one of five interdependent domains, the remaining four being land, air, maritime, and space. "See United States Cyber Command" While cyberspace should not be confused with the Internet, the term is often used to refer to objects and identities that exist largely within the communication network itself, so that a website, for example, might be metaphorically said to "exist in cyberspace". According to this interpretation, events taking place on the Internet are not happening in the locations where participants or servers are physically located, but "in cyberspace". The philosopher Michel Foucault used the term heterotopias, to describe such spaces which are simultaneously physical and mental. Firstly, cyberspace describes the flow of digital data through the network of interconnected computers: it is at once not "real", since one could not spatially locate it as a tangible object, and clearly "real" in its effects. There have been several attempts to create a concise model about how cyberspace works since it is not a physical thing that can be looked at. Secondly, cyberspace is the site of computer-mediated communication (CMC), in which online relationships and alternative forms of online identity were enacted, raising important questions about the social psychology of Internet use, the relationship between "online" and "offline" forms of life and interaction, and the relationship between the "real" and the virtual. Cyberspace draws attention to remediation of culture through new media technologies: it is not just a communication tool but a social destination and is culturally significant in its own right. Finally, cyberspace can be seen as providing new opportunities to reshape society and culture through "hidden" identities, or it can be seen as borderless communication and culture. The "space" in cyberspace has more in common with the abstract, mathematical meanings of the term (see space) than physical space. It does not have the duality of positive and negative volume (while in physical space, for example, a room has the negative volume of usable space delineated by positive volume of walls, Internet users cannot enter the screen and explore the unknown part of the Internet as an extension of the space they are in), but spatial meaning can be attributed to the relationship between different pages (of books as well as web servers), considering the unturned pages to be somewhere "out there." The concept of cyberspace, therefore, refers not to the content being presented to the surfer, but rather to the possibility of surfing among different sites, with feedback loops between the user and the rest of the system creating the potential to always encounter something unknown or unexpected. Video games differ from text-based communication in that on-screen images are meant to be figures that actually occupy a space and the animation shows the movement of those figures. Images are supposed to form the positive volume that delineates the empty space. A game adopts the cyberspace metaphor by engaging more players in the game, and then figuratively representing them on the screen as avatars. Games do not have to stop at the avatar-player level, but current implementations aiming for more immersive playing space (i.e. Laser tag) take the form of augmented reality rather than cyberspace, fully immersive virtual realities remaining impractical. Although the more radical consequences of the global communication network predicted by some cyberspace proponents (i.e. the diminishing of state influence envisioned by John Perry Barlow) failed to materialize and the word lost some of its novelty appeal, it remains current . Some virtual communities explicitly refer to the concept of cyberspace, for example Linden Lab calling their customers "Residents" of Second Life, while all such communities can be positioned "in cyberspace" for explanatory and comparative purposes (as did Sterling in "The Hacker Crackdown", followed by many journalists), integrating the metaphor into a wider cyber-culture. The metaphor has been useful in helping a new generation of thought leaders to reason through new military strategies around the world, led largely by the US Department of Defense (DoD). The use of cyberspace as a metaphor has had its limits, however, especially in areas where the metaphor becomes confused with physical infrastructure. It has also been critiqued as being unhelpful for falsely employing a spatial metaphor to describe what is inherently a network. A forerunner of the modern ideas of cyberspace is the Cartesian notion that people might be deceived by an evil demon that feeds them a false reality. This argument is the direct predecessor of modern ideas of a brain in a vat and many popular conceptions of cyberspace take Descartes's ideas as their starting point. Visual arts have a tradition, stretching back to antiquity, of artifacts meant to fool the eye and be mistaken for reality. This questioning of reality occasionally led some philosophers and especially theologians to distrust art as deceiving people into entering a world which was not real (see Aniconism). The artistic challenge was resurrected with increasing ambition as art became more and more realistic with the invention of photography, film (see "Arrival of a Train at La Ciotat"), and immersive computer simulations. American counterculture exponents like William S. Burroughs (whose literary influence on Gibson and cyberpunk in general is widely acknowledged) and Timothy Leary were among the first to extol the potential of computers and computer networks for individual empowerment. Some contemporary philosophers and scientists (e.g. David Deutsch in "The Fabric of Reality") employ virtual reality in various thought experiments. For example, Philip Zhai in "Get Real: A Philosophical Adventure in Virtual Reality" connects cyberspace to the Platonic tradition: Note that this brain-in-a-vat argument conflates cyberspace with reality, while the more common descriptions of cyberspace contrast it with the "real world". The “Geography of Notopia” (Papadimitriou, 2006) theorizes about the complex interplay of cyber-cultures and the geographical space. This interplay has several philosophical and psychological facets (Papadimitriou, 2009). The technological convergence of the mass media is the result of a long adaptation process of their communicative resources to the evolutionary changes of each historical moment. Thus, the new media became (plurally) an extension of the traditional media on the cyberspace, allowing to the public access information in a wide range of digital devices. In other words, it is a cultural virtualization of human reality as a result of the migration from physical to virtual space (mediated by the ICTs), ruled by codes, signs and particular social relationships. Forwards, arise instant ways of communication, interaction and possible quick access to information, in which we are no longer mere senders, but also producers, reproducers, co-workers and providers. New technologies also help to "connect" people from different cultures outside the virtual space, which was unthinkable fifty years ago. In this giant relationships web, we mutually absorb each other’s beliefs, customs, values, laws and habits, cultural legacies perpetuated by a physical-virtual dynamics in constant metamorphosis (ibidem). In this sense, Professor Doctor Marcelo Mendonça Teixeira created, in 2013, a new model of communication to the virtual universe, based in Claude Elwood Shannon (1948) article "A Mathematical Theory of Communication". Having originated among writers, the concept of cyberspace remains most popular in literature and film. Although artists working with other media have expressed interest in the concept, such as Roy Ascott, "cyberspace" in digital art is mostly used as a synonym for immersive virtual reality and remains more discussed than enacted. Cyberspace also brings together every service and facility imaginable to expedite money laundering. One can purchase anonymous credit cards, bank accounts, encrypted global mobile telephones, and false passports. From there one can pay professional advisors to set up IBCs (International Business Corporations, or corporations with anonymous ownership) or similar structures in OFCs (Offshore Financial Centers). Such advisors are loath to ask any penetrating questions about the wealth and activities of their clients, since the average fees criminals pay them to launder their money can be as much as 20 percent. In 2010, a five-level model was designed in France. According to this model, cyberspace is composed of five layers based on information discoveries: language, writing, printing, Internet, etc. This original model links the world of information to telecommunication technologies.
https://en.wikipedia.org/wiki?curid=7381
The Maritimes The Maritimes, also called the Maritime provinces or the Canadian Maritimes, is a region of Eastern Canada consisting of three provinces: New Brunswick, Nova Scotia, and Prince Edward Island (PEI). The Maritimes had a population of 1,813,606 in 2016. Together with Canada's easternmost province, Newfoundland and Labrador, the Maritime provinces make up the region of Atlantic Canada. Located along the Atlantic coast, various aquatic sub-basins are located in the Maritimes, such as the Gulf of Maine and Gulf of St. Lawrence. The region is located northeast of New England, southeast of Quebec's Gaspé Peninsula, and southwest of the island of Newfoundland. The notion of a Maritime Union has been proposed at various times in Canada's history; the first discussions in 1864 at the Charlottetown Conference contributed to Canadian Confederation. This movement formed the larger Dominion of Canada. The Mi'kmaq, Maliseet and Passamaquoddy people are indigenous to the Maritimes, while Acadian and British settlements date to the 17th century. The word maritime is an adjective that means "of the sea", thus any land associated with the sea can be considered a maritime state or province (all provinces of Canada except Alberta and Saskatchewan border the sea). But the term "Maritimes" has historically been collectively applied to New Brunswick, Nova Scotia and Prince Edward Island, all of which border the Atlantic Ocean. In other provinces except Newfoundland and Labrador and British Columbia, human settlement along the sea is sparse. The Hudson Bay area is northerly and has a severe climate, and the majority of the populations of Ontario, Quebec and Manitoba reside far inland. The prehistory of the Canadian Maritimes begins after the northerly retreat of glaciers at the end of the Wisconsin glaciation over 10,000 years ago; human settlement by First Nations began in the Maritimes with Paleo-Indians during the "Early Period", ending around 6,000 years ago. The "Middle Period", starting 6,000 years ago, and ending 3,000 years ago, was dominated by rising sea levels from the melting glaciers in polar regions. This is also when what is called the "Laurentian tradition" started among Archaic Indians, the term used for First Nations peoples of the time. Evidence of Archaic Indian burial mounds and other ceremonial sites existing in the Saint John River valley has been uncovered. The "Late Period" extended from 3,000 years ago until first contact with European settlers. This period was dominated by the organization of First Nations peoples into the Algonquian-speaking Abenaki Nation, which occupied territory largely in present-day interior Vermont, New Hampshire, and Maine, and the Mi'kmaq Nation, which inhabited all of Nova Scotia, Prince Edward Island, eastern New Brunswick and the southern Gaspé. The primarily agrarian Maliseet Nation settled throughout the Saint John River and Allagash River valleys of present-day New Brunswick and Maine. The Passamaquoddy Nation inhabited the northwestern coastal regions of the present-day Bay of Fundy. The Mi'kmaq Nation is also believed to have crossed the present-day Cabot Strait at around this time to settle on the south coast of Newfoundland, but they were a minority compared to the Beothuk Nation. After Newfoundland, the Maritimes were the second area in Canada to be settled by Europeans. There is evidence that Viking explorers discovered and settled in the Vinland region around 1000 AD, which is when the L'Anse aux Meadows settlement in Newfoundland and Labrador has been dated. They may have made further exploration into the present-day Maritimes and northeastern United States. Both Giovanni Caboto (John Cabot) and Giovanni da Verrazzano are reported to have sailed in or near Maritime waters during their voyages of discovery for England and France, respectively. Several Portuguese explorers/cartographers have also documented various parts of the Maritimes, namely Diogo Homem. However, it was French explorer Jacques Cartier who made the first detailed reconnaissance of the region for a European power and, in so doing, claimed the region for the King of France. Cartier was followed by nobleman Pierre Dugua, Sieur de Monts, who was accompanied by explorer/cartographer Samuel de Champlain in a 1604 expedition. During this they established the second permanent European settlement in what is now the United States and Canada, following Spain's settlement at St. Augustine in present-day Florida in the American South. Champlain's settlement at Saint Croix Island, later moved to Port-Royal, survived. By contrast, the ill-fated English settlement at Roanoke Colony off the southern American coast did not. The French settlement pre-dated the more successful English settlement at Jamestown in present-day Virginia by three years. Champlain was considered the founder of New France's province of Canada, which comprises much of the present-day lower St. Lawrence River valley in the province of Quebec. Champlain's success in the region, which came to be called "Acadie", led to the fertile tidal marshes surrounding the southeastern and northeastern reaches of the Bay of Fundy being populated by French immigrants who called themselves "Acadien". The Acadians eventually built small settlements throughout what is today mainland Nova Scotia and New Brunswick, as well as Île-Saint-Jean (Prince Edward Island), Île-Royale (Cape Breton Island), and other shorelines of the Gulf of St. Lawrence in present-day Newfoundland and Labrador, and Quebec. Acadian settlements had primarily agrarian economies. Early examples of Acadian fishing settlements developed in southwestern Nova Scotia and in Île-Royale, as well as along the south and west coasts of Newfoundland, the Gaspé Peninsula, and the present-day Côte-Nord region of Quebec. Most Acadian fishing activities were overshadowed by the much larger seasonal European fishing fleets that were based out of Newfoundland and took advantage of proximity to the Grand Banks. The growing English colonies along the American seaboard to the south and various European wars between England and France during the 17th and 18th centuries brought Acadia to the centre of world-scale geopolitical forces. In 1613, Virginian raiders captured Port-Royal, and in 1621 France ceded Acadia to Scotland's Sir William Alexander, who renamed it as "Nova Scotia". By 1632, Acadia was returned from Scotland to France under the "Treaty of Saint-Germain-en-Laye." The Port Royale settlement was moved to the site of nearby present-day Annapolis Royal. More French immigrant settlers, primarily from the Brittany, Normandie, and Vienne regions of France, continued to populate the colony of Acadia during the latter part of the 17th and early part of the 18th centuries. Important settlements also began in the Beaubassin region of the present-day Isthmus of Chignecto, and in the Saint John River valley, as well as smaller communities on Île-Saint-Jean and Île-Royale. In 1654, New England raiders attacked Acadian settlements on the Annapolis Basin. Acadians lived with uncertainty throughout the English constitutional crises under Oliver Cromwell, and it was not until the Treaty of Breda in 1667 that France's claim to the region was reaffirmed. Colonial administration by France throughout the history of Acadia was of low priority. France's priorities were in settling and strengthening its claim on the larger territory of New France and the exploration and settlement of interior North America and the Mississippi River valley. Over 74 years (1689–1763) there were six colonial wars, which involved continuous warfare between New England and Acadia (see the French and Indian Wars reflecting English and French tensions in Europe, as well as Father Rale's War and Father Le Loutre's War). Throughout these wars, New England was allied with the Iroquois Confederacy based around the southern Great Lakes and west of the Hudson River. Acadian settlers were allied with the Wabanaki Confederacy. In the first war, King William's War (the North American theater of the Nine Years' War), natives from the Maritime region participated in numerous attacks with the French on the Acadia/ New England border in southern Maine (e.g., Raid on Salmon Falls). New England retaliatory raids on Acadia, such as the Raid on Chignecto (1696), were conducted by Benjamin Church. In the second war, Queen Anne's War (the North American theater of the War of the Spanish Succession), the British conducted the Conquest of Acadia, while the region remained primarily in control of Maliseet militia, Acadia militia and Mi'kmaq militia. In 1719, to further protect strategic interests in the Gulf of St. Lawrence and St. Lawrence River, France began the 20-year construction of a large fortress at Louisbourg on Île-Royale. Massachusetts was increasingly concerned over reports of the capabilities of this fortress, and of privateers staging out of its harbour to raid New England fishermen on the Grand Banks. In the fourth war, King George's War (the North American theater of the War of the Austrian Succession), the British engaged successfully in the Siege of Louisbourg (1745). The British returned control of Île-Royale to France with the fortress virtually intact three years later under the Treaty of Aix-la-Chapelle and the French reestablished their forces there. In 1749, to counter the rising threat of Louisbourg, Halifax was founded and the Royal Navy established a major naval base and citadel. The founding of Halifax sparked Father Le Loutre's War. During the sixth and final colonial war, the French and Indian War (the North American theater of the Seven Years' War), the military conflicts in Nova Scotia continued. The British Conquest of Acadia happened in 1710. Over the next forty-five years the Acadians refused to sign an unconditional oath of allegiance to Britain. During this time period Acadians participated in various militia operations against the British and maintained vital supply lines to the French Fortress of Louisbourg and Fort Beausejour. The British sought to neutralize any military threat Acadians posed and to interrupt the vital supply lines Acadians provided to Louisbourg by deporting Acadians from Acadia. The British began the Expulsion of the Acadians with the Bay of Fundy Campaign (1755). Over the next nine years over 12,000 Acadians of 15,000 were removed from Nova Scotia. In 1758, the fortress of Louisbourg was laid siege for a second time within 15 years, this time by more than 27,000 British soldiers and sailors with over 150 warships. After the French surrender, Louisbourg was thoroughly destroyed by British engineers to ensure it would never be reclaimed. With the fall of Louisbourg, French and Mi'kmaq resistance in the region crumbled. British forces seized remaining French control over Acadia in the coming months, with Île-Saint-Jean falling in 1759 to British forces on their way to Quebec City for the Siege of Quebec and ensuing Battle of the Plains of Abraham. The war ended and Britain had gained control over the entire Maritime region and the Indigenous people signed the Halifax Treaties. Following the Seven Years' War, empty Acadian lands were settled first by 8,000 New England Planters and then by immigrants brought from Yorkshire. Île-Royale was renamed Cape Breton Island and incorporated into the Colony of Nova Scotia. Some of the Acadians who had been deported came back but went to New Brunswick. Both the colonies of Nova Scotia (present-day Nova Scotia and New Brunswick) and St. John's Island (Prince Edward Island) were affected by the American Revolutionary War, largely by privateering against American shipping, but several coastal communities were also the targets of American raiders. Charlottetown, the capital of the new colony of St. John's Island, was ransacked in 1775 with the provincial secretary kidnapped and the Great Seal stolen. The largest military action in the Maritimes during the revolutionary war was the attack on Fort Cumberland (the renamed Fort Beausejour) in 1776 by a force of American sympathizers led by Jonathan Eddy. The fort was partially overrun after a month-long siege, but the attackers were ultimately repelled after the arrival of British reinforcements from Halifax. The most significant impact from this war was the settling of large numbers of Loyalist refugees in the region (34,000 to the 17,000 settlers already there), especially in Shelburne and Parrtown (Saint John). Following the Treaty of Paris in 1783, Loyalist settlers in what would become New Brunswick persuaded British administrators to split the Colony of Nova Scotia to create the new colony of New Brunswick in 1784. At the same time, another part of the Colony of Nova Scotia, Cape Breton Island, was split off to become the Colony of Cape Breton Island. The Colony of St. John's Island was renamed to Prince Edward Island on November 29, 1798. The War of 1812 had some effect on the shipping industry in the Maritime colonies of New Brunswick, Nova Scotia, Prince Edward Island, and Cape Breton Island; however, the significant Royal Navy presence in Halifax and other ports in the region prevented any serious attempts by American raiders. Maritime and American privateers targeted unprotected shipping of both the United States and Britain respectively, further reducing trade. New Brunswick's section of the Canada–US border did not have any significant action during this conflict, although British forces did occupy a portion of coastal Maine at one point. The most significant incident from this war which occurred in the Maritimes was the British capture and detention of the American frigate USS "Chesapeake" in Halifax. In 1820, the Colony of Cape Breton Island was merged back into the Colony of Nova Scotia for the second time by the British government. British settlement of the Maritimes, as the colonies of Nova Scotia, New Brunswick and Prince Edward Island came to be known, accelerated throughout the late 18th century and into the 19th century with significant immigration to the region as a result of Scottish migrants displaced by the Highland Clearances and Irish escaping the Great Irish Famine (1845–1849). As a result, significant portions of the three provinces are influenced by Celtic heritages, with Scottish Gaelic (and to a lesser degree, Irish Gaelic) having been widely spoken, particularly in Cape Breton, although it is less prevalent today. During the American Civil War, a significant number of Maritimers volunteered to fight for the armies of the Union, while a small handful joined the Confederate Army. However, the majority of the conflict's impact was felt in the shipping industry. Maritime shipping boomed during the war due to large-scale Northern imports of war supplies which were often carried by Maritime ships as Union ships were vulnerable to Confederate naval raiders. Diplomatic tensions between Britain and the Unionist North had deteriorated after some interests in Britain expressed support for the secessionist Confederate South. The Union navy, although much smaller than the British Royal Navy and no threat to the Maritimes, did posture off Maritime coasts at times chasing Confederate naval ships which sought repairs and reprovisioning in Maritime ports, especially Halifax. The immense size of the Union army (the largest on the planet toward the end of the Civil War), however, was viewed with increasing concern by Maritimers throughout the early 1860s. Another concern was the rising threat of Fenian raids on border communities in New Brunswick by those seeking to end British rule of Ireland. This combination of events, coupled with an ongoing decline in British military and economic support to the region as the Home Office favoured newer colonial endeavours in Africa and elsewhere, led to a call among Maritime politicians for a conference on Maritime Union, to be held in early September 1864 in Charlottetown – chosen in part because of Prince Edward Island's reluctance to give up its jurisdictional sovereignty in favour of uniting with New Brunswick and Nova Scotia into a single colony. New Brunswick and Nova Scotia felt that if the union conference were held in Charlottetown, they might be able to convince Island politicians to support the proposal. The Charlottetown Conference, as it came to be called, was also attended by a slew of visiting delegates from the neighbouring colony of Canada, who had largely arrived at their own invitation with their own agenda. This agenda saw the conference dominated by discussions of creating an even larger union of the entire territory of British North America into a united colony. The Charlottetown Conference ended with an agreement to meet the following month in Quebec City, where more formal discussions ensued, culminating with meetings in London and the signing of the British North America Act. Of the Maritime provinces, only Nova Scotia and New Brunswick were initially party to the BNA Act, Prince Edward Island's reluctance, combined with a booming agricultural and fishing export economy having led to that colony opting not to sign on. The major communities of the region include Halifax and Cape Breton in Nova Scotia, Saint John, Fredericton and Moncton in New Brunswick, and Charlottetown in Prince Edward Island. In spite of its name, The Maritimes has a humid continental climate of the warm-summer subtype. Especially in coastal Nova Scotia, differences between summers and winters are narrow compared to the rest of Canada. The inland climate of New Brunswick is in stark contrast during winter, resembling more continental areas. Summers are somewhat tempered by the marine influence throughout the provinces, but due to the southerly parallels still remain similar to more continental areas further west. Yarmouth in Nova Scotia has significant marine influence to have a borderline oceanic microclimate, but winter nights are still cold even in all coastal areas. The northernmost areas of New Brunswick are only just above subarctic with very cold continental winters. Maritime society is based upon a mixture of traditions and class backgrounds. Predominantly rural until recent decades, the region traces many of its cultural activities to those rural resource-based economies of fishing, agriculture, forestry, and coal mining. While Maritimers are predominantly of west European heritage (Scottish, Irish, English, and Acadian), immigration to Industrial Cape Breton during the heyday of coal mining and steel manufacturing brought people from eastern Europe as well as from Newfoundland. The Maritimes also have a black population who are mostly descendants of African American loyalists or refugees from the War of 1812, largely concentrated in Nova Scotia but also in various communities throughout southern New Brunswick, Cape Breton (where the black population is largely of West Indian descent), and Prince Edward Island. The Mi'kmaq Nation's reserves throughout Nova Scotia, Prince Edward Island and eastern New Brunswick dominate aboriginal culture in the region, compared to the much smaller population of the Maliseet Nation in western New Brunswick. New Brunswick, in general, differs from the other two Maritime Provinces in that its French population plays a significant role in the everyday cultural experience. Being the only officially bilingual province in Canada, many of New Brunswick's inhabitants speak both French and English, especially in Moncton and the capital region of Fredericton. Cultural activities are fairly diverse throughout the region, with the music, dance, theatre, and literary art forms tending to follow the particular cultural heritage of specific locales. Notable Nova Scotian folklorist and cultural historian Helen Creighton spent the majority of her lifetime recording the various Celtic musical and folk traditions of rural Nova Scotia during the mid-20th century, prior to this knowledge being wiped out by mass media assimilation with the rest of North America. A fragment of Gaelic culture remains in Nova Scotia, specifically on Cape Breton Island. PEI and New Brunswick share this historical tie to Gaelic culture with Nova Scotia but it plays a far less significant role in their respective public provincial images. Canada has witnessed a "Celtic revival" in which many Maritime musicians and songs have risen to prominence in recent decades. Some companies, particularly breweries such as Alexander Keith's and Moosehead have played up a connection between folklore with alcohol consumption during their marketing campaigns. The Maritimes were among the strongest supporters of prohibition (Prince Edward Island lasting until 1949), and some predominantly rural communities maintain "dry" status, banning the retail sale of alcohol as a vestige of the original temperance movement in the region. Given the small population of the region (compared with the Central Canadian provinces or the New England states), the regional economy is a net exporter of natural resources, manufactured goods, and services. The regional economy has long been tied to natural resources such as fishing, logging, farming, and mining activities. Significant industrialization in the second half of the 19th century brought steel to Trenton, Nova Scotia, and subsequent creation of a widespread industrial base to take advantage of the region's large underground coal deposits. After Confederation, however, this industrial base withered with technological change, and trading links to Europe and the U.S. were reduced in favour of those with Ontario and Quebec. In recent years, however, the Maritime regional economy has begun increased contributions from manufacturing again and the steady transition to a service economy. Important manufacturing centres in the region include Pictou County, Truro, the Annapolis Valley and the South Shore, and the Strait of Canso area in Nova Scotia, as well as Summerside in Prince Edward Island, and the Miramichi area, the North Shore and the upper Saint John River valley of New Brunswick. Some predominantly coastal areas have become major tourist centres, such as parts of Prince Edward Island, Cape Breton Island, the South Shore of Nova Scotia and the Gulf of St. Lawrence and Bay of Fundy coasts of New Brunswick. Additional service-related industries in information technology, pharmaceuticals, insurance and financial sectors—as well as research-related spin-offs from the region's numerous universities and colleges—are significant economic contributors. Another important contribution to Nova Scotia's provincial economy is through spin-offs and royalties relating to off-shore petroleum exploration and development. Mostly concentrated on the continental shelf of the province's Atlantic coast in the vicinity of Sable Island, exploration activities began in the 1960s and resulted in the first commercial production field for oil beginning in the 1980s. Natural gas was also discovered in the 1980s during exploration work, and this is being commercially recovered, beginning in the late 1990s. Initial optimism in Nova Scotia about the potential of off-shore resources appears to have diminished with the lack of new discoveries, although exploration work continues and is moving farther off-shore into waters on the continental margin. Regional transportation networks have also changed significantly in recent decades with port modernizations, with new freeway and ongoing arterial highway construction, the abandonment of various low-capacity railway branchlines (including the entire railway system of Prince Edward Island and southwestern Nova Scotia), and the construction of the Canso Causeway and the Confederation Bridge. There have been airport improvements at various centres providing improved connections to markets and destinations in the rest of North America and overseas. Improvements in infrastructure and the regional economy notwithstanding, the three provinces remain one of the poorer regions of Canada. While urban areas are growing and thriving, economic adjustments have been harsh in rural and resource-dependent communities, and emigration has been an ongoing phenomenon for some parts of the region. Another problem is seen in the lower average wages and family incomes within the region. Property values are depressed, resulting in a smaller tax base for these three provinces, particularly when compared with the national average which benefits from central and western Canadian economic growth. This has been particularly problematic with the growth of the welfare state in Canada since the 1950s, resulting in the need to draw upon equalization payments to provide nationally mandated social services. Since the 1990s the region has experienced an exceptionally tumultuous period in its regional economy with the collapse of large portions of the ground fishery throughout Atlantic Canada, the closing of coal mines and a steel mill on Cape Breton Island, and the closure of military bases in all three provinces. That being said, New Brunswick has one of the largest military bases in the British Commonwealth (CFB Gagetown), which plays a significant role in the cultural and economic spheres of Fredericton, the province's capital city. While the economic underperformance of the Maritime economy has been long lasting, it has not always been present. The mid-19th century, especially the 1850s and 1860s, has long been seen as a "Golden Age" in the Maritimes. Growth was strong, and the region had one of British North America's most extensive manufacturing sectors as well as a large international shipping industry. The question of why the Maritimes fell from being a centre of Canadian manufacturing to being an economic hinterland is thus a central one to the study of the region's pecuniary difficulties. The period in which the decline occurred had a great many potential culprits. In 1867 Nova Scotia and New Brunswick merged with the Canadas in Confederation, with Prince Edward Island joining them six years later in 1873. Canada was formed only a year after free trade with the United States (in the form of the Reciprocity Agreement) had ended. In the 1870s John A. Macdonald's National Policy was implemented, creating a system of protective tariffs around the new nation. Throughout the period there was also significant technological change both in the production and transportation of goods. Several scholars have explored the so-called "golden age" of the Maritimes in the years just before Confederation. In Nova Scotia, the population grew steadily from 277,000 in 1851 to 388,000 in 1871, mostly from natural increase since immigration was slight. The era has been called a golden age, but that was a myth created in the 1930s to lure tourists to a romantic era of tall ships and antiques. Recent historians using census data have shown that is a fallacy. In 1851–1871 there was an overall increase in per capita wealth holding. However most of the gains went to the urban elite class, especially businessmen and financiers living in Halifax. The wealth held by the top 10% rose considerably over the two decades, but there was little improvement in the wealth levels in rural areas, which comprised the great majority of the population. Likewise Gwyn reports that gentlemen, merchants, bankers, colliery owners, shipowners, shipbuilders, and master mariners flourished. However the great majority of families were headed by farmers, fishermen, craftsmen and laborers. Most of them—and many widows as well—lived in poverty. Out migration became an increasingly necessary option. Thus the era was indeed a golden age but only for a small but powerful and highly visible elite. The cause of economic malaise in the Maritimes is an issue of great debate and controversy among historians, economists, and geographers. The differing opinions can approximately be divided into the "structuralists," who argue that poor policy decisions are to blame, and the others, who argue that unavoidable technological and geographical factors caused the decline. The exact date that the Maritimes began to fall behind the rest of Canada is difficult to determine. Historian Kris Inwood places the date very early, at least in Nova Scotia, finding clear signs that the Maritimes "Golden Age" of the mid-19th century was over by 1870, before Confederation or the National Policy could have had any significant impact. Richard Caves places the date closer to 1885. T.W. Acheson takes a similar view and provides considerable evidence that the early 1880s were in fact a booming period in Nova Scotia and this growth was only undermined towards the end of that decade. David Alexander argues that any earlier declines were simply part of the global Long Depression, and that the Maritimes first fell behind the rest of Canada when the great boom period of the early 20th century had little effect on the region. E.R. Forbes, however, emphasizes that the precipitous decline did not occur until after the First World War during the 1920s when new railway policies were implemented. Forbes also contends that significant Canadian defence spending during the Second World War favoured powerful political interests in Central Canada such as C.D. Howe, when major Maritime shipyards and factories, as well as Canada's largest steel mill, located in Cape Breton Island, fared poorly. One of the most important changes, and one that almost certainly had an effect, was the revolution in transportation that occurred at this time. The Maritimes were connected to central Canada by the Intercolonial Railway in the 1870s, removing a longstanding barrier to trade. For the first time this placed the Maritime manufacturers in direct competition with those of Central Canada. Maritime trading patterns shifted considerably from mainly trading with New England, Britain, and the Caribbean, to being focused on commerce with the Canadian interior, enforced by the federal government's tariff policies. Coincident with the construction of railways in the region, the age of the wooden sailing ship began to come to an end, being replaced by larger and faster steel steam ships. The Maritimes had long been a centre for shipbuilding, and this industry was hurt by the change. The larger ships were also less likely to call on the smaller population centres such as Saint John and Halifax, preferring to travel to cities like New York and Montreal. Even the Cunard Line, founded by Maritime-born Samuel Cunard, stopped making more than a single ceremonial voyage to Halifax each year. More controversial than the role of technology is the argument over the role of politics in the origins of the region's decline. Confederation and the tariff and railway freight policies that followed have often been blamed for having a deleterious effect on the Maritime economies. Arguments have been made that the Maritimes' poverty was caused by control over policy by Central Canada which used the national structures for its own enrichment. This was the central view of the Maritime Rights Movement of the 1920s, which advocated greater local control over the region's finances. T.W. Acheson is one of the main proponents of this theory. He notes the growth that was occurring during the early years of the National Policy in Nova Scotia demonstrates how the effects of railway fares and the tariff structure helped undermine this growth. Capitalists from Central Canada purchased the factories and industries of the Maritimes from their bankrupt local owners and proceeded to close down many of them, consolidating the industry in Central Canada. The policies in the early years of Confederation were designed by Central Canadian interests, and they reflected the needs of that region. The unified Canadian market and the introduction of railroads created a relative weakness in the Maritime economies. Central to this concept, according to Acheson, was the lack of metropolises in the Maritimes. Montreal and Toronto were well suited to benefit from the development of large-scale manufacturing and extensive railway systems in Quebec and Ontario, these being the goals of the Macdonald and Laurier governments. In the Maritimes the situation was very different. Today New Brunswick has several mid-sized centres in Saint John, Moncton, and Fredericton but no significant population centre. Nova Scotia has a growing metropolitan area surrounding Halifax, but a contracting population in industrial Cape Breton, and several smaller centres in Bridgewater, Kentville, Yarmouth, and Pictou County. Prince Edward Island's only significant population centres are in Charlottetown and Summerside. During the late 19th and early 20th centuries, just the opposite was the case with little to no population concentration in major industrial centres as the predominantly rural resource-dependent Maritime economy continued on the same path as it had since European settlement on the region's shores. Despite the region's absence of economic growth on the same scale as other parts of the nation, the Maritimes has changed markedly throughout the 20th century, partly as a result of global and national economic trends, and partly as a result of government intervention. Each sub-region within the Maritimes has developed over time to exploit different resources and expertise. Saint John became a centre of the timber trade and shipbuilding and is currently a centre for oil refining and some manufacturing. The northern New Brunswick communities of Edmundston, Campbellton, Dalhousie, Bathurst, and Miramichi are focused on the pulp and paper industry and some mining activity. Moncton was a centre for railways and has changed its focus to becoming a multi-modal transportation centre with associated manufacturing and retail interests. The Halifax metropolitan area has come to dominate peninsular Nova Scotia as a retail and service centre, but that province's industries were spread out from the coal and steel industries of industrial Cape Breton and Pictou counties, the mixed farming of the North Shore and Annapolis Valley, and the fishing industry was primarily focused on the South Shore and Eastern Shore. Prince Edward Island is largely dominated by farming, fishing, and tourism. Given the geographic diversity of the various sub-regions within the Maritimes, policies to centralize the population and economy were not initially successful, thus Maritime factories closed while those in Ontario and Quebec prospered. The traditional staples thesis, advocated by scholars such as S.A. Saunders, looks at the resource endowments of the Maritimes and argues that it was the decline of the traditional industries of shipbuilding and fishing that led to Maritime poverty, since these processes were rooted in geography, and thus all but inevitable. Kris Inwood has revived the staples approach and looks at a number of geographic weaknesses relative to Central Canada. He repeats Acheson's argument that the region lacks major urban centres, but adds that the Maritimes were also lacking the great rivers that led to the cheap and abundant hydro-electric power, key to Quebec and Ontario's urban and manufacturing development, that the extraction costs of Maritime resources were higher (particularly in the case of Cape Breton coal), and that the soils of the region were poorer and thus the agricultural sector weaker. The Maritimes are the only provinces in Canada which entered Confederation in the 19th century and have kept their original colonial boundaries. All three provinces have the smallest land base in the country and have been forced to make do with resources within. By comparison, the former colony of the United Province of Canada (divided into the District of Canada East, and the District of Canada West) and the western provinces were dozens of times larger and in some cases were expanded to take in territory formerly held in British Crown grants to companies such as the Hudson's Bay Company; in particular the November 19, 1869 sale of Rupert's Land to the Government of Canada under the "Rupert's Land Act 1868" was facilitated in part by Maritime taxpayers. The economic riches of energy and natural resources held within this larger land base were only realized by other provinces during the 20th century. The maritime provinces' main industry is fishing. Fishing can be found in any maritime province. This includes fishing for lobster, mackerel, tuna, salmon and many more kinds of fish. Oysters and salmonid aquiculture is also increasingly important economically. Nova Scotia is very strong in agriculture, forestry and fishing. Tourism is important to the economy of PEI. "Anne of Green Gables" was written in PEI. Because of this, every year it attracts families to see where the beloved story was based, and in the summers, to enjoy beaches and sunlight. PEI is also known for its agriculture, mainly the potato, and fishing industries. Agriculture and forestry are two prominent industries found in New Brunswick. Despite having an extensive coastline, New Brunswick's industrial sector has never been entirely reliant on the success of the fisheries. Likewise, the strong shipbuilding heritage of the province directly relates to its forest resources. Because of this, New Brunswickers tend to attribute their cultural heritage less with the sea and more with their forests and rivers. Maritime conservatism since the Second World War has been very much part of the Red Tory tradition, key influences being former Premier of Nova Scotia and federal Progressive Conservative Party leader Robert Stanfield and New Brunswick Tory strategist Dalton Camp. In recent years, the social democratic New Democratic Party (NDP) has made significant inroads both federally and provincially in the region. The NDP has elected Members of Parliament (MPs) from New Brunswick, but most of the focus of the party at the federal and provincial levels is currently in the Halifax area of Nova Scotia. Industrial Cape Breton has historically been a region of labour activism, electing Co-operative Commonwealth Federation (and later NDP) MPs, and even produced many early members of the Communist Party of Canada in the pre-World War II era. In the 2004 federal election, the NDP captured 28.45% of the vote in Nova Scotia, more than any other province. In the 2009 provincial election the NDP formed a majority government, the first in the region. The Maritimes are generally socially conservative but unlike Alberta, they also have fiscally socialist tendencies. It is because of the lack of support for fiscal conservatism that federal parties such as the Canadian Alliance never had much success in the region, In the 2004 federal election, the Conservatives had one of the worst showings in the region for a right-wing party, going back to Confederation, with the exception of the 1993 election. The Conservative party improved its seat count in the 2008 and elected 13 MPs in the 2011 election. However, in the 2015 election the Liberal Party won every seat in the region, defeating all of the Conservative (and NDP) challengers. An area within the region where both fiscal and social conservatism coincide and where the federal Reform Party and Canadian Alliance had some success is in central-western part of New Brunswick, in the Saint John River valley north of Saint John and south of Grand Falls. Contributing demographics include a predominantly Anglophone population residing in a largely rural agrarian setting. One influence might be proximity to the Canada–US border and the state of Maine. The valley is also settled by descendants of United Empire Loyalists whose influence continues in the area. There are also a large number of active and retired military personnel located in the Fredericton and Oromocto area as a result of the large military base at CFB Gagetown. Another area in the region with smatterings of coinciding fiscal and social conservatism is the Annapolis Valley of Nova Scotia. The Liberal Party of Canada has done well in the Maritimes in the past because of its interventionist policies. The Acadian Peninsula region of New Brunswick, long dependent upon seasonal employment in the Gulf of St. Lawrence fishery, tends to vote for the Liberals or NDP for this reason. In the 1997 federal election, Prime Minister Jean Chrétien's Liberals endured a bitter defeat to the PCs and NDP in many ridings as a result of unpopular cuts to unemployment benefits for seasonal workers, as well as closures of several Canadian Forces Bases, the refusal to honour a promise to rescind the Goods and Services Tax, cutbacks to provincial equalization payments, health care, post-secondary education and regional transportation infrastructure such as airports, fishing harbours, seaports, and railways. The Liberals held onto seats in Prince Edward Island and New Brunswick, while being shut out of Nova Scotia entirely, the second time in history (the only other time being the Diefenbaker sweep). In 2015 the Liberals won every seat in The Maritimes, defeating Conservative and NDP incumbents. The Maritimes is currently represented in the Canadian Parliament by 25 Members of the House of Commons (Nova Scotia – 11, New Brunswick – 10, Prince Edward Island – 4) and 24 Senators (Nova Scotia and New Brunswick – 10 each, Prince Edward Island – 4). This level of representation was established at the time of Confederation when the Maritimes had a much larger proportion of the national population. The comparatively large population growth of western and central Canada during the immigration boom of the 20th century has reduced the Maritimes' proportion of the national population to less than 10%, resulting in an over-representation in Parliament, with some federal ridings having fewer than 35,000 people, compared to central and western Canada where ridings typically contain 100,000–120,000 people. The Senate of Canada is structured along regional lines, giving an equal number of seats (24) to the Maritimes, Ontario, Quebec, and western Canada, in addition to the later entry of Newfoundland and Labrador, as well as the three territories. Enshrined in the Constitution, this model was developed to ensure that no area of the country is able to exert undue influence in the Senate. The Maritimes, with its much smaller proportion of the national population (compared to the time of Confederation) also have an over-representation in the Senate, particularly compared to the population growth of Ontario and the western provinces. This has led to calls to reform the Senate; however, such a move would entail constitutional changes. Another factor related to the number of Senate seats is that a constitutional amendment in the early 20th century mandated that no province can have fewer Members of Parliament than it has senators. This court decision resulted from a complaint by the Government of Prince Edward Island after that province's number of MPs was proposed to change from 4 to 3, accounting for its declining proportion of the national population at that time. When PEI entered Confederation in 1873, it was accorded 6 MPs and 4 Senators; however this was reduced to 4 MPs by the early 20th century. Senators being appointed for life at this time, these coveted seats rarely went unfilled for a long period of time anywhere in Canada. As a result, PEI's challenge was accepted by the federal government, and its level of federal representation was secured. In the aftermath of the 1989 budget, which saw a fillibuster by Liberal Senators in attempt to kill legislation creating the Goods and Services Tax, Prime Minister Brian Mulroney "stacked" the Senate by creating additional seats in several provinces across Canada, including New Brunswick; however, there was no attempt by these provinces to increase the number of MPs to reflect this change in Senate representation.
https://en.wikipedia.org/wiki?curid=7382
Cyril of Alexandria Cyril of Alexandria (; ;  376 – 444) was the Patriarch of Alexandria from 412 to 444. He was enthroned when the city was at the height of its influence and power within the Roman Empire. Cyril wrote extensively and was a leading protagonist in the Christological controversies of the late-4th and 5th centuries. He was a central figure in the Council of Ephesus in 431, which led to the deposition of Nestorius as Patriarch of Constantinople. Cyril is counted among the Church Fathers and the Doctors of the Church, and his reputation within the Christian world has resulted in his titles "Pillar of Faith" and "Seal of all the Fathers", but Theodosius II, the Roman Emperor, condemned him for behaving like a "proud pharaoh", and the Nestorian bishops at the Council of Ephesus declared him a heretic, labelling him as a "monster, born and educated for the destruction of the church." Cyril is well known for his dispute with Nestorius and his supporter Patriarch John of Antioch, whom Cyril excluded from the Council of Ephesus for arriving late. He is also known for his expulsion of Novatians and Jews from Alexandria and for inflaming tensions that led to the murder of the Hellenistic philosopher Hypatia by a Christian mob. Historians disagree over the extent of his responsibility in this. Cyril tried to oblige the pious Christian emperor Theodosius II (AD 408-450) to himself by dedicating his Paschal table to him. It is also important to note that Cyril’s Paschal table was provided with a Metonic basic structure in the form of a Metonic 19-year lunar cycle adopted by him around AD 425, which was very different from the very first Metonic 19-year lunar cycle invented around AD 260 by Anatolius, but exactly equal to the similar lunar cycle which had been introduced around AD 412 by Annianus; the julian equivalent of this Alexandrian lunar cycle adopted by Cyril and nowadays referred to as the ‘classical (Alexandrian) 19-year lunar cycle’ would only much later emerge again: a century later in Rome as the basic structure of Dionysius Exiguus’ Paschal table (AD 525) and two more centuries later in England as the one of Beda’s Easter table (AD 725). The Roman Catholic Church did not commemorate Saint Cyril in the Tridentine Calendar: it added his feast only in 1882, assigning to it the date of 9 February. This date is used by the Western Rite Orthodox Church. Yet the 1969 Catholic Calendar revision moved it to 27 June, considered to be the day of the saint's death, as celebrated by the Coptic Orthodox Church. The same date has been chosen for the Lutheran calendar. The Eastern Orthodox and Byzantine Catholic Churches celebrate his feast day on 9 June and also, together with Pope Athanasius I of Alexandria, on 18 January. Little is known for certain of Cyril's early life. He was born  376, in the small town of Didouseya, Egypt, modern-day El-Mahalla El-Kubra. A few years after his birth, his maternal uncle Theophilus rose to the powerful position of Patriarch of Alexandria. His mother remained close to her brother and under his guidance, Cyril was well educated. His writings show his knowledge of Christian writers of his day, including Eusebius, Origen, Didymus the Blind, and writers of the Church of Alexandria. He received the formal Christian education standard for his day: he studied grammar from age twelve to fourteen (390–392), rhetoric and humanities from fifteen to twenty (393–397) and finally theology and biblical studies (398–402). In 403 he accompanied his uncle to attend the "Synod of the Oak" in Constantinople, which deposed John Chrysostom as Archbishop of Constantinople. The prior year, Theophilus had been summoned by the emperor to Constantinople to apologize before a synod, over which Chrysostom would preside, on account of several charges which were brought against him by certain Egyptian monks. Theophilus had them persecuted as Origenists. Placing himself at the head of soldiers and armed servants Theophilus had marched against the monks, burned their dwellings, and ill-treated those whom he captured. Theophilus arrived at Constantinople with twenty-nine of his suffragan bishops, and conferring with those opposed to the Archbishop, drafted a long list of largely unfounded accusations against Chrysostom, who refused to recognize the legality of a synod in which his open enemies were judges. Chrysostom was subsequently deposed. Theophilus died on 15 October 412, and Cyril was made Pope or Patriarch of Alexandria on 18 October 412, but only after a riot between his supporters and those of his rival Archdeacon Timotheus. According to Socrates Scholasticus, the Alexandrians were always rioting. Thus, Cyril followed his uncle in a position that had become powerful and influential, rivalling that of the prefect in a time of turmoil and frequently violent conflict between the cosmopolitan city's pagan, Jewish, and Christian inhabitants. He began to exert his authority by causing the churches of the Novatianists to be closed and their sacred vessels to be seized. Orestes, "Praefectus augustalis" of the Diocese of Egypt, steadfastly resisted Cyril's ecclesiastical encroachment onto secular prerogatives. Tension between the parties increased when in 415, Orestes published an edict that outlined new regulations regarding mime shows and dancing exhibitions in the city, which attracted large crowds and were commonly prone to civil disorder of varying degrees. Crowds gathered to read the edict shortly after it was posted in the city's theater. Cyril sent the "grammaticus" Hierax to discover the content of the edict. The edict angered Christians as well as Jews. At one such gathering, Hierax read the edict and applauded the new regulations, prompting a disturbance. Many people felt that Hierax was attempting to incite the crowd into sedition. Orestes had Hierax tortured in public in a theatre. This order had two aims: one to quell the riot, the other to mark Orestes' authority over Cyril. Socrates Scholasticus recounts that upon hearing of Hierex's severe and public punishment, Cyril threatened to retaliate against the Jews of Alexandria with "the utmost severities" if the harassment of Christians did not cease immediately. In response to Cyril's threat, the Jews of Alexandria grew even more furious, eventually resorting to violence against the Christians. They plotted to flush the Christians out at night by running through the streets claiming that the Church of Alexander was on fire. When Christians responded to what they were led to believe was the burning down of their church, "the Jews immediately fell upon and slew them" by using rings to recognize one another in the dark and killing everyone else in sight. When the morning came, Cyril, along with many of his followers, took to the city's synagogues in search of the perpetrators of the massacre. According to Socrates Scholasticus, after Cyril rounded up all the Jews in Alexandria he ordered them to be stripped of all possessions, banished them from Alexandria, and allowed their goods to be pillaged by the remaining citizens of Alexandria. Scholasticus indicates that all the Jews were banished, while John of Nikiû says only those involved in the ambush. Susan Wessel says that, while it is not clear whether Scholasticus was a Novationist (whose churches Cyril had closed), he was apparently sympathetic towards them, and makes clear Cyril's habit of abusing his episcopal power by infringing on the rights and duties of the secular authorities. Wessel says "...Socrates probably does not provide accurate and unambiguous information about Cyril's relationship to imperial authority". Nonetheless, with Cyril's banishment of the Jews, however many, "Orestes [...] was filled with great indignation at these transactions, and was excessively grieved that a city of such magnitude should have been suddenly bereft of so large a portion of its population." Because of this, the feud between Cyril and Orestes intensified, and both men wrote to the emperor regarding the situation. Eventually, Cyril attempted to reach out to Orestes through several peace overtures, including attempted mediation and, when that failed, showed him the Gospels, which he interpreted to indicate that the religious authority of Cyril would require Orestes' acquiescence in the bishop's policy. Nevertheless, Orestes remained unmoved by such gestures. This refusal almost cost Orestes his life. Nitrian monks came from the desert and instigated a riot against Orestes among the population of Alexandria. These monks had resorted to violence 15 years before, during a controversy between Theophilus (Cyril's uncle) and the "Tall Brothers"; the monks assaulted Orestes and accused him of being a pagan. Orestes rejected the accusations, showing that he had been baptised by the Archbishop of Constantinople. A monk named Ammonius threw a stone hitting Orestes in the head. The prefect had Ammonius tortured to death, whereupon the Patriarch honored him as a martyr. However, according to Scholasticus, the Christian community displayed a general lack of enthusiasm for Ammonius's case for martyrdom. The prefect then wrote to the emperor Theodosius II, as did Cyril. The Prefect Orestes enjoyed the political backing of Hypatia, an astronomer, philosopher and mathematician who had considerable moral authority in the city of Alexandria, and who had extensive influence. At the time of her death, she was probably over sixty years of age. Indeed, many students from wealthy and influential families came to Alexandria purposely to study privately with Hypatia, and many of these later attained high posts in government and the Church. Several Christians thought that Hypatia's influence had caused Orestes to reject all conciliatory offerings by Cyril. Modern historians think that Orestes had cultivated his relationship with Hypatia to strengthen a bond with the pagan community of Alexandria, as he had done with the Jewish one, in order to better manage the tumultuous political life of the Egyptian capital. A mob, led by a lector named Peter, took Hypatia from her chariot and murdered her, hacking her body apart and burning the pieces outside the city walls. Neoplatonist historian Damascius ( 458 –  538) was "anxious to exploit the scandal of Hypatia's death", and attributed responsibility for her murder to Bishop Cyril and his Christian followers. Damascius's account of the Christian murder of Hypatia is the sole historical source attributing direct responsibility to Bishop Cyril. Some modern studies represent Hypatia's death as the result of a struggle between two Christian factions, the moderate Orestes, supported by Hypatia, and the more rigid Cyril. According to lexicographer William Smith, "She was accused of too much familiarity with Orestes, prefect of Alexandria, and the charge spread among the clergy, who took up the notion that she interrupted the friendship of Orestes with their archbishop, Cyril." Scholasticus writes that Hypatia ultimately fell "victim to the political jealousy which at the time prevailed". News of Hypatia's murder provoked great public denunciation, not only of Cyril but of the whole Alexandrian Christian community. Another major conflict was between the Alexandrian and Antiochian schools of ecclesiastical reflection, piety, and discourse. This long running conflict widened with the third canon of the First Council of Constantinople which granted the see of Constantinople primacy over the older sees of Alexandria and Antioch. Thus, the struggle between the sees of Alexandria and Antioch now included Constantinople. The conflict came to a head in 428 after Nestorius, who originated in Antioch, was made Archbishop of Constantinople. Cyril gained an opportunity to restore Alexandria's pre-eminence over both Antioch and Constantinople when an Antiochine priest who was in Constantinople at Nestorius' behest began to preach against calling Mary the "Mother of God" ("Theotokos"). As the term "Mother of God" had long been attached to Mary, the laity in Constantinople complained against the priest. Rather than repudiating the priest, Nestorius intervened on his behalf. Nestorius argued that Mary was neither a "Mother of Man" nor "Mother of God" as these referred to Christ's two natures; rather, Mary was the "Mother of Christ" (Greek: "Christotokos"). Christ, according to Nestorius, was the conjunction of the Godhead with his "temple" (which Nestorius was fond of calling his human nature). The controversy seemed to be centered on the issue of the suffering of Christ. Cyril maintained that the Son of God or the divine Word, truly suffered "in the flesh." However, Nestorius claimed that the Son of God was altogether incapable of suffering, even within his union with the flesh. Eusebius of Dorylaeum went so far as to accuse Nestorius of adoptionism. By this time, news of the controversy in the capital had reached Alexandria. At Easter 429 A.D., Cyril wrote a letter to the Egyptian monks warning them of Nestorius' views. A copy of this letter reached Constantinople where Nestorius preached a sermon against it. This began a series of letters between Cyril and Nestorius which gradually became more strident in tone. Finally, Emperor Theodosius II convoked the Council of Ephesus (in 431) to solve the dispute. Cyril selected Ephesus as the venue since it supported the veneration of Mary. The council was convoked before Nestorius's supporters from Antioch and Syria had arrived and thus Nestorius refused to attend when summoned. Predictably, the Council ordered the deposition and exile of Nestorius for heresy. However, when John of Antioch and the other pro-Nestorius bishops finally reached Ephesus, they assembled their own Council, condemned Cyril for heresy, deposed him from his see, and labelled him as a "monster, born and educated for the destruction of the church". Theodosius, by now old enough to hold power by himself, annulled the verdict of the Council and arrested Cyril, but Cyril eventually escaped. Having fled to Egypt, Cyril bribed Theodosius' courtiers, and sent a mob led by Dalmatius, a hermit, to besiege Theodosius' palace, and shout abuse; the Emperor eventually gave in, sending Nestorius into minor exile (Upper Egypt). Cyril died about 444, but the controversies were to continue for decades, from the "Robber Synod" of Ephesus (449) to the Council of Chalcedon (451) and beyond. Cyril regarded the embodiment of God in the person of Jesus Christ to be so mystically powerful that it spread out from the body of the God-man into the rest of the race, to reconstitute human nature into a graced and deified condition of the saints, one that promised immortality and transfiguration to believers. Nestorius, on the other hand, saw the incarnation as primarily a moral and ethical example to the faithful, to follow in the footsteps of Jesus. Cyril's constant stress was on the simple idea that it was God who walked the streets of Nazareth (hence Mary was "Theotokos", meaning "God bearer", which became in Latin "Mater Dei or Dei Genetrix", or Mother of God), and God who had appeared in a transfigured humanity. Nestorius spoke of the distinct "Jesus the man" and "the divine Logos" in ways that Cyril thought were too dichotomous, widening the ontological gap between man and God in a way that some of his contemporaries believed would annihilate the person of Christ. The main issue that prompted this dispute between Cyril and Nestorius was the question which arose at the Council of Constantinople: What exactly was the being to which Mary gave birth? Cyril affirmed that the Holy Trinity consists of a singular divine nature, essence, and being ("ousia") in three distinct aspects, instantiations, or subsistencies of being ("hypostases"). These distinct hypostases are the Father or God in Himself, the Son or Word ("Logos"), and the Holy Spirit. Then, when the Son became flesh and entered the world, the pre-Incarnate divine nature and assumed human nature both remained, but became "united" in the person of Jesus. This resulted in the miaphysite slogan "One Nature united out of two" being used to encapsulate the theological position of this Alexandrian bishop. According to Cyril's theology, there were two states for the Son of God: the state that existed "prior" to the Son (or Word/Logos) becoming enfleshed in the person of Jesus and the state that actually became enfleshed. The Logos Incarnate suffered and died on the Cross, and therefore the Son was able to suffer without suffering. Cyril passionately argued for the continuity of a single subject, God the Word, from the pre-Incarnate state to the Incarnate state. The divine Logos was really present in the flesh and in the world—not merely bestowed upon, semantically affixed to, or morally associated with the man Jesus, as the adoptionists and, he believed, Nestorius had taught. Cyril of Alexandria became noted in Church history because of his spirited fight for the title "Theotokos" during the First Council of Ephesus (431). His writings include the homily given in Ephesus and several other sermons. Some of his alleged homilies are in dispute as to his authorship. In several writings, Cyril focuses on the love of Jesus to his mother. On the Cross, he overcomes his pain and thinks of his mother. At the wedding in Cana, he bows to her wishes. Cyril created the basis for all other mariological developments through his teaching of the blessed Virgin Mary, as the "Mother of God." The conflict with Nestorius was mainly over this issue, and it has often been misunderstood. "[T]he debate was not so much about Mary as about Jesus. The question was not what honors were due to Mary, but how one was to speak of the birth of Jesus." St. Cyril received an important recognition of his preachings by the Second Council of Constantinople (553 d.C.) which declared; Cyril was a scholarly archbishop and a prolific writer. In the early years of his active life in the Church he wrote several exegetical documents. Among these were: "Commentaries on the Old Testament", "Thesaurus", "Discourse Against Arians", "Commentary on St. John's Gospel", and "Dialogues on the Trinity". In 429 as the Christological controversies increased, the output of his writings was so extensive that his opponents could not match it. His writings and his theology have remained central to the tradition of the Fathers and to all Orthodox to this day.
https://en.wikipedia.org/wiki?curid=7383
Hanukkah Hanukkah ( ; ', Tiberian: ', usually spelled , pronounced in Modern Hebrew, or in Yiddish; a transliteration also romanized as Chanukah or Ḥanukah) is a Jewish festival commemorating the rededication of the Second Temple in Jerusalem at the time of the Maccabean Revolt against the Seleucid Empire. It is also known as the Festival of Lights (, ""). Hanukkah is observed for eight nights and days, starting on the 25th day of Kislev according to the Hebrew calendar, which may occur at any time from late November to late December in the Gregorian calendar. The festival is observed by lighting the candles of a candelabrum with nine branches, called a menorah (or hanukkiah). One branch is typically placed above or below the others and its candle is used to light the other eight candles. This unique candle is called the "shamash" (, "attendant"). Each night, one additional candle is lit by the "shamash" until all eight candles are lit together on the final night of the festival. Other Hanukkah festivities include playing the game of dreidel and eating oil-based foods, such as latkes and sufganiyot, and dairy foods. Since the 1970s, the worldwide Chabad Hasidic movement has initiated public menorah lightings in open public places in many countries. Although a relatively minor holiday in strictly religious terms, Hanukkah has attained major cultural significance in North America and elsewhere among secular Jews as a Jewish alternative to Christmas, and is often celebrated correspondingly fervently. The name "Hanukkah" derives from the Hebrew verb "", meaning "to dedicate". On Hanukkah, the Maccabean Jews regained control of Jerusalem and rededicated the Temple. Many homiletical explanations have been given for the name: In Hebrew, the word Hanukkah is written or (). It is most commonly transliterated to English as "Hanukkah" or "". The former spelling ("Hanukkah"), which is based on using characters of the English alphabet as symbols to re-create the word's correct spelling in Hebrew, is the most common and the preferred choice of Merriam–Webster, "Collins English Dictionary", the "Oxford Style Manual", and the style guides of "The New York Times" and "The Guardian". The sound represented by "Ch" (, similar to the Scottish pronunciation of "loch") is not native to the English language. Furthermore, the letter "ḥeth" (), which is the first letter in the Hebrew spelling, is pronounced differently in modern Hebrew (voiceless uvular fricative) from in classical Hebrew (voiceless pharyngeal fricative ), and neither of those sounds is unambiguously representable in English spelling. However, its original sound is closer to the English "H" than to the Scottish "Ch", and "Hanukkah" more accurately represents the spelling in the Hebrew alphabet. Moreover, the 'kaf' consonant is geminate in classical (but not modern) Hebrew. Adapting the classical Hebrew pronunciation with the geminate and pharyngeal can lead to the spelling "Hanukkah", while adapting the modern Hebrew pronunciation with no gemination and uvular leads to the spelling . The story of Hanukkah is preserved in the books of the First and Second Maccabees, which describe in detail the re-dedication of the Temple in Jerusalem and the lighting of the menorah. These books are not part of the canonized Tanakh (Hebrew Bible) used by modern Jews, though the Catholic and Orthodox Churches consider them part of the Old Testament. The eight-day rededication of the temple is described in , though the name of the festival and the miracle of the lights do not appear here. A story similar in character, and older in date, is the one alluded to in according to which the relighting of the altar fire by Nehemiah was due to a miracle which occurred on the 25th of Kislev, and which appears to be given as the reason for the selection of the same date for the rededication of the altar by Judah Maccabee. The above account in 1 Maccabees 4, as well as portrays the feast as a delayed observation of the eight-day Feast of Booths (Sukkot)"; similarly explains the length of the feast as "in the manner of the Feast of Booths". Megillat Taanit (1st century) contains a list of festive days on which fasting or eulogizing is forbidden. It specifies, "On the 25th of [Kislev] is Hanukkah of eight days, and one is not to eulogize" but gives no further details. The Mishna (late 2nd century) mentions Hanukkah in several places, but never describes its laws in detail and never mentions any aspect of the history behind it. To explain the Mishna's lack of a systematic discussion of Hanukkah, Rav Nissim Gaon postulated that information on the holiday was so commonplace that the Mishna felt no need to explain it. Modern scholar Reuvein Margolies suggests that as the Mishnah was redacted after the Bar Kochba revolt, its editors were reluctant to include explicit discussion of a holiday celebrating another relatively recent revolt against a foreign ruler, for fear of antagonizing the Romans. The miracle of the one-day supply of oil miraculously lasting eight days is first described in the Talmud, committed to writing about 600 years after the events described in the books of Maccabees. The Talmud says that after the forces of Antiochus IV had been driven from the Temple, the Maccabees discovered that almost all of the ritual olive oil had been profaned. They found only a single container that was still sealed by the High Priest, with enough oil to keep the menorah in the Temple lit for a single day. They used this, yet it burned for eight days (the time it took to have new oil pressed and made ready). The Talmud presents three options: Except in times of danger, the lights were to be placed outside one's door, on the opposite side of the mezuza, or in the window closest to the street. Rashi, in a note to "Shabbat 21b," says their purpose is to publicize the miracle. The blessings for Hanukkah lights are discussed in tractate "Succah," p. 46a. Megillat Antiochus (probably composed in the 2nd century) concludes with the following words: The Al HaNissim prayer is recited on Hanukkah as an addition to the Amidah prayer, which was formalized in the late 1st century. "Al HaNissim" describes the history of the holiday as follows: The Jewish historian Titus Flavius Josephus narrates in his book, Jewish Antiquities XII, how the victorious Judas Maccabeus ordered lavish yearly eight-day festivities after rededicating the Temple in Jerusalem that had been profaned by Antiochus IV Epiphanes. Josephus does not say the festival was called Hanukkah but rather the "Festival of Lights": In the New Testament, John 10:22–23 says, "Then came the Festival of Dedication at Jerusalem. It was winter, and Jesus was in the temple courts walking in Solomon’s Colonnade" (NIV). The Greek noun used appears in the neuter plural as "the renewals" or "the consecrations" (Greek: τὰ ἐγκαίνια; "ta enkaínia"). The same root appears in 2 Esdras 6:16 in the Septuagint to refer specifically to Hanukkah. This Greek word was chosen because the Hebrew word for "consecration" or "dedication" is "Hanukkah" (חנכה). The Aramaic New Testament uses the Aramaic word "Khawdata" (a close synonym), which literally means "renewal" or "to make new." Judea was part of the Ptolemaic Kingdom of Egypt until 200 BCE when King Antiochus III the Great of Syria defeated King Ptolemy V Epiphanes of Egypt at the Battle of Panium. Judea then became part of the Seleucid Empire of Syria. King Antiochus III the Great, wanting to conciliate his new Jewish subjects, guaranteed their right to "live according to their ancestral customs" and to continue to practice their religion in the Temple of Jerusalem. However, in 175 BCE, Antiochus IV Epiphanes, the son of Antiochus III, invaded Judea, at the request of the sons of Tobias. The Tobiads, who led the Hellenizing Jewish faction in Jerusalem, were expelled to Syria around 170 BCE when the high priest Onias and his pro-Egyptian faction wrested control from them. The exiled Tobiads lobbied Antiochus IV Epiphanes to recapture Jerusalem. As Flavius Josephus relates: When the Second Temple in Jerusalem was looted and services stopped, Judaism was outlawed. In 167 BCE, Antiochus ordered an altar to Zeus erected in the Temple. He banned brit milah (circumcision) and ordered pigs to be sacrificed at the altar of the temple. Antiochus's actions provoked a large-scale revolt. Mattathias (Mattityahu), a Jewish priest, and his five sons Jochanan, Simeon, Eleazar, Jonathan, and Judah led a rebellion against Antiochus. It started with Mattathias killing first a Jew who wanted to comply with Antiochus's order to sacrifice to Zeus, and then a Greek official who was to enforce the government's behest (1 Mac. 2, 24–25). Judah became known as Yehuda HaMakabi ("Judah the Hammer"). By 166 BCE Mattathias had died, and Judah took his place as leader. By 165 BCE the Jewish revolt against the Seleucid monarchy was successful. The Temple was liberated and rededicated. The festival of Hanukkah was instituted to celebrate this event. Judah ordered the Temple to be cleansed, a new altar to be built in place of the polluted one and new holy vessels to be made. According to the Talmud, unadulterated and undefiled pure olive oil with the seal of the kohen gadol (high priest) was needed for the menorah in the Temple, which was required to burn throughout the night every night. The story goes that one flask was found with only enough oil to burn for one day, yet it burned for eight days, the time needed to prepare a fresh supply of kosher oil for the menorah. An eight-day festival was declared by the Jewish sages to commemorate this miracle. The version of the story in 1 Maccabees states that an eight-day celebration of songs and sacrifices was proclaimed upon re-dedication of the altar, and makes no specific mention of the miracle of the oil. Some modern scholars argue that the king was intervening in an internal civil war between the Maccabean Jews and the Hellenized Jews in Jerusalem. These competed violently over who would be the High Priest, with traditionalists with Hebrew/Aramaic names like Onias contesting with Hellenizing High Priests with Greek names like Jason and Menelaus. In particular, Jason's Hellenistic reforms would prove to be a decisive factor leading to eventual conflict within the ranks of Judaism. Other authors point to possible socioeconomic reasons in addition to the religious reasons behind the civil war. What began in many respects as a civil war escalated when the Hellenistic kingdom of Syria sided with the Hellenizing Jews in their conflict with the traditionalists. As the conflict escalated, Antiochus took the side of the Hellenizers by prohibiting the religious practices the traditionalists had rallied around. This may explain why the king, in a total departure from Seleucid practice in all other places and times, banned a traditional religion. The miracle of the oil is widely regarded as a legend and its authenticity has been questioned since the Middle Ages. However, given the famous question Rabbi Yosef Karo posed concerning why Hanukkah is celebrated for eight days when the miracle was only for seven days (since there was enough oil for one day), it was clear that he believed it was a historical event. This belief has been adopted by most of Orthodox Judaism, in as much as Rabbi Karo's "Shulchan Aruch" is a main Code of Jewish Law. Selected battles between the Maccabees and the Seleucid Syrian-Greeks: Hanukkah is celebrated with a series of rituals that are performed every day throughout the 8-day holiday, some are family-based and others communal. There are special additions to the daily prayer service, and a section is added to the blessing after meals. Hanukkah is not a "Sabbath-like" holiday, and there is no obligation to refrain from activities that are forbidden on the Sabbath, as specified in the "Shulkhan Arukh". Adherents go to work as usual but may leave early in order to be home to kindle the lights at nightfall. There is no religious reason for schools to be closed, although in Israel schools close from the second day for the whole week of Hanukkah. Many families exchange gifts each night, such as books or games, and "Hanukkah Gelt" is often given to children. Fried foods (such as latkes (potato pancakes), jelly doughnuts (sufganiyot), and Sephardic bimuelos) are eaten to commemorate the importance of oil during the celebration of Hanukkah. Some also have a custom of eating dairy products to remember Judith and how she overcame Holofernes by feeding him cheese, which made him thirsty, and giving him wine to drink. When Holofernes became very drunk, Judith cut off his head. Each night throughout the 8 day holiday, a candle or oil-based light is lit. As a universally practiced "beautification" (hiddur mitzvah) of the mitzvah, the number of lights lit is increased by one each night. An extra light called a "shamash", meaning "attendant" or "sexton," is also lit each night, and is given a distinct location, usually higher, lower, or to the side of the others. Among Ashkenazim the tendency is for every male member of the household (and in many families, girls as well) to light a full set of lights each night, while among Sephardim the prevalent custom is to have one set of lights for the entire household. The purpose of the "shamash" is to adhere to the prohibition, specified in the Talmud, against using the Hanukkah lights for anything other than publicizing and meditating on the Hanukkah miracle. This differs from Sabbath candles which are meant to be used for illumination and lighting. Hence, if one were to need extra illumination on Hanukkah, the "shamash" candle would be available, and one would avoid using the prohibited lights. Some, especially Ashkenazim, light the "shamash" candle first and then use it to light the others. So altogether, including the "shamash", two lights are lit on the first night, three on the second and so on, ending with nine on the last night, for a total of 44 (36, excluding the "shamash"). It is Sephardic custom not to light the shamash first and use it to light the rest. Instead, the shamash candle is the last to be lit, and a different candle or a match is used to light all the candles. Some Hasidic Jews follow this Sephardic custom as well. The lights can be candles or oil lamps. Electric lights are sometimes used and are acceptable in places where open flame is not permitted, such as a hospital room, or for the very elderly and infirm; however, those who permit reciting a blessing over electric lamps only allow it if it is incandescent and battery operated (an incandescent flashlight would be acceptable for this purpose), while a blessing may not be recited over a plug-in menorah or lamp. Most Jewish homes have a special candelabrum referred to as either a "Chanukiah" (the modern Israeli term) or a "menorah" (the traditional name, simply Hebrew for 'lamp'). Many families use an oil lamp (traditionally filled with olive oil) for Hanukkah. Like the candle Chanukiah, it has eight wicks to light plus the additional "shamash" light. In the United States, Hanukkah became a more visible festival in the public sphere from the 1970s when Rabbi Menachem M. Schneerson called for public awareness and observance of the festival and encouraged the lighting of public menorahs. Diane Ashton attributed the increased visibility and reinvention of Hanukkah by some of the American Jewish community as a way to adapt to American life, re-inventing the festival in "the language of individualism and personal conscience derived from both Protestantism and the Enlightenment". The reason for the Hanukkah lights is not for the "lighting of the house within", but rather for the "illumination of the house without," so that passersby should see it and be reminded of the holiday's miracle (i.e. that the sole cruse of pure oil found which held enough oil to burn for one night actually burned for eight nights). Accordingly, lamps are set up at a prominent window or near the door leading to the street. It is customary amongst some Ashkenazi Jews to have a separate menorah for each family member (customs vary), whereas most Sephardi Jews light one for the whole household. Only when there was danger of antisemitic persecution were lamps supposed to be hidden from public view, as was the case in Persia under the rule of the Zoroastrians, or in parts of Europe before and during World War II. However, most Hasidic groups light lamps near an inside doorway, not necessarily in public view. According to this tradition, the lamps are placed on the opposite side from the "mezuzah", so that when one passes through the door s/he is surrounded by the holiness of "mitzvot" (the commandments). Generally, women are exempt in Jewish law from time-bound positive commandments, although the Talmud requires that women engage in the mitzvah of lighting Hanukkah candles "for they too were involved in the miracle." Hanukkah lights should usually burn for at least half an hour after it gets dark. The custom of many is to light at sundown, although most Hasidim light later. Many Hasidic Rebbes light much later to fulfill the obligation of publicizing the miracle by the presence of their Hasidim when they kindle the lights. Inexpensive small wax candles sold for Hanukkah burn for approximately half an hour so should be lit no earlier than nightfall. Friday night presents a problem, however. Since candles may not be lit on Shabbat itself, the candles must be lit before sunset. However, they must remain lit through the lighting of the Shabbat candles. Therefore, the Hanukkah menorah is lit first with larger candles than usual, followed by the Shabbat candles. At the end of the Shabbat, there are those who light the Hanukkah lights before Havdalah and those who make Havdalah before the lighting Hanukkah lights. If for whatever reason one didn't light at sunset or nightfall, the lights should be kindled later, as long as there are people in the streets. Later than that, the lights should still be kindled, but the blessings should be recited only if there is at least somebody else awake in the house and present at the lighting of the Hannukah lights. Typically two blessings ("brachot"; singular: "brachah") are recited during this eight-day festival when lighting the candles. On the first night, the shehecheyanu blessing is added, making a total of three blessings. The blessings are said before or after the candles are lit depending on tradition. On the first night of Hanukkah one light (candle or oil) is lit on the right side of the menorah, on the following night a second light is placed to the left of the first but it is lit first, and so on, proceeding from placing candles right to left but lighting them from left to right over the eight nights. Transliteration: Translation: "Blessed are You, our God, King of the universe, Who has sanctified us with His commandments and commanded us to kindle the Hanukkah light[s]." Transliteration: Translation: "Blessed are You, LORD our God, King of the universe, Who performed miracles for our ancestors in those days at this time..." After the lights are kindled the hymn "Hanerot Halalu" is recited. There are several different versions; the version presented here is recited in many Ashkenazic communities: In the Ashkenazi tradition, each night after the lighting of the candles, the hymn Ma'oz Tzur is sung. The song contains six stanzas. The first and last deal with general themes of divine salvation, and the middle four deal with events of persecution in Jewish history, and praises God for survival despite these tragedies (the exodus from Egypt, the Babylonian captivity, the miracle of the holiday of Purim, the Hasmonean victory), and a longing for the days when Judea will finally triumph over Rome. The song was composed in the thirteenth century by a poet only known through the acrostic found in the first letters of the original five stanzas of the song: Mordechai. The familiar tune is most probably a derivation of a German Protestant church hymn or a popular folk song. After lighting the candles and Ma'oz Tzur, singing other Hanukkah songs is customary in many Jewish homes. Some Hasidic and Sephardi Jews recite Psalms, such as , , and . In North America and in Israel it is common to exchange presents or give children presents at this time. In addition, many families encourage their children to give tzedakah (charity) in lieu of presents for themselves. An addition is made to the ""hoda'ah"" (thanksgiving) benediction in the Amidah (thrice-daily prayers), called "Al HaNissim" ("On/about the Miracles"). This addition refers to the victory achieved over the Syrians by the Hasmonean Mattathias and his sons. The same prayer is added to the grace after meals. In addition, the "Hallel" (praise) Psalms ( – ) are sung during each morning service and the "Tachanun" penitential prayers are omitted. The Torah is read every day in the shacharit morning services in synagogue, on the first day beginning from (according to some customs, ), and the last day ending with . Since Hanukkah lasts eight days it includes at least one, and sometimes two, Jewish Sabbaths (Saturdays). The weekly Torah portion for the first Sabbath is almost always "Miketz", telling of Joseph's dream and his enslavement in Egypt. The "Haftarah" reading for the first Sabbath Hanukkah is – . When there is a second Sabbath on Hanukkah, the "Haftarah" reading is from – . The Hanukkah "menorah" is also kindled daily in the synagogue, at night with the blessings and in the morning without the blessings. The menorah is not lit during Shabbat, but rather prior to the beginning of Shabbat as described above and not at all during the day. During the Middle Ages "Megillat Antiochus" was read in the Italian synagogues on Hanukkah just as the Book of Esther is read on Purim. It still forms part of the liturgy of the Yemenite Jews. The last day of Hanukkah is known by some as "Zot Hanukkah" and by others as "Chanukat HaMizbeach", from the verse read on this day in the synagogue , "Zot Hanukkat Hamizbe'ach": "This was the dedication of the altar". According to the teachings of Kabbalah and Hasidism, this day is the final "seal" of the High Holiday season of Yom Kippur and is considered a time to repent out of love for God. In this spirit, many Hasidic Jews wish each other "Gmar chatimah tovah" ("may you be sealed totally for good"), a traditional greeting for the Yom Kippur season. It is taught in Hasidic and Kabbalistic literature that this day is particularly auspicious for the fulfillment of prayers. It is customary for women not to work for at least the first half-hour of the candles' burning, and some have the custom not to work for the entire time of burning. It is also forbidden to fast or to eulogize during Hanukkah. A large number of songs have been written on Hanukkah themes, perhaps more so than for any other Jewish holiday. Some of the best known are ""Ma'oz Tzur"" (Rock of Ages), ""Latke'le Latke'le"" (Yiddish song about cooking Latkes), ""Hanukkiah Li Yesh"" ("I Have a Hanukkah Menorah"), ""Ocho Kandelikas"" ("Eight Little Candles"), ""Kad Katan"" ("A Small Jug"), ""S'vivon Sov Sov Sov"" ("Dreidel, Spin and Spin"), ""Haneirot Halolu"" ("These Candles which we light"), ""Mi Yimalel"" ("Who can Retell") and ""Ner Li, Ner Li"" ("I have a Candle"). Among the most well known songs in English-speaking countries are "Dreidel, Dreidel, Dreidel" and "Oh Chanukah". Among the Rebbes of the Nadvorna Hasidic dynasty, it is customary for the Rebbes to play violin after the menorah is lit. Penina Moise's Hannukah Hymn published in the 1842 "Hymns Written for the Use of Hebrew Congregations" was instrumental in the beginning of Americanization of Hanukkah. There is a custom of eating foods fried or baked in oil (preferably olive oil) to commemorate the miracle of a small flask of oil keeping the Second Temple's Menorah alight for eight days. Traditional foods include latkes, a fried potato fritter, especially among Ashkenazi families. Sephardi, Polish, and Israeli families eat jam-filled doughnuts called sufganiyot, which were ( "pontshkes") by Ashkenazi Jews living in Eastern and Central Europe prior to the Holocaust, bimuelos (spherical doughnuts) and sufganiyot which are deep-fried in oil. Italkim and Hungarian Jews traditionally eat cheese pancakes known as "cassola" or "cheese latkes". Latkes are not popular in Israel, as they are more commonly made at home and are an Ashkenazi Jewish dish. The Sephardi Jews eat fritas de prasa, a similar fried dish made with mashed potato and leek. As the majority of the population in Israel is of Sephardi and Mizrahi Jewish descent, and these groups have their own Hanukkah dishes such as fritas de prasa, sfinj, cassola, and shamlias, among others. Latkes have also been largely replaced by sufganiyot due to local economic factors, convenience and the influence of trade unions. Bakeries in Israel have popularized many new types of fillings for "sufganiyot" besides the traditional strawberry jelly filling, including chocolate cream, vanilla cream, caramel, cappuccino and others. In recent years, downsized, "mini" sufganiyot containing half the calories of the regular, 400-to-600-calorie version, have become popular. Rabbinic literature also records a tradition of eating cheese and other dairy products during Hanukkah. This custom, as mentioned above, commemorates the heroism of Judith during the Babylonian captivity of the Jews and reminds us that women also played an important role in the events of Hanukkah. The deuterocanonical book of Judith (Yehudit or Yehudis in Hebrew), which is not part of the Tanakh, records that Holofernes, an Assyrian general, had surrounded the village of Bethulia as part of his campaign to conquer Judea. After intense fighting, the water supply of the Jews was cut off and the situation became desperate. Judith, a pious widow, told the city leaders that she had a plan to save the city. Judith went to the Assyrian camps and pretended to surrender. She met Holofernes, who was smitten by her beauty. She went back to his tent with him, where she plied him with cheese and wine. When he fell into a drunken sleep, Judith beheaded him and escaped from the camp, taking the severed head with her (the beheading of Holofernes by Judith has historically been a popular theme in art). When Holofernes' soldiers found his corpse, they were overcome with fear; the Jews, on the other hand, were emboldened and launched a successful counterattack. The town was saved, and the Assyrians defeated. Roast goose has historically been a traditional Hanukkah food among Eastern European and American Jews, although the custom has declined in recent decades.
https://en.wikipedia.org/wiki?curid=7388
Christian views on marriage From the earliest days of the Christian faith, Christians have honored marriage, or "holy matrimony", as a divinely blessed, lifelong, monogamous union, between a man and a woman. According to the Episcopal Book of Common Prayer (1979), reflecting the traditional view, "Christian marriage is a solemn and public covenant between a man and a woman in the presence of God," "intended by God for their mutual joy; for the help and comfort given one another in prosperity and adversity; and, when it is God's will, for the procreation of children and their nurture." However, while many Christians might agree with the traditional definition, the terminology and theological views of marriage have varied through time in different countries, and among Christian denominations. Many Protestants consider marriage to be a sacred institution or "holy ordinance" of God. Roman Catholics and Eastern Orthodox Christians consider marriage a holy sacrament or sacred mystery. However, there have been and are differing attitudes among denominations and individual Christians towards not only the concept of Christian marriage, but also concerning divorce, remarriage, gender roles, family authority (the "headship" of the husband), the legal status of married women, birth control, marriageable age, cousin marriage, marriage of in-laws, interfaith marriage, same-sex marriage, and polygamy, among other topics, so that in the 21st century there cannot be said to be a single, uniform, worldwide view of marriage among all who profess to be Christians. Christian teaching has never held that marriage is necessary for everyone; for many centuries in Western Europe, priestly or monastic celibacy was valued as highly as, if not higher than, marriage. Christians who did not marry were expected to refrain from all sexual activity, as were those who took holy orders or monastic vows . In some Western countries, a separate and secular civil wedding ceremony is required for recognition by the state, while in other Western countries, couples must merely obtain a marriage license from a local government authority and can be married by Christian or other clergy if they are authorized by law to conduct weddings. In this case, the state recognizes the religious marriage as a civil marriage as well; and Christian couples married in this way have all the rights of civil marriage, including, for example, divorce, even if their church forbids divorce. Since the beginning of the 21st century, same-sex couples have been allowed to marry civilly in many countries, and some Christian churches in those countries allow religious marriage of same-sex couples, though others forbid it, along with all other same-sex relationships. Christians believe that marriage is considered in its ideal according to the purpose of God. At the heart of God's design for marriage is companionship and intimacy. The biblical picture of marriage expands into something much broader, with the husband and wife relationship illustrating the relationship between Christ and the church. It is also considered in its actual occurrence, sometimes involving failure. Therefore, the Bible speaks on the subject of divorce. The New Testament recognizes a place for singleness. Salvation within Christianity is not dependent on the continuation of a biological lineage. Polygyny, or men having multiple wives at once, is one of the most common marital arrangements represented in the Old Testament, yet scholars doubt that it was common among average Israelites because of the wealth needed to practice it. Both the biblical patriarchs and kings of Israel are described as engaged in polygamous relationships. Betrothal ("erusin"), which is merely a binding promise to get married, is distinct from marriage itself ("nissu'in"), with the time between these events varying substantially. Since a wife was regarded as property in biblical times, the betrothal ("erusin") was effected simply by purchasing her from her father (or guardian); the girl's consent is not explicitly required by any biblical law. Like the adjacent Arabic culture (in the pre-Islamic period), the act of marriage appears mainly to have consisted of the groom fetching the bride, although among the Israelites the procession was a festive occasion, accompanied by music, dancing, and lights. To celebrate the marriage, week-long feasts were sometimes held. In Old Testament times, a wife was regarded as chattel, belonging to her husband. The descriptions of the Bible suggest that she would be expected to perform tasks such as spinning, sewing, weaving, manufacture of clothing, fetching of water, baking of bread, and animal husbandry. However, wives were usually looked after with care, and bigamous men were expected to ensure that they give their first wife food, clothing, and sexual activity. Since a wife was regarded as property, her husband was originally free to divorce her with little restriction, at any time. A divorced couple could get back together unless the wife had married someone else after her divorce. The Bible clearly addresses marriage and divorce. Those in troubled marriages are encouraged to seek counseling and restoration because most divorces are neither necessary nor unavoidable. In both Matthew and Mark, Jesus appealed to God's will in creation. He builds upon the narratives in where male and female are created together and for one another. Thus Jesus takes a firm stand on the permanence of marriage in the original will of God. This corresponds closely with the position of the Pharisee school of thought led by Shammai, at the start of the first millennium, with which Jesus would have been familiar. By contrast, Rabbinic Judaism subsequently took the opposite view, espoused by Hillel, the leader of the other major Pharisee school of thought at the time; in Hillel's view, men were allowed to divorce their wives for any reason. Some hold that marriage vows are unbreakable, so that even in the distressing circumstances in which a couple separates, they are still married from God's point of view. This is the Roman Catholic church's position, although occasionally the church will declare a marriage to be “null” (in other words, it never really was a marriage). William Barclay (1907-1978) has written: Jesus brought together two passages from Genesis, reinforcing the basic position on marriage found in Jewish scripture. Thus, he implicitly emphasized that it is God-made ("God has joined together"), "male and female," lifelong ("let no one separate"), and monogamous ("a man…his wife"). Jesus used the image of marriage and the family to teach the basics about the Kingdom of God. He inaugurated his ministry by blessing the wedding feast at Cana. In the Sermon on the Mount he set forth a new commandment concerning marriage, teaching that lustful looking constitutes adultery. He also superseded a Mosaic Law allowing divorce with his teaching that "…anyone who divorces his wife, except for sexual immorality (Gk. "porneia"), causes her to become an adulteress, and anyone who marries the divorced woman commits adultery". Similar Pauline teachings are found in . The exception clause—"except for…"—uses the Greek word "porneia" which is variously translated "fornication" (KJV), "marital unfaithfulness" (NIV 1984), "sexual immorality" (NIV 2011), "unchastity" (RSV), "et al". "The KJV New Testament Greek Lexicon, KJV" says "porneia" includes a variety of sexual "deviations" to include "illicit sexual intercourse, adultery, fornication, homosexuality, lesbianism, intercourse with animals, etc., sexual intercourse with close relatives…." Theologian Frank Stagg says that manuscripts disagree as to the presence in the original text of the phrase "except for fornication". Stagg writes: "Divorce always represents failure…a deviation from God's will…. There is grace and redemption where there is contrition and repentance…. There is no clear authorization in the New Testament for remarriage after divorce." Stagg interprets the chief concern of as being "to condemn the criminal act of the man who divorces an innocent wife…. Jesus was rebuking the husband who victimizes an innocent wife and thinks that he makes it right with her by giving her a divorce". He points out that Jesus refused to be trapped by the Pharisees into choosing between the strict and liberal positions on divorce as held at the time in Judaism. When they asked him, "Is it lawful for a man to divorce his wife for any cause?" he answered by reaffirming God's will as stated in Genesis and , that in marriage husband and wife are made "one flesh", and what God has united man must not separate. There is no evidence that Jesus himself ever married, and considerable evidence that he remained single. In contrast to Judaism and many other traditions, he taught that there is a place for voluntary singleness in Christian service. He believed marriage could be a distraction from an urgent mission, that he was living in a time of crisis and urgency where the Kingdom of God would be established where there would be no marriage nor giving in marriage: In Jesus is asked about the continuing state of marriage after death and he affirms that at the resurrection "people will neither marry nor be given in marriage; they will be like the angels in heaven." The Apostle Paul quoted passages from Genesis almost verbatim in two of his New Testament books. He used marriage not only to describe the kingdom of God, as Jesus had done, but to define also the nature of the 1st-century Christian church. His theological view was a Christian development of the Old Testament parallel between marriage and the relationship between God and Israel. He analogized the church as a bride and Christ as the bridegroom─drawing parallels between Christian marriage and the relationship between Christ and the Church. There is no hint in the New Testament that Jesus was ever married, and no clear evidence that Paul was ever married. However, both Jesus and Paul seem to view marriage as a legitimate calling from God for Christians. Paul elevates singleness to that of the preferable position, but does offer a caveat suggesting this is "because of the impending crisis"—which could itself extend to present times (see also Pauline privilege). Paul's primary issue was that marriage adds concerns to one's life that detract from their ability to serve God without distraction. Some scholars have speculated that Paul may have been a widower since prior to his conversion to Christianity he was a Pharisee and member of the Sanhedrin, positions in which the social norm of the day required the men to be married. But it is just as likely that he never married at all. Yet, Paul acknowledges the mutuality of marital relations, and recognizes that his own singleness is "a particular gift from God" that others may not necessarily have. He writes: "Now to the unmarried and the widows I say: It is good for them to stay unmarried, as I am. But if they cannot control themselves, they should marry, for it is better to marry than to burn with passion." Paul indicates that bishops, deacons, and elders must be "husbands of one wife", and that women must have one husband. This is usually understood to legislate against polygamy rather than to require marriage: In the Roman Age, female widows who did not remarry were considered more pure than those who did. Such widows were known as "one man woman" ("enos andros gune") in the epistles of Paul. Paul writes: Paul allowed widows to remarry. Paul says that only "one-man women" older than 60 years can make the list of Christian widows who did special tasks in the community, but that younger widows should remarry to hinder sin. Building on what they saw the example of Jesus and Paul advocating, some early Church Fathers placed less value on the family and saw celibacy and freedom from family ties as a preferable state. Nicene Fathers such as Augustine believed that marriage was a sacrament because it was a symbol used by Paul to express Christ's love of the Church. However, there was also an apocalyptic dimension in his teaching, and he was clear that if everybody stopped marrying and having children that would be an admirable thing; it would mean that the Kingdom of God would return all the sooner and the world would come to an end. Such a view reflects the Manichaean past of Augustine. While upholding the New Testament teaching that marriage is "honourable in all and the bed undefiled," Augustine believed that "yet, whenever it comes to the actual process of generation, the very embrace which is lawful and honourable cannot be effected without the ardour of lust...This is the carnal concupiscence, which, while it is no longer accounted sin in the regenerate, yet in no case happens to nature except from sin." Both Tertullian and Gregory of Nyssa were church fathers who were married. They each stressed that the happiness of marriage was ultimately rooted in misery. They saw marriage as a state of bondage that could only be cured by celibacy. They wrote that at the very least, the virgin woman could expect release from the "governance of a husband and the chains of children." Tertullian argued that second marriage, having been freed from the first by death,"will have to be termed no other than a species of fornication," partly based on the reasoning that this involves desiring to marry a woman out of sexual ardor, which a Christian convert is to avoid. Also advocating celibacy and virginity as preferable alternatives to marriage, Jerome wrote: "It is not disparaging wedlock to prefer virginity. No one can make a comparison between two things if one is good and the other evil." On First Corinthians 7:1 he reasons, "It is good, he says, for a man not to touch a woman. If it is good not to touch a woman, it is bad to touch one: for there is no opposite to goodness but badness. But if it be bad and the evil is pardoned, the reason for the concession is to prevent worse evil." St. John Chrysostom wrote: "...virginity is better than marriage, however good... Celibacy is...an imitation of the angels. Therefore, virginity is as much more honorable than marriage, as the angel is higher than man. But why do I say angel? Christ, Himself, is the glory of virginity." Cyprian, Bishop of Carthage, said that the first commandment given to men was to increase and multiply, but now that the earth was full there was no need to continue this process of multiplication. This view of marriage was reflected in the lack of any formal liturgy formulated for marriage in the early Church. No special ceremonial was devised to celebrate Christian marriage—despite the fact that the Church had produced liturgies to celebrate the Eucharist, Baptism and Confirmation. It was not important for a couple to have their nuptials blessed by a priest. People could marry by mutual agreement in the presence of witnesses. At first, the old Roman pagan rite was used by Christians, although modified superficially. The first detailed account of a Christian wedding in the West dates from the 9th century. This system, known as Spousals, persisted after the Reformation. Today all Christian denominations regard marriage as a sacred institution, a covenant. Roman Catholics consider it to be a sacrament. Marriage was officially recognized as a sacrament at the 1184 Council of Verona. Before then, no specific ritual was prescribed for celebrating a marriage: "Marriage vows did not have to be exchanged in a church, nor was a priest's presence required. A couple could exchange consent anywhere, anytime." In the decrees on marriage of the Council of Trent (twenty-fourth session from 1563), the validity of marriage was made dependent upon the wedding taking place before a priest and two witnesses, although the lack of a requirement for parental consent ended a debate that had proceeded from the 12th century. In the case of a divorce, the right of the innocent party to marry again was denied so long as the other party was alive, even if the other party had committed adultery.
https://en.wikipedia.org/wiki?curid=7390
Class (computer programming) In object-oriented programming, a class is an extensible program-code-template for creating objects, providing initial values for state (member variables) and implementations of behavior (member functions or methods). In many languages, the class name is used as the name for the class (the template itself), the name for the default constructor of the class (a subroutine that creates objects), and as the type of objects generated by instantiating the class; these distinct concepts are easily conflated. When an object is created by a constructor of the class, the resulting object is called an instance of the class, and the member variables specific to the object are called instance variables, to contrast with the class variables shared across the class. In some languages, classes are only a compile-time feature (new classes cannot be declared at run-time), while in other languages classes are first-class citizens, and are generally themselves objects (typically of type or similar). In these languages, a class that creates classes is called a metaclass. In casual use, people often refer to the "class" of an object, but narrowly speaking objects have "type": the interface, namely the types of member variables, the signatures of member functions (methods), and properties these satisfy. At the same time, a class has an implementation (specifically the implementation of the methods), and can create objects of a given type, with a given implementation. In the terms of type theory, a class is an implementationa "concrete" data structure and collection of subroutineswhile a type is an interface. Different (concrete) classes can produce objects of the same (abstract) type (depending on type system); for example, the type might be implemented with two classes (fast for small stacks, but scales poorly) and (scales well but high overhead for small stacks). Similarly, a given class may have several different constructors. Class types generally represent nouns, such as a person, place or thing, or something nominalized, and a class represents an implementation of these. For example, a type might represent the properties and functionality of bananas in general, while the and classes would represent ways of producing bananas (say, banana suppliers or data structures and functions to represent and draw bananas in a video game). The class could then produce particular bananas: instances of the class would be objects of type . Often only a single implementation of a type is given, in which case the class name is often identical with the type name. Classes are composed from structural and behavioral constituents. Programming languages that include classes as a programming construct offer support, for various class-related features, and the syntax required to use these features varies greatly from one programming language to another. A class contains data field descriptions (or "properties", "fields", "data members", or "attributes"). These are usually field types and names that will be associated with state variables at program run time; these state variables either belong to the class or specific instances of the class. In most languages, the structure defined by the class determines the layout of the memory used by its instances. Other implementations are possible: for example, objects in Python use associative key-value containers. Some programming languages such as Eiffel support specification of invariants as part of the definition of the class, and enforce them through the type system. Encapsulation of state is necessary for being able to enforce the invariants of the class. The behavior of class or its instances is defined using methods. Methods are subroutines with the ability to operate on objects or classes. These operations may alter the state of an object or simply provide ways of accessing it. Many kinds of methods exist, but support for them varies across languages. Some types of methods are created and called by programmer code, while other special methods—such as constructors, destructors, and conversion operators—are created and called by compiler-generated code. A language may also allow the programmer to define and call these special methods. Every class "implements" (or "realizes") an interface by providing structure and behavior. Structure consists of data and state, and behavior consists of code that specifies how methods are implemented. There is a distinction between the definition of an interface and the implementation of that interface; however, this line is blurred in many programming languages because class declarations both define and implement an interface. Some languages, however, provide features that separate interface and implementation. For example, an abstract class can define an interface without providing implementation. Languages that support class inheritance also allow classes to inherit interfaces from the classes that they are derived from. For example, if "class A" inherits from "class B" and if "class B" implements the interface "interface B" then "class A" also inherits the functionality(constants and methods declaration) provided by "interface B". In languages that support access specifiers, the interface of a class is considered to be the set of public members of the class, including both methods and attributes (via implicit getter and setter methods); any private members or internal data structures are not intended to be depended on by external code and thus are not part of the interface. Object-oriented programming methodology dictates that the operations of any interface of a class are to be independent of each other. It results in a layered design where clients of an interface use the methods declared in the interface. An interface places no requirements for clients to invoke the operations of one interface in any particular order. This approach has the benefit that client code can assume that the operations of an interface are available for use whenever the client has access to the object. The buttons on the front of your television set are the interface between you and the electrical wiring on the other side of its plastic casing. You press the "power" button to toggle the television on and off. In this example, your particular television is the instance, each method is represented by a button, and all the buttons together compose the interface (other television sets that are the same model as yours would have the same interface). In its most common form, an interface is a specification of a group of related methods without any associated implementation of the methods. A television set also has a myriad of "attributes", such as size and whether it supports colour, which together comprise its structure. A class represents the full description of a television, including its attributes (structure) and buttons (interface). Getting the total number of televisions manufactured could be a "static method" of the television class. This method is clearly associated with the class, yet is outside the domain of each individual instance of the class. A static method that finds a particular instance out of the set of all television objects is another example. The following is a common set of access specifiers: Although many object-oriented languages support the above access specifiers, their semantics may differ. Object-oriented design uses the access specifiers in conjunction with careful design of public method implementations to enforce class invariants—constraints on the state of the objects. A common usage of access specifiers is to separate the internal data of a class from its interface: the internal structure is made private, while public accessor methods can be used to inspect or alter such private data. Access specifiers do not necessarily control "visibility", in that even private members may be visible to client external code. In some languages, an inaccessible but visible member may be referred to at run-time (for example, by a pointer returned from a member function), but an attempt to use it by referring to the name of the member from client code will be prevented by the type checker. The various object-oriented programming languages enforce member accessibility and visibility to various degrees, and depending on the language's type system and compilation policies, enforced at either compile-time or run-time. For example, the Java language does not allow client code that accesses the private data of a class to compile. In the C++ language, private methods are visible, but not accessible in the interface; however, they may be made invisible by explicitly declaring fully abstract classes that represent the interfaces of the class. Some languages feature other accessibility schemes: In addition to the design of standalone classes, programming languages may support more advanced class design based upon relationships between classes. The inter-class relationship design capabilities commonly provided are "compositional" and "hierarchical". Classes can be composed of other classes, thereby establishing a compositional relationship between the enclosing class and its embedded classes. Compositional relationship between classes is also commonly known as a "has-a" relationship. For example, a class "Car" could be composed of and contain a class "Engine". Therefore, a Car "has an" Engine. One aspect of composition is containment, which is the enclosure of component instances by the instance that has them. If an enclosing object contains component instances by value, the components and their enclosing object have a similar lifetime. If the components are contained by reference, they may not have a similar lifetime. For example, in Objective-C 2.0: @interface Car : NSObject @property NSString *name; @property Engine *engine @property NSArray *tires; @end This class "has" an instance of (a string object), , and (an array object). Classes can be "derived" from one or more existing classes, thereby establishing a hierarchical relationship between the derived-from classes ("base classes", "parent classes" or ') and the derived class ("child class" or "subclass") . The relationship of the derived class to the derived-from classes is commonly known as an is-a relationship. For example, a class 'Button' could be derived from a class 'Control'. Therefore, a Button is a"' Control. Structural and behavioral members of the parent classes are "inherited" by the child class. Derived classes can define additional structural members (data fields) and behavioral members (methods) in addition to those that they "inherit" and are therefore "specializations" of their superclasses. Also, derived classes can override inherited methods if the language allows. Not all languages support multiple inheritance. For example, Java allows a class to implement multiple interfaces, but only inherit from one class. If multiple inheritance is allowed, the hierarchy is a directed acyclic graph (or DAG for short), otherwise it is a tree. The hierarchy has classes as nodes and inheritance relationships as links. Classes in the same level are more likely to be associated than classes in different levels. The levels of this hierarchy are called layers or levels of abstraction. Example (Simplified Objective-C 2.0 code, from iPhone SDK): @interface UIResponder : NSObject //... @interface UIView : UIResponder //... @interface UIScrollView : UIView //... @interface UITableView : UIScrollView //... In this example, a UITableView is a UIScrollView is a UIView is a UIResponder is an NSObject. Conceptually, a superclass is a superset of its subclasses. For example, a common class hierarchy would involve as a superclass of and , while would be a subclass of . These are all subset relations in set theory as well, i.e., all squares are rectangles but not all rectangles are squares. A common conceptual error is to mistake a "part of" relation with a subclass. For example, a car and truck are both kinds of vehicles and it would be appropriate to model them as subclasses of a vehicle class. However, it would be an error to model the component parts of the car as subclass relations. For example, a car is composed of an engine and body, but it would not be appropriate to model engine or body as a subclass of car. In object-oriented modeling these kinds of relations are typically modeled as object properties. In this example, the class would have a property called . would be typed to hold a collection of objects, such as instances of , , , etc. Object modeling languages such as UML include capabilities to model various aspects of "part of" and other kinds of relations – data such as the cardinality of the objects, constraints on input and output values, etc. This information can be utilized by developer tools to generate additional code beside the basic data definitions for the objects, such as error checking on get and set methods. One important question when modeling and implementing a system of object classes is whether a class can have one or more superclasses. In the real world with actual sets it would be rare to find sets that didn't intersect with more than one other set. However, while some systems such as Flavors and CLOS provide a capability for more than one parent to do so at run time introduces complexity that many in the object-oriented community consider antithetical to the goals of using object classes in the first place. Understanding which class will be responsible for handling a message can get complex when dealing with more than one superclass. If used carelessly this feature can introduce some of the same system complexity and ambiguity classes were designed to avoid. Most modern object-oriented languages such as Smalltalk and Java require single inheritance at run time. For these languages, multiple inheritance may be useful for modeling but not for an implementation. However, semantic web application objects do have multiple superclasses. The volatility of the Internet requires this level of flexibility and the technology standards such as the Web Ontology Language (OWL) are designed to support it. A similar issue is whether or not the class hierarchy can be modified at run time. Languages such as Flavors, CLOS, and Smalltalk all support this feature as part of their meta-object protocols. Since classes are themselves first-class objects, it is possible to have them dynamically alter their structure by sending them the appropriate messages. Other languages that focus more on strong typing such as Java and C++ do not allow the class hierarchy to be modified at run time. Semantic web objects have the capability for run time changes to classes. The rational is similar to the justification for allowing multiple superclasses, that the Internet is so dynamic and flexible that dynamic changes to the hierarchy are required to manage this volatility. Although class-based languages are commonly assumed to support inheritance, inheritance is not an intrinsic aspect of the concept of classes. Some languages, often referred to as "object-based languages", support classes yet do not support inheritance. Examples of object-based languages include earlier versions of Visual Basic. In object-oriented analysis and in UML, an association between two classes represents a collaboration between the classes or their corresponding instances. Associations have direction; for example, a bi-directional association between two classes indicates that both of the classes are aware of their relationship. Associations may be labeled according to their name or purpose. An association role is given end of an association and describes the role of the corresponding class. For example, a "subscriber" role describes the way instances of the class "Person" participate in a "subscribes-to" association with the class "Magazine". Also, a "Magazine" has the "subscribed magazine" role in the same association. Association role multiplicity describes how many instances correspond to each instance of the other class of the association. Common multiplicities are "0..1", "1..1", "1..*" and "0..*", where the "*" specifies any number of instances. There are many categories of classes, some of which overlap. In a language that supports inheritance, an abstract class, or "abstract base class" (ABC), is a class that cannot be instantiated because it is either labeled as abstract or it simply specifies abstract methods (or "virtual methods"). An abstract class may provide implementations of some methods, and may also specify virtual methods via signatures that are to be implemented by direct or indirect descendants of the abstract class. Before a class derived from an abstract class can be instantiated, all abstract methods of its parent classes must be implemented by some class in the derivation chain. Most object-oriented programming languages allow the programmer to specify which classes are considered abstract and will not allow these to be instantiated. For example, in Java, C# and PHP, the keyword "abstract" is used. In C++, an abstract class is a class having at least one abstract method given by the appropriate syntax in that language (a pure virtual function in C++ parlance). A class consisting of only virtual methods is called a Pure Abstract Base Class (or "Pure ABC") in C++ and is also known as an "interface" by users of the language. Other languages, notably Java and C#, support a variant of abstract classes called an interface via a keyword in the language. In these languages, multiple inheritance is not allowed, but a class can implement multiple interfaces. Such a class can only contain abstract publicly accessible methods. A concrete class is a class that can be instantiated, as opposed to abstract classes, which cannot. In some languages, classes can be declared in scopes other than the global scope. There are various types of such classes. An inner class is a class defined within another class. The relationship between an inner class and its containing class can also be treated as another type of class association. An inner class is typically neither associated with instances of the enclosing class nor instantiated along with its enclosing class. Depending on language, it may or may not be possible to refer to the class from outside the enclosing class. A related concept is "inner types", also known as "inner data type" or "nested type", which is a generalization of the concept of inner classes. C++ is an example of a language that supports both inner classes and inner types (via "typedef" declarations). Another type is a local class, which is a class defined within a procedure or function. This limits references to the class name to within the scope where the class is declared. Depending on the semantic rules of the language, there may be additional restrictions on local classes compared to non-local ones. One common restriction is to disallow local class methods to access local variables of the enclosing function. For example, in C++, a local class may refer to static variables declared within its enclosing function, but may not access the function's automatic variables. Metaclasses are classes whose instances are classes. A metaclass describes a common structure of a collection of classes and can implement a design pattern or describe particular kinds of classes. Metaclasses are often used to describe frameworks. In some languages, such as Python, Ruby or Smalltalk, a class is also an object; thus each class is an instance of a unique metaclass that is built into the language. The Common Lisp Object System (CLOS) provides metaobject protocols (MOPs) to implement those classes and metaclasses. Non-subclassable classes allow programmers to design classes and hierarchies of classes where at some level in the hierarchy, further derivation is prohibited (a stand-alone class may be also designated as non-subclassable, preventing the formation of any hierarchy). Contrast this to "abstract" classes, which imply, encourage, and require derivation in order to be used at all. A non-subclassable class is implicitly "concrete". A non-subclassable class is created by declaring the class as in C# or as in Java or PHP. For example, Java's class is designated as "final". Non-subclassable classes may allow a compiler (in compiled languages) to perform optimizations that are not available for subclassable classes. An open class is one that can be changed. Typically, an executable program cannot be changed by customers. Developers can often change some classes, but typically cannot change standard or built-in ones. In Ruby, all classes are open. In Python, classes can be created at runtime, and all can be modified afterwards. Objective-C categories permit the programmer to add methods to an existing class without the need to recompile that class or even have access to its source code. Some languages have special support for mixins, though in any language with multiple inheritance a mixin is simply a class that does not represent an is-a-type-of relationship. Mixins are typically used to add the same methods to multiple classes; for example, a class might provide a method called when included in classes and that do not share a common parent. In languages supporting the feature, a partial class is a class whose definition may be split into multiple pieces, within a single source-code file or across multiple files. The pieces are merged at compile-time, making compiler output the same as for a non-partial class. The primary motivation for introduction of partial classes is to facilitate the implementation of code generators, such as visual designers. It is otherwise a challenge or compromise to develop code generators that can manage the generated code when it is interleaved within developer-written code. Using partial classes, a code generator can process a separate file or coarse-grained partial class within a file, and is thus alleviated from intricately interjecting generated code via extensive parsing, increasing compiler efficiency and eliminating the potential risk of corrupting developer code. In a simple implementation of partial classes, the compiler can perform a phase of precompilation where it "unifies" all the parts of a partial class. Then, compilation can proceed as usual. Other benefits and effects of the partial class feature include: Partial classes have existed in Smalltalk under the name of "Class Extensions" for considerable time. With the arrival of the .NET framework 2, Microsoft introduced partial classes, supported in both C# 2.0 and Visual Basic 2005. WinRT also supports partial classes. This simple example, written in Visual Basic .NET, shows how parts of the same class are defined in two different files. Partial Class MyClass End Class Partial Class MyClass End Class When compiled, the result is the same as if the two files were written as one, like this: Class MyClass End Class In Objective-C, partial classes, also known as categories, may even spread over multiple libraries and executables, like the following example. But a key difference is that Objective-C's categories can overwrite definitions in another interface declaration, and that categories aren't equal to original class definition (the first requires the last). Instead, .NET partial class can't have conflicting definitions, and all partial definitions are equal to the others. In Foundation, header file NSData.h: @interface NSData : NSObject - (id)initWithContentsOfURL:(NSURL *)URL; @end In user-supplied library, a separate binary from Foundation framework, header file NSData+base64.h: @interface NSData (base64) - (NSString *)base64String; - (id)initWithBase64String:(NSString *)base64String; @end And in an app, yet another separate binary file, source code file main.m: int main(int argc, char *argv[]) The dispatcher will find both methods called over the NSData instance and invoke both of them correctly. Uninstantiable classes allow programmers to group together per-class fields and methods that are accessible at runtime without an instance of the class. Indeed, instantiation is prohibited for this kind of class. For example, in C#, a class marked "static" can not be instantiated, can only have static members (fields, methods, other), may not have "instance constructors", and is "sealed". An unnamed class or anonymous class is a class that is not bound to a name or identifier upon definition. This is analogous to named versus unnamed functions. The benefits of organizing software into object classes fall into three categories: Object classes facilitate rapid development because they lessen the semantic gap between the code and the users. System analysts can talk to both developers and users using essentially the same vocabulary, talking about accounts, customers, bills, etc. Object classes often facilitate rapid development because most object-oriented environments come with powerful debugging and testing tools. Instances of classes can be inspected at run time to verify that the system is performing as expected. Also, rather than get dumps of core memory, most object-oriented environments have interpreted debugging capabilities so that the developer can analyze exactly where in the program the error occurred and can see which methods were called to which arguments and with what arguments. Object classes facilitate ease of maintenance via encapsulation. When developers need to change the behavior of an object they can localize the change to just that object and its component parts. This reduces the potential for unwanted side effects from maintenance enhancements. Software re-use is also a major benefit of using Object classes. Classes facilitate re-use via inheritance and interfaces. When a new behavior is required it can often be achieved by creating a new class and having that class inherit the default behaviors and data of its superclass and then tailor some aspect of the behavior or data accordingly. Re-use via interfaces (also known as methods) occurs when another object wants to invoke (rather than create a new kind of) some object class. This method for re-use removes many of the common errors that can make their way into software when one program re-uses code from another. As a data type, a class is usually considered as a compile-time construct. A language or library may also support prototype or factory metaobjects that represent run-time information about classes, or even represent metadata that provides access to reflection facilities and ability to manipulate data structure formats at run-time. Many languages distinguish this kind of run-time type information about classes from a class on the basis that the information is not needed at run-time. Some dynamic languages do not make strict distinctions between run-time and compile-time constructs, and therefore may not distinguish between metaobjects and classes. For example, if Human is a metaobject representing the class Person, then instances of class Person can be created by using the facilities of the Human metaobject.
https://en.wikipedia.org/wiki?curid=7392
Color blindness Color blindness, also known as color vision deficiency, is the decreased ability to see color or differences in color. Simple tasks such as selecting ripe fruit, choosing clothing, and reading traffic lights can be more challenging. Color blindness may also make some educational activities more difficult. However, problems are generally minor, and most people find that they can adapt. People with total color blindness (achromatopsia) may also have decreased visual acuity and be uncomfortable in bright environments. The most common cause of color blindness is an inherited problem in the development of one or more of the three sets of color-sensing cones in the eye. Males are more likely to be color blind than females, as the genes responsible for the most common forms of color blindness are on the X chromosome. As females have two X chromosomes, a defect in one is typically compensated for by the other, therefore females can be carriers. Males only have one X chromosome and therefore express the genetic disorder. Color blindness can also result from physical or chemical damage to the eye, optic nerve or parts of the brain. Diagnosis is typically with the Ishihara color test; however, a number of other testing methods, including genetic testing, also exist. There is no cure for color blindness. Diagnosis may allow a person's teacher to change their method of teaching to accommodate the decreased ability to recognize colors. Special lenses may help people with red–green color blindness when under bright conditions. There are also mobile apps that can help people identify colors. Red–green color blindness is the most common form, followed by blue–yellow color blindness and total color blindness. Red–green color blindness affects up to 8% of males and 0.5% of females of Northern European descent. The ability to see color also decreases in old age. Being color blind may make people ineligible for certain jobs in certain countries. This may include being a pilot, train driver, crane operator, and working in the armed forces. The effect of color blindness on artistic ability, however, is controversial. The ability to draw appears to be unchanged, and a number of famous artists are believed to have been color blind. In almost all cases, color blind people retain blue–yellow discrimination, and most color blind individuals are anomalous trichromats rather than complete dichromats. In practice, this means that they often retain a limited discrimination along the red–green axis of color space, although their ability to separate colors in this dimension is reduced. Color blindness very rarely refers to complete monochromatism. Dichromats often confuse red and green items. For example, they may find it difficult to distinguish a Braeburn apple from a Granny Smith or red from green of traffic lights without other clues—for example, shape or position. Dichromats tend to learn to use texture and shape clues and so may be able to penetrate camouflage that has been designed to deceive individuals with normal color vision. Colors of traffic lights are confusing to some dichromats as there is insufficient apparent difference between the red/amber traffic lights and sodium street lamps; also, the green can be confused with a grubby white lamp. This is a risk on high-speed undulating roads where angular cues cannot be used. British Rail color lamp signals use more easily identifiable colors: The red is blood red, the amber is yellow and the green is a bluish color. Most British road traffic lights are mounted vertically on a black rectangle with a white border (forming a "sighting board") and so dichromats can more easily look for the position of the light within the rectangle—top, middle or bottom. In the eastern provinces of Canada horizontally mounted traffic lights are generally differentiated by shape to facilitate identification for those with color blindness. In the United States, this is not done by shape but by position, as the red light is always on the left if the light is horizontal, or on top if the light is vertical. However, a lone flashing light (e.g. red for stop, yellow for caution) is still problematic. Color vision deficiencies can be classified as acquired or inherited. Color blindness is typically an inherited genetic disorder. It is most commonly inherited from mutations on the X chromosome, but the mapping of the human genome has shown there are many causative mutations—mutations capable of causing color blindness originate from at least 19 different chromosomes and 56 different genes (as shown online at the Online Mendelian Inheritance in Man (OMIM)). Two of the most common inherited forms of color blindness are protanomaly (and, more rarely, protanopia—the two together often known as "protans") and deuteranomaly (or, more rarely, deuteranopia—the two together often referred to as "deutans"). Both "protans" and "deutans" (of which the deutans are by far the most common) are known as "red–green color-blind". They comprise about 8% of human males and 0.6% of females of Northern European ancestry. Some of the inherited diseases known to cause color blindness are: Inherited color blindness can be congenital (from birth), or it can commence in childhood or adulthood. Depending on the mutation, it can be stationary, that is, remain the same throughout a person's lifetime, or progressive. As progressive phenotypes involve deterioration of the retina and other parts of the eye, certain forms of color blindness can progress to legal blindness, i.e. an acuity of 6/60 (20/200) or worse, and often leave a person with complete blindness. Color blindness always pertains to the cone photoreceptors in retinas, as it is the cones that detect the color frequencies of light. About 8% of males, and 0.4% of females, are red–green color blind in some way or another, whether it is one color, a color combination, or another mutation. Males are at a greater risk of inheriting an X-linked mutation because males only have one X chromosome (XY, with the Y chromosome carrying altogether different genes from the X chromosome), and females have two (XX); if a woman inherits a normal X chromosome in addition to the one that carries the mutation, she will not display the mutation. Men do not have a second X chromosome to override the chromosome that carries the mutation. If 8% of variants of a given gene are defective, the probability of a single copy being defective is 8%, but the probability that two copies are both defective is 0.082, i.e. 0.64%. Other causes of color blindness include brain or retinal damage caused by accidents and other traumas which produce swelling of the brain in the occipital lobe, and damage to the retina caused by exposure to ultraviolet light (wavelengths 10 to 300 nm). Damage often presents itself later in life. Color blindness may also present itself in the range of degenerative diseases of the eye, such as age-related macular degeneration, and as part of the retinal damage caused by diabetes. Vitamin A deficiency may also cause color blindness. Some subtle forms of color blindness may be associated with chronic solvent-induced encephalopathy (CSE), caused by long-time exposure to solvent vapors. Red–green color blindness can be caused by ethambutol, a drug used in the treatment of tuberculosis. The typical human retina contains two kinds of light cells: the rod cells (active in low light) and the cone cells (active in normal daylight). Normally, there are three kinds of cone cells, each containing a different pigment, which are activated when the pigments absorb light. The spectral sensitivities of the cones differ; one is most sensitive to short wavelengths, one to medium wavelengths, and the third to medium-to-long wavelengths within the visible spectrum, with their peak sensitivities in the blue, green, and yellow–green regions of the spectrum, respectively. The absorption spectra of the three systems overlap, and combine to cover the visible spectrum. These receptors are known as short (S), medium (M), and long (L) wavelength cones, but are also often referred to as blue, green, and red cones, although this terminology is inaccurate. The receptors are each responsive to a wide range of wavelengths. For example, the long wavelength "red" receptor has its peak sensitivity in the yellow–green, some way from the red end (longest wavelength) of the visible spectrum. The sensitivity of normal color vision actually depends on the overlap between the absorption ranges of the three systems: different colors are recognized when the different types of cone are stimulated to different degrees. Red light, for example, stimulates the long wavelength cones much more than either of the others, and reducing the wavelength causes the other two cone systems to be increasingly stimulated, causing a gradual change in hue. Many of the genes involved in color vision are on the X chromosome, making color blindness much more common in males than in females because males only have one X chromosome, while females have two. An estimated 2–3% of women have two slightly different red colour cones and can be considered tetrachromats. One such woman has been reported to be a true or functional tetrachromat, as she can discriminate colors most other people can't. The Ishihara color test, which consists of a series of pictures of colored spots, is the test most often used to diagnose red–green color deficiencies. A figure (usually one or more Arabic digits) is embedded in the picture as a number of spots in a slightly different color, and can be seen with normal color vision, but not with a particular color defect. The full set of tests has a variety of figure/background color combinations, and enable diagnosis of which particular visual defect is present. The anomaloscope, described above, is also used in diagnosing anomalous trichromacy. Because the Ishihara color test contains only numerals, it may not be useful in diagnosing young children, who have not yet learned to use numbers. In the interest of identifying these problems early on in life, alternative color vision tests were developed using only symbols (square, circle, car). Besides the Ishihara color test, the US Navy and US Army also allow testing with the Farnsworth Lantern Test. This test allows 30% of color deficient individuals, whose deficiency is not too severe, to pass. Another test used by clinicians to measure chromatic discrimination is the Farnsworth–Munsell 100 hue test. The patient is asked to arrange a set of colored caps or chips to form a gradual transition of color between two anchor caps. The HRR color test (developed by Hardy, Rand, and Rittler) is a red–green color test that, unlike the Ishihara, also has plates for the detection of the tritan defects. Most clinical tests are designed to be fast, simple, and effective at identifying broad categories of color blindness. In academic studies of color blindness, on the other hand, there is more interest in developing flexible tests to collect thorough datasets, identify copunctal points, and measure just noticeable differences. Types of color blindness and the terms used Based on clinical appearance, color blindness may be described as total or partial. Total color blindness is much less common than partial color blindness. There are two major types of color blindness: difficulty distinguishing between red and green, and difficulty distinguishing between blue and yellow. Immunofluorescent imaging is a way to determine red–green color coding. Conventional color coding is difficult for individuals with red–green color blindness (protanopia or deuteranopia) to discriminate. Replacing red with magenta or green with turquoise improves visibility for such individuals. The different kinds of inherited color blindness result from partial or complete loss of function of one or more of the three different cone systems. When one cone system is compromised, dichromacy results. The most frequent forms of human color blindness result from problems with either the middle (green) or long (red) wavelength sensitive cone systems, and make it hard to discriminate reds, yellows, and greens from one another. They are collectively referred to as "red–green color blindness", though the term is an over-simplification and is somewhat misleading. Other forms of color blindness are much more rare. They include problems in discriminating blues from greens and yellows from reds/pinks, and the rarest form of all, complete color blindness or "monochromacy", where one cannot distinguish any color from grey, as in a black-and-white movie or photograph. Protanopes, deuteranopes, and tritanopes are dichromats; that is, they can match any color they see with some mixture of just two primary colors (in contrast to those with normal sight (trichromats) who can distinguish three primary colors). Dichromats usually know they have a color vision problem, and it can affect their daily lives. Out of the male population, 2% have severe difficulties distinguishing between red, orange, yellow, and green. (Orange and yellow are different combinations of red and green light.) Colors in this range, which appear very different to a normal viewer, appear to a dichromat to be the same or a similar color. The terms protanopia, deuteranopia, and tritanopia come from Greek, and respectively mean "inability to see ("anopia") with the first ("prot-"), second ("deuter-"), or third ("trit-") [cone]". Anomalous trichromacy is the least serious type of color deficiency. People with protanomaly, deuteranomaly, or tritanomaly are trichromats, but the color matches they make differ from the normal. They are called anomalous trichromats. In order to match a given spectral yellow light, protanomalous observers need more red light in a red/green mixture than a normal observer, and deuteranomalous observers need more green. From a practical standpoint though, many protanomalous and deuteranomalous people have very little difficulty carrying out tasks that require normal color vision. Some may not even be aware that their color perception is in any way different from normal. Protanomaly and deuteranomaly can be diagnosed using an instrument called an anomaloscope, which mixes spectral red and green lights in variable proportions, for comparison with a fixed spectral yellow. If this is done in front of a large audience of males, as the proportion of red is increased from a low value, first a small proportion of the audience will declare a match, while most will see the mixed light as greenish; these are the deuteranomalous observers. Next, as more red is added the majority will say that a match has been achieved. Finally, as yet more red is added, the remaining, protanomalous, observers will declare a match at a point where normal observers will see the mixed light as definitely reddish. Protanopia, deuteranopia, protanomaly, and deuteranomaly are commonly inherited forms of red–green color blindness which affect a substantial portion of the human population. Those affected have difficulty with discriminating red and green hues due to the absence or mutation of the red or green retinal photoreceptors. It is sex-linked: genetic red–green color blindness affects males much more often than females, because the genes for the red and green color receptors are located on the X chromosome, of which males have only one and females have two. Females (XX) are red–green color blind only if "both" their X chromosomes are defective with a similar deficiency, whereas males (XY) are color blind if their single X chromosome is defective. The gene for red–green color blindness is transmitted from a color blind male to all his daughters, who are usually heterozygote carriers and are thus unaffected. In turn, a carrier woman has a 50% chance of passing on a mutated X chromosome region to each of her male offspring. The sons of an affected male will not inherit the trait from him, since they receive his Y chromosome and not his (defective) X chromosome. Should an affected male have children with a carrier or colorblind woman, their daughters may be colorblind by inheriting an affected X chromosome from each parent. Because one X chromosome is inactivated at random in each cell during a woman's development, deuteranomalous heterozygotes (i.e. female carriers of deuteranomaly) may be tetrachromats, because they will have the normal long wave (red) receptors, the normal medium wave (green) receptors, the abnormal medium wave (deuteranomalous) receptors and the normal autosomal short wave (blue) receptors in their retinas. The same applies to the carriers of protanomaly (who have two types of long wave receptors, normal medium wave receptors, and normal autosomal short wave receptors in their retinas). If, by rare chance, a woman is heterozygous for "both" protanomaly and deuteranomaly, she could be pentachromatic. This situation could arise if, for instance, she inherited the X chromosome with the abnormal long wave gene (but normal medium wave gene) from her mother who is a carrier of protanomaly, and her other X chromosome from a deuteranomalous father. Such a woman would have a normal and an abnormal long wave receptor, a normal and abnormal medium wave receptor, and a normal autosomal short wave receptor—5 different types of color receptors in all. The degree to which women who are carriers of either protanomaly or deuteranomaly are demonstrably tetrachromatic and require a mixture of four spectral lights to match an arbitrary light is very variable. In many cases it is almost unnoticeable, but in a minority the tetrachromacy is very pronounced. However, Jameson "et al." have shown that with appropriate and sufficiently sensitive equipment it can be demonstrated that any female carrier of red–green color blindness (i.e. heterozygous protanomaly, or heterozygous deuteranomaly) is a tetrachromat to a greater or lesser extent. People in whom deuteranopia or deuteranomaly manifest are sometimes referred to as deutans, those with protanopia or protanomaly as protans. Since deuteranomaly is by far the most common form of red–green blindness among men of northwestern European descent (with an incidence of 8%) it follows that the proportion of carriers (and of potential deuteranomalous tetrachromats) among the females of that genetic stock is 14.7% (i.e. 92% × 8% × 2), based on the Hardy–Weinberg principle. Those with tritanopia and tritanomaly have difficulty discerning between bluish and greenish hues, as well as yellowish and reddish hues. Color blindness involving the inactivation of the short-wavelength sensitive cone system (whose absorption spectrum peaks in the bluish-violet) is called tritanopia or, loosely, blue–yellow color blindness. The tritanope's neutral point occurs near a yellowish 570 nm; green is perceived at shorter wavelengths and red at longer wavelengths. Mutation of the short-wavelength sensitive cones is called tritanomaly. Tritanopia is equally distributed among males and females. Jeremy H. Nathans (with the Howard Hughes Medical Institute) demonstrated that the gene coding for the blue receptor lies on chromosome 7, which is shared equally by males and females. Therefore, it is not sex-linked. This gene does not have any neighbor whose DNA sequence is similar. Blue color blindness is caused by a simple mutation in this gene. Total color blindness is defined as the inability to see color. Although the term may refer to acquired disorders such as cerebral achromatopsia also known as color agnosia, it typically refers to congenital color vision disorders (i.e. more frequently rod monochromacy and less frequently cone monochromacy). In cerebral achromatopsia, a person cannot perceive colors even though the eyes are capable of distinguishing them. Some sources do not consider these to be true color blindness, because the failure is of perception, not of vision. They are forms of visual agnosia. Monochromacy is the condition of possessing only a single channel for conveying information about color. Monochromats possess a complete inability to distinguish any colors and perceive only variations in brightness. It occurs in two primary forms: There is no cure for color deficiencies. ″The American Optometric Association reports a contact lens on one eye can increase the ability to differentiate between colors, though nothing can make you truly see the deficient color.″ Optometrists can supply colored spectacle lenses or a single red-tint contact lens to wear on the non-dominant eye, but although this may improve discrimination of some colors, it can make other colors more difficult to distinguish. A 1981 review of various studies to evaluate the effect of the X-chrom contact lens concluded that, while the lens may allow the wearer to achieve a better score on certain color vision tests, it did not correct color vision in the natural environment. A case history using the X-Chrom lens for a rod monochromat is reported and an X-Chrom manual is online. Lenses that filter certain wavelengths of light can allow people with a cone anomaly, but not dichromacy, to see better separation of colors, especially those with classic "red/green" color blindness. They work by notching out wavelengths that strongly stimulate both red and green cones in a deuter- or protanomalous person, improving the distinction between the two cones' signals. As of 2012, sunglasses that notch out color wavelengths are available commercially. Many mobile and computer applications have been developed to help color blind individual to view better differentiate between colors. Some applications launch a simulation of colorblindness to allow people with typical vision to understand how people with color blindness see the world, which can improve inclusive design for both groups. Color blindness affects a large number of individuals, with protans and deutans being the most common types. In individuals with Northern European ancestry, as many as 8 percent of men and 0.4 percent of women experience congenital color deficiency. The number affected varies among groups. Isolated communities with a restricted gene pool sometimes produce high proportions of color blindness, including the less usual types. Examples include rural Finland, Hungary, and some of the Scottish islands. In the United States, about 7 percent of the male population—or about 10.5 million men—and 0.4 percent of the female population either cannot distinguish red from green, or see red and green differently from how others do (Howard Hughes Medical Institute, 2006). More than 95 percent of all variations in human color vision involve the red and green receptors in male eyes. It is very rare for males or females to be "blind" to the blue end of the spectrum. The first scientific paper on the subject of color blindness, "Extraordinary facts relating to the vision of colours", was published by the English chemist John Dalton in 1798 after the realization of his own color blindness. Because of Dalton's work, the general condition has been called "daltonism", although in English this term is now used only for deuteranopia. Color codes present particular problems for those with color deficiencies as they are often difficult or impossible for them to perceive. Good graphic design avoids using color coding or using color contrasts alone to express information; this not only helps color blind people, but also aids understanding by normally sighted people by providing them with multiple reinforcing cues. Designers need to take into account that color-blindness is highly sensitive to differences in material. For example, a red–green colorblind person who is incapable of distinguishing colors on a map printed on paper may have no such difficulty when viewing the map on a computer screen or television. In addition, some color blind people find it easier to distinguish problem colors on artificial materials, such as plastic or in acrylic paints, than on natural materials, such as paper or wood. Third, for some color blind people, color can only be distinguished if there is a sufficient "mass" of color: thin lines might appear black, while a thicker line of the same color can be perceived as having color. Designers should also note that red–blue and yellow–blue color combinations are generally safe. So instead of the ever-popular "red means bad and green means good" system, using these combinations can lead to a much higher ability to use color coding effectively. This will still cause problems for those with monochromatic color blindness, but it is still something worth considering. When the need to process visual information as rapidly as possible arises, for example in an emergency situation, the visual system may operate only in shades of gray, with the extra information load in adding color being dropped. This is an important possibility to consider when designing, for example, emergency brake handles or emergency phones. Color blindness may make it difficult or impossible for a person to engage in certain occupations. Persons with color blindness may be legally or practically barred from occupations in which color perception is an essential part of the job ("e.g.," mixing paint colors), or in which color perception is important for safety ("e.g.," operating vehicles in response to color-coded signals). This occupational safety principle originates from the Lagerlunda train crash of 1875 in Sweden. Following the crash, Professor Alarik Frithiof Holmgren, a physiologist, investigated and concluded that the color blindness of the engineer (who had died) had caused the crash. Professor Holmgren then created the first test using different-colored skeins to exclude people from jobs in the transportation industry on the basis of color blindness. However, there is a claim that there is no firm evidence that color deficiency did cause the collision, or that it might have not been the sole cause. Color vision is important for occupations using telephone or computer networking cabling, as the individual wires inside the cables are color-coded using green, orange, brown, blue and white colors. Electronic wiring, transformers, resistors, and capacitors are color-coded as well, using black, brown, red, orange, yellow, green, blue, violet, gray, white, silver, gold. Some countries have refused to grant driving licenses to individuals with color blindness. In Romania, there is an ongoing campaign to remove the legal restrictions that prohibit colorblind citizens from getting drivers' licenses. The usual justification for such restrictions is that drivers of motor vehicles must be able to recognize color-coded signals, such as traffic lights or warning lights. While many aspects of aviation depend on color coding, only a few of them are critical enough to be interfered with by some milder types of color blindness. Some examples include color-gun signaling of aircraft that have lost radio communication, color-coded glide-path indications on runways, and the like. Some jurisdictions restrict the issuance of pilot credentials to persons who suffer from color blindness for this reason. Restrictions may be partial, allowing color-blind persons to obtain certification but with restrictions, or total, in which case color-blind persons are not permitted to obtain piloting credentials at all. In the United States, the Federal Aviation Administration requires that pilots be tested for normal color vision as part of their medical clearance in order to obtain the required medical certificate, a prerequisite to obtaining a pilot's certification. If testing reveals color blindness, the applicant may be issued a license with restrictions, such as no night flying and no flying by color signals—such a restriction effectively prevents a pilot from holding certain flying occupations, such as that of an airline pilot, although commercial pilot certification is still possible, and there are a few flying occupations that do not require night flight and thus are still available to those with restrictions due to color blindness (e.g., agricultural aviation). The government allows several types of tests, including medical standard tests ("e.g.," the Ishihara, Dvorine, and others) and specialized tests oriented specifically to the needs of aviation. If an applicant fails the standard tests, they will receive a restriction on their medical certificate that states: "Not valid for night flying or by color signal control". They may apply to the FAA to take a specialized test, administered by the FAA. Typically, this test is the "color vision light gun test". For this test an FAA inspector will meet the pilot at an airport with an operating control tower. The color signal light gun will be shone at the pilot from the tower, and they must identify the color. If they pass they may be issued a waiver, which states that the color vision test is no longer required during medical examinations. They will then receive a new medical certificate with the restriction removed. This was once a Statement of Demonstrated Ability (SODA), but the SODA was dropped, and converted to a simple waiver (letter) early in the 2000s. Research published in 2009 carried out by the City University of London's Applied Vision Research Centre, sponsored by the UK's Civil Aviation Authority and the U.S. Federal Aviation Administration, has established a more accurate assessment of color deficiencies in pilot applicants' red/green and yellow–blue color range which could lead to a 35% reduction in the number of prospective pilots who fail to meet the minimum medical threshold. Inability to distinguish color does not necessarily preclude the ability to become a celebrated artist. The 20th century expressionist painter Clifton Pugh, three-time winner of Australia's Archibald Prize, on biographical, gene inheritance and other grounds has been identified as a protanope. 19th century French artist Charles Méryon became successful by concentrating on etching rather than painting after he was diagnosed as having a red–green deficiency. Jin Kim's red–green color blindness did not stop him from becoming first an animator and later a character designer with Walt Disney Animation Studios. A Brazilian court ruled that people with color blindness are protected by the Inter-American Convention on the Elimination of All Forms of Discrimination against Person with Disabilities. At trial, it was decided that the carriers of color blindness have a right of access to wider knowledge, or the full enjoyment of their human condition. In the United States, under federal anti-discrimination laws such as the Americans with Disabilities Act, color vision deficiencies have not been found to constitute a disability that triggers protection from workplace discrimination. A famous traffic light on Tipperary Hill in Syracuse, New York, is upside-down due to the sentiments of its Irish American community, but has been criticized due to the potential hazard it poses for color-blind persons. Some tentative evidence finds that color blind people are better at penetrating certain color camouflages. Such findings may give an evolutionary reason for the high rate of red–green color blindness. There is also a study suggesting that people with some types of color blindness can distinguish colors that people with normal color vision are not able to distinguish. In World War II, color blind observers were used to penetrate camouflage. In September 2009, the journal "Nature" reported that researchers at the University of Washington and University of Florida were able to give trichromatic vision to squirrel monkeys, which normally have only dichromatic vision, using gene therapy. In 2003, a cybernetic device called eyeborg was developed to allow the wearer to hear sounds representing different colors. Achromatopsic artist Neil Harbisson was the first to use such a device in early 2004; the eyeborg allowed him to start painting in color by memorizing the sound corresponding to each color. In 2012, at a TED Conference, Harbisson explained how he could now perceive colors outside the ability of human vision.
https://en.wikipedia.org/wiki?curid=7397
Computer security Computer security, cybersecurity or information technology security (IT security) is the protection of computer systems and networks from the theft of or damage to their hardware, software, or electronic data, as well as from the disruption or misdirection of the services they provide. The field is becoming more important due to increased reliance on computer systems, the Internet and wireless network standards such as Bluetooth and Wi-Fi, and due to the growth of "smart" devices, including smartphones, televisions, and the various devices that constitute the "Internet of things". Owing to its complexity, both in terms of politics and technology, cybersecurity is also one of the major challenges in the contemporary world. A vulnerability is a weakness in design, implementation, operation or internal control. Most of the vulnerabilities that have been discovered are documented in the Common Vulnerabilities and Exposures (CVE) database. An "exploitable" vulnerability is one for which at least one working attack or "exploit" exists. Vulnerabilities are often hunted or exploited with the aid of automated tools or manually using customized scripts. To secure a computer system, it is important to understand the attacks that can be made against it, and these threats can typically be classified into one of these categories below: A backdoor in a computer system, a cryptosystem or an algorithm, is any secret method of bypassing normal authentication or security controls. They may exist for a number of reasons, including by original design or from poor configuration. They may have been added by an authorized party to allow some legitimate access, or by an attacker for malicious reasons; but regardless of the motives for their existence, they create a vulnerability. Backdoors can be very hard to detect, and detection of backdoors are usually discovered by someone who has access to application source code or intimate knowledge of the computer's Operating System. Denial of service attacks (DoS) are designed to make a machine or network resource unavailable to its intended users. Attackers can deny service to individual victims, such as by deliberately entering a wrong password enough consecutive times to cause the victims account to be locked, or they may overload the capabilities of a machine or network and block all users at once. While a network attack from a single IP address can be blocked by adding a new firewall rule, many forms of Distributed denial of service (DDoS) attacks are possible, where the attack comes from a large number of points – and defending is much more difficult. Such attacks can originate from the zombie computers of a botnet, or from a range of other possible techniques, including reflection and amplification attacks, where innocent systems are fooled into sending traffic to the victim. An unauthorized user gaining physical access to a computer is most likely able to directly copy data from it. They may also compromise security by making operating system modifications, installing software worms, keyloggers, covert listening devices or using wireless mice. Even when the system is protected by standard security measures, these may be able to be by-passed by booting another operating system or tool from a CD-ROM or other bootable media. Disk encryption and Trusted Platform Module are designed to prevent these attacks. Eavesdropping is the act of surreptitiously listening to a private computer "conversation" (communication), typically between hosts on a network. For instance, programs such as Carnivore and NarusInSight have been used by the FBI and NSA to eavesdrop on the systems of internet service providers. Even machines that operate as a closed system (i.e., with no contact to the outside world) can be eavesdropped upon via monitoring the faint electromagnetic transmissions generated by the hardware; TEMPEST is a specification by the NSA referring to these attacks. Surfacing in 2017, a new class of multi-vector, polymorphic cyber threats surfaced that combined several types of attacks and changed form to avoid cybersecurity controls as they spread. These threats have been classified as fifth-generation cyberattacks. Phishing is the attempt to acquire sensitive information such as usernames, passwords, and credit card details directly from users by deceiving the users. Phishing is typically carried out by email spoofing or instant messaging, and it often directs users to enter details at a fake website whose "look" and "feel" are almost identical to the legitimate one. The fake website often asks for personal information, such as log-in details and passwords. This information can then be used to gain access to the individual's real account on the real website. Preying on a victim's trust, phishing can be classified as a form of social engineering. Attackers are using creative ways to gain access to real accounts. A common scam is for attackers to send fake electronic invoices to individuals showing that they recently purchased music, apps, or other, and instructing them to click on a link if the purchases were not authorized. Privilege escalation describes a situation where an attacker with some level of restricted access is able to, without authorization, elevate their privileges or access level. For example, a standard computer user may be able to exploit a vulnerability in the system to gain access to restricted data; or even become "root" and have full unrestricted access to a system. Social engineering, insofar as computer security is concerned, aims to convince a user to disclose secrets such as passwords, card numbers, etc. by, for example, impersonating a bank, a contractor, or a customer. Social engineering, in the context of information security, is the psychological manipulation of people into performing actions or divulging confidential information. A common scam involves fake CEO emails sent to accounting and finance departments. In early 2016, the FBI reported that the scam has cost US businesses more than $2billion in about two years. In May 2016, the Milwaukee Bucks NBA team was the victim of this type of cyber scam with a perpetrator impersonating the team's president Peter Feigin, resulting in the handover of all the team's employees' 2015 W-2 tax forms. Spoofing is the act of masquerading as a valid entity through falsification of data (such as an IP address or username), in order to gain access to information or resources that one is otherwise unauthorized to obtain. There are several types of spoofing, including: Tampering describes a malicious modification or alteration of data. So-called Evil Maid attacks and security services planting of surveillance capability into routers are examples. Employee behavior can have a big impact on information security in organizations. Cultural concepts can help different segments of the organization work effectively or work against effectiveness towards information security within an organization. Information security culture is the "...totality of patterns of behavior in an organization that contribute to the protection of information of all kinds.″ Andersson and Reimers (2014) found that employees often do not see themselves as part of their organization's information security effort and often take actions that impede organizational changes. Research shows information security culture needs to be improved continuously. In ″Information Security Culture from Analysis to Change″, authors commented, ″It's a never ending process, a cycle of evaluation and change or maintenance.″ To manage the information security culture, five steps should be taken: pre-evaluation, strategic planning, operative planning, implementation, and post-evaluation. The growth in the number of computer systems and the increasing reliance upon them by individuals, businesses, industries and governments means that there are an increasing number of systems at risk. The computer systems of financial regulators and financial institutions like the U.S. Securities and Exchange Commission, SWIFT, investment banks, and commercial banks are prominent hacking targets for cybercriminals interested in manipulating markets and making illicit gains. Web sites and apps that accept or store credit card numbers, brokerage accounts, and bank account information are also prominent hacking targets, because of the potential for immediate financial gain from transferring money, making purchases, or selling the information on the black market. In-store payment systems and ATMs have also been tampered with in order to gather customer account data and PINs. Computers control functions at many utilities, including coordination of telecommunications, the power grid, nuclear power plants, and valve opening and closing in water and gas networks. The Internet is a potential attack vector for such machines if connected, but the Stuxnet worm demonstrated that even equipment controlled by computers not connected to the Internet can be vulnerable. In 2014, the Computer Emergency Readiness Team, a division of the Department of Homeland Security, investigated 79 hacking incidents at energy companies. Vulnerabilities in smart meters (many of which use local radio or cellular communications) can cause problems with billing fraud. The aviation industry is very reliant on a series of complex systems which could be attacked. A simple power outage at one airport can cause repercussions worldwide, much of the system relies on radio transmissions which could be disrupted, and controlling aircraft over oceans is especially dangerous because radar surveillance only extends 175 to 225 miles offshore. There is also potential for attack from within an aircraft. In Europe, with the (Pan-European Network Service) and NewPENS, and in the US with the NextGen program, air navigation service providers are moving to create their own dedicated networks. The consequences of a successful attack range from loss of confidentiality to loss of system integrity, air traffic control outages, loss of aircraft, and even loss of life. Desktop computers and laptops are commonly targeted to gather passwords or financial account information, or to construct a botnet to attack another target. Smartphones, tablet computers, smart watches, and other mobile devices such as quantified self devices like activity trackers have sensors such as cameras, microphones, GPS receivers, compasses, and accelerometers which could be exploited, and may collect personal information, including sensitive health information. WiFi, Bluetooth, and cell phone networks on any of these devices could be used as attack vectors, and sensors might be remotely activated after a successful breach. The increasing number of home automation devices such as the Nest thermostat are also potential targets. Large corporations are common targets. In many cases attacks are aimed at financial gain through identity theft and involve data breaches. Examples include loss of millions of clients' credit card details by Home Depot, Staples, Target Corporation, and the most recent breach of Equifax. Some cyberattacks are ordered by foreign governments, which engage in cyberwarfare with the intent to spread their propaganda, sabotage, or spy on their targets. Many people believe the Russian government played a major role in the US presidential election of 2016 by using Twitter and Facebook to affect the results of the election. Medical records have been targeted in general identify theft, health insurance fraud, and impersonating patients to obtain prescription drugs for recreational purposes or resale. Although cyber threats continue to increase, 62% of all organizations did not increase security training for their business in 2015. Not all attacks are financially motivated, however: security firm HBGary Federal suffered a serious series of attacks in 2011 from hacktivist group Anonymous in retaliation for the firm's CEO claiming to have infiltrated their group, and Sony Pictures was hacked in 2014 with the apparent dual motive of embarrassing the company through data leaks and crippling the company by wiping workstations and servers. Vehicles are increasingly computerized, with engine timing, cruise control, anti-lock brakes, seat belt tensioners, door locks, airbags and advanced driver-assistance systems on many models. Additionally, connected cars may use WiFi and Bluetooth to communicate with onboard consumer devices and the cell phone network. Self-driving cars are expected to be even more complex. All of these systems carry some security risk, and such issues have gained wide attention. Simple examples of risk include a malicious compact disc being used as an attack vector, and the car's onboard microphones being used for eavesdropping. However, if access is gained to a car's internal controller area network, the danger is much greater – and in a widely publicized 2015 test, hackers remotely carjacked a vehicle from 10 miles away and drove it into a ditch. Manufacturers are reacting in a number of ways, with Tesla in 2016 pushing out some security fixes "over the air" into its cars' computer systems. In the area of autonomous vehicles, in September 2016 the United States Department of Transportation announced some initial safety standards, and called for states to come up with uniform policies. Government and military computer systems are commonly attacked by activists and foreign powers. Local and regional government infrastructure such as traffic light controls, police and intelligence agency communications, personnel records, student records, and financial systems are also potential targets as they are now all largely computerized. Passports and government ID cards that control access to facilities which use RFID can be vulnerable to cloning. The Internet of things (IoT) is the network of physical objects such as devices, vehicles, and buildings that are embedded with electronics, software, sensors, and network connectivity that enables them to collect and exchange data – and concerns have been raised that this is being developed without appropriate consideration of the security challenges involved. While the IoT creates opportunities for more direct integration of the physical world into computer-based systems, it also provides opportunities for misuse. In particular, as the Internet of Things spreads widely, cyberattacks are likely to become an increasingly physical (rather than simply virtual) threat. If a front door's lock is connected to the Internet, and can be locked/unlocked from a phone, then a criminal could enter the home at the press of a button from a stolen or hacked phone. People could stand to lose much more than their credit card numbers in a world controlled by IoT-enabled devices. Thieves have also used electronic means to circumvent non-Internet-connected hotel door locks. An attack that targets physical infrastructure and/or human lives is classified as a Cyber-kinetic attack. As IoT devices and appliances gain currency, cyber-kinetic attacks can become pervasive and significantly damaging. Medical devices have either been successfully attacked or had potentially deadly vulnerabilities demonstrated, including both in-hospital diagnostic equipment and implanted devices including pacemakers and insulin pumps. There are many reports of hospitals and hospital organizations getting hacked, including ransomware attacks, Windows XP exploits, viruses, and data breaches of sensitive data stored on hospital servers. On 28 December 2016 the US Food and Drug Administration released its recommendations for how medical device manufacturers should maintain the security of Internet-connected devices – but no structure for enforcement. In distributed generation systems, the risk of a cyber attack is real, according to "Daily Energy Insider". An attack could cause a loss of power in a large area for a long period of time, and such an attack could have just as severe consequences as a natural disaster. The District of Columbia is considering creating a Distributed Energy Resources (DER) Authority within the city, with the goal being for customers to have more insight into their own energy use and giving the local electric utility, Pepco, the chance to better estimate energy demand. The D.C. proposal, however, would "allow third-party vendors to create numerous points of energy distribution, which could potentially create more opportunities for cyber attackers to threaten the electric grid." Serious financial damage has been caused by security breaches, but because there is no standard model for estimating the cost of an incident, the only data available is that which is made public by the organizations involved. "Several computer security consulting firms produce estimates of total worldwide losses attributable to virus and worm attacks and to hostile digital acts in general. The 2003 loss estimates by these firms range from $13 billion (worms and viruses only) to $226 billion (for all forms of covert attacks). The reliability of these estimates is often challenged; the underlying methodology is basically anecdotal." Security breaches continue to cost businesses billions of dollars but a survey revealed that 66% of security staffs do not believe senior leadership takes cyber precautions as a strategic priority. However, reasonable estimates of the financial cost of security breaches can actually help organizations make rational investment decisions. According to the classic Gordon-Loeb Model analyzing the optimal investment level in information security, one can conclude that the amount a firm spends to protect information should generally be only a small fraction of the expected loss (i.e., the expected value of the loss resulting from a cyber/information security breach). As with physical security, the motivations for breaches of computer security vary between attackers. Some are thrill-seekers or vandals, some are activists, others are criminals looking for financial gain. State-sponsored attackers are now common and well resourced but started with amateurs such as Markus Hess who hacked for the KGB, as recounted by Clifford Stoll in "The Cuckoo's Egg". Additionally, recent attacker motivations can be traced back to extremist organizations seeking to gain political advantage or disrupt social agendas. The growth of the internet, mobile technologies and inexpensive computing devices that has led to a rise in capabilities but also risk to environments that are deemed as vital to operations. All critical targeted environments are susceptible to compromise and has led to a series of proactive studies on how to migrate the risk by taking into consideration motivations by these type of actors. Several stark differences exist between the hacker motivation and that of nation state actors seeking to attack based an ideological preference. A standard part of threat modeling for any particular system is to identify what might motivate an attack on that system, and who might be motivated to breach it. The level and detail of precautions will vary depending on the system to be secured. A home personal computer, bank, and classified military network face very different threats, even when the underlying technologies in use are similar. In computer security a countermeasure is an action, device, procedure or technique that reduces a threat, a vulnerability, or an attack by eliminating or preventing it, by minimizing the harm it can cause, or by discovering and reporting it so that corrective action can be taken. Some common countermeasures are listed in the following sections: Security by design, or alternately secure by design, means that the software has been designed from the ground up to be secure. In this case, security is considered as a main feature. Some of the techniques in this approach include: The Open Security Architecture organization defines IT security architecture as "the design artifacts that describe how the security controls (security countermeasures) are positioned, and how they relate to the overall information technology architecture. These controls serve the purpose to maintain the system's quality attributes: confidentiality, integrity, availability, accountability and assurance services". Techopedia defines security architecture as "a unified security design that addresses the necessities and potential risks involved in a certain scenario or environment. It also specifies when and where to apply security controls. The design process is generally reproducible." The key attributes of security architecture are: Practicing security architecture provides the right foundation to systematically address business, IT and security concerns in an organization. A state of computer "security" is the conceptual ideal, attained by the use of the three processes: threat prevention, detection, and response. These processes are based on various policies and system components, which include the following: Today, computer security comprises mainly "preventive" measures, like firewalls or an exit procedure. A firewall can be defined as a way of filtering network data between a host or a network and another network, such as the Internet, and can be implemented as software running on the machine, hooking into the network stack (or, in the case of most UNIX-based operating systems such as Linux, built into the operating system kernel) to provide real-time filtering and blocking. Another implementation is a so-called "physical firewall", which consists of a separate machine filtering network traffic. Firewalls are common amongst machines that are permanently connected to the Internet. Some organizations are turning to big data platforms, such as Apache Hadoop, to extend data accessibility and machine learning to detect advanced persistent threats. However, relatively few organizations maintain computer systems with effective detection systems, and fewer still have organized response mechanisms in place. As a result, as Reuters points out: "Companies for the first time report they are losing more through electronic theft of data than physical stealing of assets". The primary obstacle to effective eradication of cybercrime could be traced to excessive reliance on firewalls and other automated "detection" systems. Yet it is basic evidence gathering by using packet capture appliances that puts criminals behind bars. In order to ensure adequate security, the confidentiality, integrity and availability of a network, better known as the CIA triad, must be protected and is considered the foundation to information security. To achieve those objectives, administrative, physical and technical security measures should be employed. The amount of security afforded to an asset can only be determined when its value is known. Vulnerability management is the cycle of identifying, and remediating or mitigating vulnerabilities, especially in software and firmware. Vulnerability management is integral to computer security and network security. Vulnerabilities can be discovered with a vulnerability scanner, which analyzes a computer system in search of known vulnerabilities, such as open ports, insecure software configuration, and susceptibility to malware. In order for these tools to be effective, they must be kept up to date with every new update the vendors release. Typically, these updates will scan for the new vulnerabilities that were introduced recently. Beyond vulnerability scanning, many organizations contract outside security auditors to run regular penetration tests against their systems to identify vulnerabilities. In some sectors, this is a contractual requirement. While formal verification of the correctness of computer systems is possible, it is not yet common. Operating systems formally verified include seL4, and SYSGO's PikeOS – but these make up a very small percentage of the market. Two factor authentication is a method for mitigating unauthorized access to a system or sensitive information. It requires "something you know"; a password or PIN, and "something you have"; a card, dongle, cellphone, or other piece of hardware. This increases security as an unauthorized person needs both of these to gain access. Social engineering and direct computer access (physical) attacks can only be prevented by non-computer means, which can be difficult to enforce, relative to the sensitivity of the information. Training is often involved to help mitigate this risk, but even in highly disciplined environments (e.g. military organizations), social engineering attacks can still be difficult to foresee and prevent. Inoculation, derived from inoculation theory, seeks to prevent social engineering and other fraudulent tricks or traps by instilling a resistance to persuasion attempts through exposure to similar or related attempts. It is possible to reduce an attacker's chances by keeping systems up to date with security patches and updates, using a security scanner and/or hiring people with expertise in security, though none of these guarantee the prevention of an attack. The effects of data loss/damage can be reduced by careful backing up and insurance. While hardware may be a source of insecurity, such as with microchip vulnerabilities maliciously introduced during the manufacturing process, hardware-based or assisted computer security also offers an alternative to software-only computer security. Using devices and methods such as dongles, trusted platform modules, intrusion-aware cases, drive locks, disabling USB ports, and mobile-enabled access may be considered more secure due to the physical access (or sophisticated backdoor access) required in order to be compromised. Each of these is covered in more detail below. One use of the term "computer security" refers to technology that is used to implement secure operating systems. In the 1980s the United States Department of Defense (DoD) used the "Orange Book" standards, but the current international standard ISO/IEC 15408, "Common Criteria" defines a number of progressively more stringent Evaluation Assurance Levels. Many common operating systems meet the EAL4 standard of being "Methodically Designed, Tested and Reviewed", but the formal verification required for the highest levels means that they are uncommon. An example of an EAL6 ("Semiformally Verified Design and Tested") system is Integrity-178B, which is used in the Airbus A380 and several military jets. In software engineering, secure coding aims to guard against the accidental introduction of security vulnerabilities. It is also possible to create software designed from the ground up to be secure. Such systems are "secure by design". Beyond this, formal verification aims to prove the correctness of the algorithms underlying a system; important for cryptographic protocols for example. Within computer systems, two of many security models capable of enforcing privilege separation are access control lists (ACLs) and capability-based security. Using ACLs to confine programs has been proven to be insecure in many situations, such as if the host computer can be tricked into indirectly allowing restricted file access, an issue known as the confused deputy problem. It has also been shown that the promise of ACLs of giving access to an object to only one person can never be guaranteed in practice. Both of these problems are resolved by capabilities. This does not mean practical flaws exist in all ACL-based systems, but only that the designers of certain utilities must take responsibility to ensure that they do not introduce flaws. Capabilities have been mostly restricted to research operating systems, while commercial OSs still use ACLs. Capabilities can, however, also be implemented at the language level, leading to a style of programming that is essentially a refinement of standard object-oriented design. An open-source project in the area is the E language. The end-user is widely recognized as the weakest link in the security chain and it is estimated that more than 90% of security incidents and breaches involve some kind of human error. Among the most commonly recorded forms of errors and misjudgment are poor password management, the inability to recognize misleading URLs and to identify fake websites and dangerous email attachments. A common mistake that users make is saving their userid/password in their browsers to make it easier to log in to banking sites. This is a gift to attackers who have obtained access to a machine by some means. The risk may be mitigated by the use of two-factor authentication. As the human component of cyber risk is particularly relevant in determining the global cyber risk an organization is facing, security awareness training, at all levels, not only provides formal compliance with regulatory and industry mandates but is considered essential in reducing cyber risk and protecting individuals and companies from the great majority of cyber threats. The focus on the end-user represents a profound cultural change for many security practitioners, who have traditionally approached cybersecurity exclusively from a technical perspective, and moves along the lines suggested by major security centers to develop a culture of cyber awareness within the organization, recognizing that a security-aware user provides an important line of defense against cyber attacks. Related to end-user training, digital hygiene or cyber hygiene is a fundamental principle relating to information security and, as the analogy with personal hygiene shows, is the equivalent of establishing simple routine measures to minimize the risks from cyber threats. The assumption is that good cyber hygiene practices can give networked users another layer of protection, reducing the risk that one vulnerable node will be used to either mount attacks or compromise another node or network, especially from common cyberattacks. As opposed to a purely technology-based defense against threats, cyber hygiene mostly regards routine measures that are technically simple to implement and mostly dependent on discipline or education. It can be thought of as an abstract list of tips or measures that have been demonstrated as having a positive effect on personal and/or collective digital security. As such, these measures can be performed by laypeople, not just security experts. Cyber hygiene relates to personal hygiene as computer viruses relate to biological viruses (or pathogens). However, while the term "computer virus" was coined almost simultaneously with the creation of the first working computer viruses, the term "cyber hygiene" is a much later invention, perhaps as late as 2000 by Internet pioneer Vint Cerf. It has since been adopted by the Congress and Senate of the United States, the FBI, EU institutions and heads of state. Cyber hygiene should also not be mistaken for proactive cyber defence, a military term. Responding forcefully to attempted security breaches (in the manner that one would for attempted physical security breaches) is often very difficult for a variety of reasons: Incident response is an organized approach to addressing and managing the aftermath of a computer security incident or compromise with the goal of preventing a breach or thwarting a cyberattack. An incident that is not identified and managed at the time of intrusion typically escalates to a more damaging event such as a data breach or system failure. The intended outcome of a computer security incident response plan is to limit damage and reduce recovery time and costs. Responding to compromises quickly can mitigate exploited vulnerabilities, restore services and processes and minimize losses. Incident response planning allows an organization to establish a series of best practices to stop an intrusion before it causes damage. Typical incident response plans contain a set of written instructions that outline the organization's response to a cyberattack. Without a documented plan in place, an organization may not successfully detect an intrusion or compromise and stakeholders may not understand their roles, processes and procedures during an escalation, slowing the organization's response and resolution. There are four key components of a computer security incident response plan: Some illustrative examples of different types of computer security breaches are given below. In 1988, only 60,000 computers were connected to the Internet, and most were mainframes, minicomputers and professional workstations. On 2 November 1988, many started to slow down, because they were running a malicious code that demanded processor time and that spread itself to other computers – the first internet "computer worm". The software was traced back to 23-year-old Cornell University graduate student Robert Tappan Morris, Jr. who said "he wanted to count how many machines were connected to the Internet". In 1994, over a hundred intrusions were made by unidentified crackers into the Rome Laboratory, the US Air Force's main command and research facility. Using trojan horses, hackers were able to obtain unrestricted access to Rome's networking systems and remove traces of their activities. The intruders were able to obtain classified files, such as air tasking order systems data and furthermore able to penetrate connected networks of National Aeronautics and Space Administration's Goddard Space Flight Center, Wright-Patterson Air Force Base, some Defense contractors, and other private sector organizations, by posing as a trusted Rome center user. In early 2007, American apparel and home goods company TJX announced that it was the victim of an unauthorized computer systems intrusion and that the hackers had accessed a system that stored data on credit card, debit card, check, and merchandise return transactions. In 2010 the computer worm known as Stuxnet reportedly ruined almost one-fifth of Iran's nuclear centrifuges. It did so by disrupting industrial programmable logic controllers (PLCs) in a targeted attack. This is generally believed to have been launched by Israel and the United States to disrupt Iranian's nuclear program – although neither has publicly admitted this. In early 2013, documents provided by Edward Snowden were published by "The Washington Post" and "The Guardian" exposing the massive scale of NSA global surveillance. There were also indications that the NSA may have inserted a backdoor in a NIST standard for encryption. This standard was later withdrawn due to widespread criticism. The NSA additionally were revealed to have tapped the links between Google's data centres. In 2013 and 2014, a Russian/Ukrainian hacking ring known as "Rescator" broke into Target Corporation computers in 2013, stealing roughly 40 million credit cards, and then Home Depot computers in 2014, stealing between 53 and 56 million credit card numbers. Warnings were delivered at both corporations, but ignored; physical security breaches using self checkout machines are believed to have played a large role. "The malware utilized is absolutely unsophisticated and uninteresting," says Jim Walter, director of threat intelligence operations at security technology company McAfee – meaning that the heists could have easily been stopped by existing antivirus software had administrators responded to the warnings. The size of the thefts has resulted in major attention from state and Federal United States authorities and the investigation is ongoing. In April 2015, the Office of Personnel Management discovered it had been hacked more than a year earlier in a data breach, resulting in the theft of approximately 21.5 million personnel records handled by the office. The Office of Personnel Management hack has been described by federal officials as among the largest breaches of government data in the history of the United States. Data targeted in the breach included personally identifiable information such as Social Security Numbers, names, dates and places of birth, addresses, and fingerprints of current and former government employees as well as anyone who had undergone a government background check. It is believed the hack was perpetrated by Chinese hackers. In July 2015, a hacker group known as "The Impact Team" successfully breached the extramarital relationship website Ashley Madison, created by Avid Life Media. The group claimed that they had taken not only company data but user data as well. After the breach, The Impact Team dumped emails from the company's CEO, to prove their point, and threatened to dump customer data unless the website was taken down permanently." When Avid Life Media did not take the site offline the group released two more compressed files, one 9.7GB and the second 20GB. After the second data dump, Avid Life Media CEO Noel Biderman resigned; but the website remained functioning. International legal issues of cyber attacks are complicated in nature. There is no global base of common rules to judge, and eventually punish, cybercrimes and cybercriminals - and where security firms or agencies do locate the cybercriminal behind the creation of a particular piece of malware or form of cyber attack, often the local authorities cannot take action due to lack of laws under which to prosecute. Proving attribution for cybercrimes and cyberattacks is also a major problem for all law enforcement agencies. "Computer viruses switch from one country to another, from one jurisdiction to another – moving around the world, using the fact that we don't have the capability to globally police operations like this. So the Internet is as if someone [had] given free plane tickets to all the online criminals of the world." The use of techniques such as dynamic DNS, fast flux and bullet proof servers add to the difficulty of investigation and enforcement. The role of the government is to make regulations to force companies and organizations to protect their systems, infrastructure and information from any cyberattacks, but also to protect its own national infrastructure such as the national power-grid. The government's regulatory role in cyberspace is complicated. For some, cyberspace was seen virtual space that was to remain free of government intervention, as can be seen in many of today's libertarian blockchain and bitcoin discussions. Many government officials and experts think that the government should do more and that there is a crucial need for improved regulation, mainly due to the failure of the private sector to solve efficiently the cybersecurity problem. R. Clarke said during a panel discussion at the RSA Security Conference in San Francisco, he believes that the "industry only responds when you threaten regulation. If the industry doesn't respond (to the threat), you have to follow through." On the other hand, executives from the private sector agree that improvements are necessary, but think that government intervention would affect their ability to innovate efficiently. Daniel R. McCarthy analyzed this public-private partnership in cybersecurity and reflected on the role of cybersecurity in the broader constitution of political order. On May 22, 2020, the UN Security Council held its second ever informal meeting on cybersecurity to focus on cyber challenges to international peace. According to UN Secretary-General António Guterres, new technologies are too often used to violate rights. Many different teams and organisations exist, including: On 14 April 2016 the European Parliament and Council of the European Union adopted The General Data Protection Regulation (GDPR) (EU) 2016/679. GDPR, which became enforceable beginning 25 May 2018, provides for data protection and privacy for all individuals within the European Union (EU) and the European Economic Area (EEA). GDPR requires that business processes that handle personal data be built with data protection by design and by default. GDPR also requires that certain organizations appoint a Data Protection Officer (DPO). Most countries have their own computer emergency response team to protect network security. Since 2010, Canada has had a cybersecurity strategy. This functions as a counterpart document to the National Strategy and Action Plan for Critical Infrastructure. The strategy has three main pillars: securing government systems, securing vital private cyber systems, and helping Canadians to be secure online. There is also a Cyber Incident Management Framework to provide a coordinated response in the event of a cyber incident. The Canadian Cyber Incident Response Centre (CCIRC) is responsible for mitigating and responding to threats to Canada's critical infrastructure and cyber systems. It provides support to mitigate cyber threats, technical support to respond and recover from targeted cyber attacks, and provides online tools for members of Canada's critical infrastructure sectors. It posts regular cybersecurity bulletins and operates an online reporting tool where individuals and organizations can report a cyber incident. To inform the general public on how to protect themselves online, Public Safety Canada has partnered with STOP.THINK.CONNECT, a coalition of non-profit, private sector, and government organizations, and launched the Cyber Security Cooperation Program. They also run the GetCyberSafe portal for Canadian citizens, and Cyber Security Awareness Month during October. Public Safety Canada aims to begin an evaluation of Canada's cybersecurity strategy in early 2015. China's Central Leading Group for Internet Security and Informatization () was established on 27 February 2014. This Leading Small Group (LSG) of the Communist Party of China is headed by General Secretary Xi Jinping himself and is staffed with relevant Party and state decision-makers. The LSG was created to overcome the incoherent policies and overlapping responsibilities that characterized China's former cyberspace decision-making mechanisms. The LSG oversees policy-making in the economic, political, cultural, social and military fields as they relate to network security and IT strategy. This LSG also coordinates major policy initiatives in the international arena that promote norms and standards favored by the Chinese government and that emphasize the principle of national sovereignty in cyberspace. Berlin starts National Cyber Defense Initiative: On 16 June 2011, the German Minister for Home Affairs, officially opened the new German NCAZ (National Center for Cyber Defense) Nationales Cyber-Abwehrzentrum located in Bonn. The NCAZ closely cooperates with BSI (Federal Office for Information Security) Bundesamt für Sicherheit in der Informationstechnik, BKA (Federal Police Organisation) Bundeskriminalamt (Deutschland), BND (Federal Intelligence Service) Bundesnachrichtendienst, MAD (Military Intelligence Service) Amt für den Militärischen Abschirmdienst and other national organizations in Germany taking care of national security aspects. According to the Minister the primary task of the new organization founded on 23 February 2011, is to detect and prevent attacks against the national infrastructure and mentioned incidents like Stuxnet. Germany has also established the largest research institution for IT security in Europe, the Center for Research in Security and Privacy (CRISP) in Darmstadt. Some provisions for cybersecurity have been incorporated into rules framed under the Information Technology Act 2000. The National Cyber Security Policy 2013 is a policy framework by Ministry of Electronics and Information Technology (MeitY) which aims to protect the public and private infrastructure from cyberattacks, and safeguard "information, such as personal information (of web users), financial and banking information and sovereign data". CERT- In is the nodal agency which monitors the cyber threats in the country. The post of National Cyber Security Coordinator has also been created in the Prime Minister's Office (PMO). The Indian Companies Act 2013 has also introduced cyber law and cybersecurity obligations on the part of Indian directors. Some provisions for cybersecurity have been incorporated into rules framed under the Information Technology Act 2000 Update in 2013. Following cyber attacks in the first half of 2013, when the government, news media, television station, and bank websites were compromised, the national government committed to the training of 5,000 new cybersecurity experts by 2017. The South Korean government blamed its northern counterpart for these attacks, as well as incidents that occurred in 2009, 2011, and 2012, but Pyongyang denies the accusations. The 1986 , the Computer Fraud and Abuse Act is the key legislation. It prohibits unauthorized access or damage of "protected computers" as defined in . Although various other measures have been proposed – none has succeeded. In 2013, executive order "Improving Critical Infrastructure Cybersecurity" was signed, which prompted the creation of the NIST Cybersecurity Framework Standardized Government Testing Services The General Services Administration (GSA) has standardized the "penetration test" service as a pre-vetted support service, to rapidly address potential vulnerabilities, and stop adversaries before they impact US federal, state and local governments. These services are commonly referred to as Highly Adaptive Cybersecurity Services (HACS) and are listed at the US GSA Advantage website. See more information here: Penetration test: Standardized government penetration test services. The Department of Homeland Security has a dedicated division responsible for the response system, risk management program and requirements for cybersecurity in the United States called the National Cyber Security Division. The division is home to US-CERT operations and the National Cyber Alert System. The National Cybersecurity and Communications Integration Center brings together government organizations responsible for protecting computer networks and networked infrastructure. The third priority of the Federal Bureau of Investigation (FBI) is to: "Protect the United States against cyber-based attacks and high-technology crimes", and they, along with the National White Collar Crime Center (NW3C), and the Bureau of Justice Assistance (BJA) are part of the multi-agency task force, The Internet Crime Complaint Center, also known as IC3. In addition to its own specific duties, the FBI participates alongside non-profit organizations such as InfraGard. In the criminal division of the United States Department of Justice operates a section called the Computer Crime and Intellectual Property Section. The CCIPS is in charge of investigating computer crime and intellectual property crime and is specialized in the search and seizure of digital evidence in computers and networks. In 2017, CCIPS published A Framework for a Vulnerability Disclosure Program for Online Systems to help organizations "clearly describe authorized vulnerability disclosure and discovery conduct, thereby substantially reducing the likelihood that such described activities will result in a civil or criminal violation of law under the Computer Fraud and Abuse Act (18 U.S.C. § 1030)." The United States Cyber Command, also known as USCYBERCOM, ""has the mission to direct, synchronize, and coordinate cyberspace planning and operations to defend and advance national interests in collaboration with domestic and international partners".""" It has no role in the protection of civilian networks. The U.S. Federal Communications Commission's role in cybersecurity is to strengthen the protection of critical communications infrastructure, to assist in maintaining the reliability of networks during disasters, to aid in swift recovery after, and to ensure that first responders have access to effective communications services. The Food and Drug Administration has issued guidance for medical devices, and the National Highway Traffic Safety Administration is concerned with automotive cybersecurity. After being criticized by the Government Accountability Office, and following successful attacks on airports and claimed attacks on airplanes, the Federal Aviation Administration has devoted funding to securing systems on board the planes of private manufacturers, and the Aircraft Communications Addressing and Reporting System. Concerns have also been raised about the future Next Generation Air Transportation System. "Computer emergency response team" is a name given to expert groups that handle computer security incidents. In the US, two distinct organization exist, although they do work closely together. There is growing concern that cyberspace will become the next theater of warfare. As Mark Clayton from the "Christian Science Monitor" described in an article titled "The New Cyber Arms Race": In the future, wars will not just be fought by soldiers with guns or with planes that drop bombs. They will also be fought with the click of a mouse a half a world away that unleashes carefully weaponized computer programs that disrupt or destroy critical industries like utilities, transportation, communications, and energy. Such attacks could also disable military networks that control the movement of troops, the path of jet fighters, the command and control of warships. This has led to new terms such as "cyberwarfare" and "cyberterrorism". The United States Cyber Command was created in 2009 and many other countries have similar forces. There are a few critical voices that question whether cybersecurity is as significant a threat as it is made out to be. Cybersecurity is a fast-growing field of IT concerned with reducing organizations' risk of hack or data breach. According to research from the Enterprise Strategy Group, 46% of organizations say that they have a "problematic shortage" of cybersecurity skills in 2016, up from 28% in 2015. Commercial, government and non-governmental organizations all employ cybersecurity professionals. The fastest increases in demand for cybersecurity workers are in industries managing increasing volumes of consumer data such as finance, health care, and retail. However, the use of the term "cybersecurity" is more prevalent in government job descriptions. Typical cybersecurity job titles and descriptions include: Student programs are also available to people interested in beginning a career in cybersecurity. Meanwhile, a flexible and effective option for information security professionals of all experience levels to keep studying is online security training, including webcasts. A wide range of certified courses are also available. In the United Kingdom, a nationwide set of cybersecurity forums, known as the U.K Cyber Security Forum, were established supported by the Government's cybersecurity strategy in order to encourage start-ups and innovation and to address the skills gap identified by the U.K Government. The following terms used with regards to computer security are explained below:
https://en.wikipedia.org/wiki?curid=7398
Chris Cunningham Chris Cunningham (born 15 October 1970) is a British video artist and music video director, best known for his music videos for electronic musicians such as Autechre, Squarepusher, and most notably Aphex Twin on videos for "Windowlicker" and "Come to Daddy", and Björk's "All is Full of Love". All were used in Chris' chapter in Director's Label. He has also created art installations and directed short movies. He was approached to direct a movie version of William Gibson's cyberpunk novel "Neuromancer"; the project has been in development hell for more than two decades. In the 2000s, Cunningham began doing music production work, and has also designed album artwork for a variety of musicians. After seeing Cunningham's work on the 1995 film version of "Judge Dredd", Stanley Kubrick head-hunted Cunningham to design and supervise animatronic tests of the central robot child character in his version of the film "A.I. Artificial Intelligence". Cunningham worked for over a year on the film before leaving to pursue a career as a director. Earlier work in film included model-making, prosthetic make-up and concept illustrations for "Hardware" and "Dust Devil" for director Richard Stanley, work on "Nightbreed" for Clive Barker, and on "Alien3" for David Fincher. Between 1990 and 1992, he contributed the occasional cover painting and strip to "Judge Dredd Megazine", working under the pseudonym "Chris Halls"; Halls is his stepfather's surname. Cunningham has had close ties to Warp Records since his first video for Autechre, "Second Bad Vilbel", which received airplay on MTV's "Amp". Videos for Aphex Twin's "Come to Daddy" and "Windowlicker" are perhaps his best known. His video for Björk's "All Is Full of Love" won multiple awards, including an MTV music video award for Breakthrough Video and was nominated for a Grammy for Best Short Form Music Video. It was also the first ever music video to win a Gold Pencil at the D&AD Awards. It can still be seen at the Museum of Modern Art in New York. His video for Aphex Twin's "Windowlicker" was nominated for the "Best Video" award at the Brit Awards 2000. He also directed Madonna's "Frozen" video which became an international hit and won the award for Best Special Effects at the 1998 MTV Music Video Awards. Cunningham also came out of a seven-year hiatus from making music videos to direct the video for "Sheena Is a Parasite" by The Horrors. His video installation "Flex" was first shown in 2000 at the Royal Academy of Arts, and subsequently at the Anthony d'Offay Gallery and other art galleries. "Flex" was commissioned by the Anthony d'Offay Gallery for the exhibition curated by Norman Rosenthal and Max Wigram at the Royal Academy of Arts in 2000. The Anthony d'Offay Gallery also commissioned "Monkey Drummer", a 2½ minute piece intended for exhibition as a companion to "Flex" at the 2000 "Apocalypse" exhibition at the Royal Academy of Arts: however, the piece was not finished in time. In it an automaton with nine appendages and the head of a monkey plays the drums to "Mt Saint Michel + Saint Michaels Mount", the 10th track on Aphex Twin's 2001 album "drukqs". "Monkey Drummer" debuted as part of Cunningham's installation at the 49th International Exhibition of Art at the 2001 Venice Biennale, which consisted of a loop of "Monkey Drummer", "Flex", and his video for Björk's "All Is Full of Love". In 2002 both "Flex" and "Monkey Drummer" were exhibited by 5th Gallery in Dublin, Ireland, in an exhibition curated by Artist/Curator Paul Murnaghan, In 2007, an excerpt from "Flex" was shown in the Barbican's exhibition Seduced: Art and Sex from Antiquity to Now curated by Martin Kemp, Marina Wallace and Joanne Bernstein. alongside other pieces by Bacon, Klimt, Rembrandt, Rodin and Picasso. In 2005, Cunningham released the short film "Rubber Johnny" as a DVD accompanied by a book of photographs and drawings. "Rubber Johnny", a six-minute experimental short film cut to a soundtrack by Aphex Twin remixed by Cunningham, was shot between 2001 and 2004. Shot on DV night-vision, it was made in Cunningham's own time as a home movie of sorts, and took three and half years of weekends to complete. The Telegraph called it "like a Looney Tunes short for a generation raised on video nasties and rave music". During this period Cunningham also made another short film for Warp Films, "Spectral Musicians", which remains unreleased. The short film was set to Squarepusher's "My Fucking Sound" from his album "Go Plastic"; and to a piece called "Mutilation Colony" which was written especially for the short, and was released on the EP "Do You Know Squarepusher". Cunningham has directed a handful of commercials for companies and brands, including Gucci, Sony (PlayStation), Levi's, Telecom Italia, Nissan, and Orange. In 2004/2005, Cunningham took a sabbatical from filmmaking to learn about music production and recording and to develop his own music projects. In December 2007 Cunningham produced two tracks, "Three Decades" and "Primary Colours", for "Primary Colours", the second album by The Horrors. In the summer of 2008, due to scheduling conflicts with his feature film script writing he could not work on the rest of the album which was subsequently recorded by Geoff Barrow from Portishead. In 2008, he produced and arranged a new version of 'I Feel Love' for the Gucci commercial that he also directed. He travelled to Nashville to work with Donna Summer to record a brand new vocal for it. In 2005, Cunningham played a 45-minute audio visual piece performed live in Tokyo and Osaka in front of 30,000+ fans over the two nights at Japan's premier electronic music event Electraglide. These performances evolved into "Chris Cunningham Live", a 55-minute long performance piece combining original and remixed music and film. It features remixed, unreleased and brand new videos and music dynamically edited together into a new live piece spread over three screens. The sound accompanying these images includes Cunningham's first publicly performed compositions interspersed with his remixes of other artist's work. "Chris Cunningham Live" debuted as one of the headline attractions at Warp 20 in Paris on 8 May 2009 with other performances scheduled at festivals in UK, and a number of European cities later in the year. "Chris Cunningham Live" continued in June 2011, with performances in London, Barcelona, and Sydney, Australia. Cunningham has created photography and cover artwork for various people including Björk's "All Is Full of Love", Aphex Twin's "Windowlicker" and "Come to Daddy". In 2008, Cunningham produced a fashion shoot for "Dazed & Confused" using Grace Jones as a model to create "Nubian versions" of Rubber Johnny. In an interview for BBC's "The Culture Show", it was suggested that the collaboration may expand into a video project. In regards to the collaboration, Cunningham stated "For me, Grace has the strongest iconography of any artist in music. She’s definitely the most inspiring person I’ve worked with so far". In November 2008, Cunningham followed on with another photoshoot for "Vice Magazine". In 2000, Cunningham and cyberpunk author William Gibson began work on the script for Gibson's 1984 novel "Neuromancer". However, because "Neuromancer" was due to be a big budget studio film, it is rumoured that Cunningham pulled out due to being a first time director without final cut approval. He also felt that too much of the original book's ideas had been cannibalised by other recent films. On 18 November 2004, in the FAQ on the William Gibson Board, Gibson was asked: In an August 1999 "Spike Magazine" interview, Gibson stated "He (Chris) was brought to my attention by someone else. We were told, third-hand, that he was extremely wary of the Hollywood process, and wouldn't return calls. But someone else told us that "Neuromancer" had been his "Wind in the Willows", that he'd read it when he was a kid. I went to London and we met." Gibson is also quoted in the article as saying "Chris is my own 100 per cent personal choice...My only choice. The only person I've met who I thought might have a hope in hell of doing it right. I went back to see him in London just after he'd finished the Bjork video, and I sat on a couch beside this dead sex little Bjork robot, except it was wearing Aphex Twin's head. We talked." It is rumoured that the character of Damien Pease in Gibson's 2003 novel "Pattern Recognition" was based on Cunningham, with the character's apartment featuring a female robot which had appeared in one of Cunningham's videos. Development funding was in place for Cunningham to direct and co-write his first feature film for Warp Films, to whom he was at the time committed "for all future full-length film projects." He has since left Warp Films to set up his own production company 'CC Co' to produce his films independently. Cunningham was married to Warpaint's bassist Jenny Lee Lindberg. They are currently no longer together. The video collection "The Work of Director Chris Cunningham" was released in November 2004 as part of the Directors Label set. This DVD includes selected highlights from 1995–2000.
https://en.wikipedia.org/wiki?curid=7400
Centaur A centaur (; , "kéntauros", ), or occasionally hippocentaur, is a creature from Greek mythology with the upper body of a human and the lower body and legs of a horse. Centaurs are thought of in many Greek myths as being as wild as untamed horses, and were said to have inhabited the region of Magnesia and Mount Pelion in Thessaly, the Foloi oak forest in Elis, and the Malean peninsula in southern Laconia. Centaurs are subsequently featured in Roman mythology, and were familiar figures in the medieval bestiary. They remain a staple of modern fantastic literature. The centaurs were usually said to have been born of Ixion and Nephele. As the story goes, Nephele was a cloud made into the likeness of Hera in a plot to trick Ixion into revealing his lust for Hera to Zeus. Ixion seduced Nephele and from that relationship centaurs were created. Another version, however, makes them children of Centaurus, a man who mated with the Magnesian mares. Centaurus was either himself the son of Ixion and Nephele (inserting an additional generation) or of Apollo and the nymph Stilbe. In the latter version of the story, Centaurus's twin brother was Lapithes, ancestor of the Lapiths. Another tribe of centaurs was said to have lived on Cyprus. According to Nonnus, they were fathered by Zeus, who, in frustration after Aphrodite had eluded him, spilled his seed on the ground of that land. Unlike those of mainland Greece, the Cyprian centaurs were horned. There were also the Lamian Pheres, twelve rustic daimones (spirits) of the Lamos river. They were set by Zeus to guard the infant Dionysos, protecting him from the machinations of Hera, but the enraged goddess transformed them into ox-horned Centaurs. The Lamian Pheres later accompanied Dionysos in his campaign against the Indians. The centaur's half-human, half-horse composition has led many writers to treat them as liminal beings, caught between the two natures they embody in contrasting myths; they are both the embodiment of untamed nature, as in their battle with the Lapiths (their kin), and conversely, teachers like Chiron. The Centaurs are best known for their fight with the Lapiths who, according to one origin myth, would have been cousins to the centaurs. The battle, called the Centauromachy, was caused by the centaurs' attempt to carry off Hippodamia and the rest of the Lapith women on the day of Hippodamia's marriage to Pirithous, who was the king of the Lapithae and a son of Ixion. Theseus, a hero and founder of cities, who happened to be present, threw the balance in favour of the Lapiths by assisting Pirithous in the battle. The Centaurs were driven off or destroyed. Another Lapith hero, Caeneus, who was invulnerable to weapons, was beaten into the earth by Centaurs wielding rocks and the branches of trees. In her article "The Centaur: Its History and Meaning in Human Culture," Elizabeth Lawrence claims that the contests between the centaurs and the Lapiths typify the struggle between civilization and barbarism. The Centauromachy is most famously portrayed in the Parthenon metopes by Phidias and in a Renaissance-era sculpture by Michelangelo. The Greek word "kentauros" is generally regarded as being of obscure origin. The etymology from "ken + tauros", "piercing bull," was a euhemerist suggestion in Palaephatus' rationalizing text on Greek mythology, "On Incredible Tales" (Περὶ ἀπίστων), which included mounted archers from a village called "Nephele" eliminating a herd of bulls that were the scourge of Ixion's kingdom. Another possible related etymology can be "bull-slayer". The most common theory holds that the idea of centaurs came from the first reaction of a non-riding culture, as in the Minoan Aegean world, to nomads who were mounted on horses. The theory suggests that such riders would appear as half-man, half-animal. Bernal Díaz del Castillo reported that the Aztecs also had this misapprehension about Spanish cavalrymen. The Lapith tribe of Thessaly, who were the kinsmen of the Centaurs in myth, were described as the inventors of horse-riding by Greek writers. The Thessalian tribes also claimed their horse breeds were descended from the centaurs. Robert Graves (relying on the work of Georges Dumézil, who argued for tracing the centaurs back to the Indian Gandharva), speculated that the centaurs were a dimly remembered, pre-Hellenic fraternal earth cult who had the horse as a totem. A similar theory was incorporated into Mary Renault's "The Bull from the Sea." Though female centaurs, called centaurides or centauresses, are not mentioned in early Greek literature and art, they do appear occasionally in later antiquity. A Macedonian mosaic of the 4th century BC is one of the earliest examples of the centauress in art. Ovid also mentions a centauress named Hylonome who committed suicide when her husband Cyllarus was killed in the war with the Lapiths. The Kalibangan cylinder seal, dated to be around 2600-1900 BC, found at the site of Indus-Valley civilization shows a battle between men in the presence of centaur-like creatures. Other sources claim the creatures represented are actually half human and half tigers, later evolving into the Hindu Goddess of War. These seals are also evidence of Indus-Mesopotamia relations in the 3rd millennium BC. In a popular legend associated with Pazhaya Sreekanteswaram Temple in Thiruvananthapuram, the curse of a saintly Brahmin transformed a handsome Yadava prince into a creature having a horse's body and the prince's head, arms, and torso in place of the head and neck of the horse. Kinnaras, another half-man, half-horse mythical creature from Indian mythology, appeared in various ancient texts, arts, and sculptures from all around India. It is shown as a horse with the torso of a man where the horse's head would be, and is similar to a Greek centaur. A centaur-like half-human, half-equine creature called "Polkan" appeared in Russian folk art and lubok prints of the 17th–19th centuries. Polkan is originally based on "Pulicane", a half-dog from Andrea da Barberino's poem "I Reali di Francia", which was once popular in the Slavonic world in prosaic translations. The extensive Mycenaean pottery found at Ugarit included two fragmentary Mycenaean terracotta figures which have been tentatively identified as centaurs. This finding suggests a Bronze Age origin for these creatures of myth. A painted terracotta centaur was found in the "Hero's tomb" at Lefkandi, and by the Geometric period, centaurs figure among the first representational figures painted on Greek pottery. An often-published Geometric period bronze of a warrior face-to-face with a centaur is at the Metropolitan Museum of Art. In Greek art of the Archaic period, centaurs are depicted in three different forms. Some centaurs are depicted with a human torso attached to the body of a horse at the withers, where the horse's neck would be; this form, designated "Class A" by Professor Paul Baur, later became standard. "Class B" centaurs are depicted with a human body and legs joined at the waist to the hindquarters of a horse; in some cases centaurs of both Class A and Class B appear together. A third type, designated "Class C", depicts centaurs with human forelegs terminating in hooves. Baur describes this as an apparent development of Aeolic art, which never became particularly widespread. At a later period, paintings on some "amphorae" depict winged centaurs. Centaurs were also frequently depicted in Roman art. One example is the pair of centaurs drawing the chariot of Constantine the Great and his family in the Great Cameo of Constantine ("circa" AD 314–16), which embodies wholly pagan imagery, and contrasts sharply with the popular image of Constantine as the patron of early Christianity. Centaurs preserved a Dionysian connection in the 12th-century Romanesque carved capitals of Mozac Abbey in the Auvergne. Other similar capitals depict harvesters, boys riding goats (a further Dionysiac theme), and griffins guarding the chalice that held the wine. Centaurs are also shown on a number of Pictish carved stones from north-east Scotland erected in the 8th–9th centuries AD (e.g., at Meigle, Perthshire). Though outside the limits of the Roman Empire, these depictions appear to be derived from Classical prototypes. The John C. Hodges library at The University of Tennessee hosts a permanent exhibit of a "Centaur from Volos" in its library. The exhibit, made by sculptor Bill Willers by combining a study human skeleton with the skeleton of a Shetland pony, is entitled "Do you believe in Centaurs?". According to the exhibitors, it was meant to mislead students in order to make them more critically aware. Another exhibit by Willers is now on long-term display at the International Wildlife Museum in Tucson, Arizona. The full-mount skeleton of a Centaur, built by Skulls Unlimited International, Inc., is on display along with several other fabled creatures, including the Cyclopes, Unicorn, and Griffin. Centaurs are common in European heraldry, although more frequent in continental than in British arms. A centaur holding a bow is referred to as a sagittarius. Jerome's version of the "Life" of St Anthony the Great, written by Athanasius of Alexandria about the hermit monk of Egypt, was widely disseminated in the Middle Ages; it relates Anthony's encounter with a centaur who challenged the saint, but was forced to admit that the old gods had been overthrown. The episode was often depicted in "The Meeting of St Anthony Abbot and St Paul the Hermit" by the painter Stefano di Giovanni, who was known as "Sassetta". Of the two episodic depictions of the hermit Anthony's travel to greet the hermit Paul, one is his encounter with the demonic figure of a centaur along the pathway in a wood. Lucretius, in his first-century BC philosophical poem "On the Nature of Things," denied the existence of centaurs based on their differing rate of growth. He states that at the age of three years, horses are in the prime of their life while humans at the same age are still little more than babies, making hybrid animals impossible. C.S. Lewis' "The Chronicles of Narnia" series depicts centaurs as the wisest and noblest of creatures. Narnian Centaurs are gifted at stargazing, prophecy, healing, and warfare; a fierce and valiant race always faithful to the High King Aslan the Lion. In J.K. Rowling's "Harry Potter" series, centaurs live in the Forbidden Forest close to Hogwarts, preferring to avoid contact with humans. They live in societies called herds and are skilled at archery, healing, and astrology, but like in the original myths, they are known to have some wild and barbarous tendencies. With the exception of Chiron, the centaurs in Rick Riordan's "Percy Jackson & the Olympians" are seen as wild party-goers who use a lot of American slang. Chiron retains his mythological role as a trainer of heroes and is skilled in archery. In Riordan's subsequent series, "Heroes of Olympus", another group of centaurs are depicted with more animalistic features (such as horns) and appear as villains, serving the Gigantes. Philip Jose Farmer's "World of Tiers" series (1965) includes centaurs, called Half-Horses or Hoi Kentauroi. His creations address several of the metabolic problems of such creatures—how could the human mouth and nose intake sufficient air to sustain both itself and the horse body and, similarly, how could the human ingest sufficient food to sustain both parts. Brandon Mull's "Fablehaven" series features centaurs that live in an area called Grunhold. The centaurs are portrayed as a proud, elitist group of beings that consider themselves superior to all other creatures. The fourth book also has a variation on the species called an Alcetaur, which is part man, part moose. The myth of the centaur appears in John Updike's novel "The Centaur". The author depicts a rural Pennsylvanian town as seen through the optics of the myth of the centaur. An unknown and marginalized local school teacher, just like the mythological Chiron did for Prometheus, gave up his life for the future of his son who had chosen to be an independent artist in New York. Other hybrid creatures appear in Greek mythology, always with some liminal connection that links Hellenic culture with archaic or non-Hellenic cultures: Also, Additionally, "Bucentaur", the name of several historically important Venetian vessels, was linked to a posited ox-centaur or "βουκένταυρος" "(boukentauros)" by fanciful and likely spurious folk-etymology.
https://en.wikipedia.org/wiki?curid=7401
Chemotaxis Chemotaxis (from "chemo-" + "taxis") is the movement of an organism in response to a chemical stimulus. Somatic cells, bacteria, and other single-cell or multicellular organisms direct their movements according to certain chemicals in their environment. This is important for bacteria to find food (e.g., glucose) by swimming toward the highest concentration of food molecules, or to flee from poisons (e.g., phenol). In multicellular organisms, chemotaxis is critical to early development (e.g., movement of sperm towards the egg during fertilization) and subsequent phases of development (e.g., migration of neurons or lymphocytes) as well as in normal function and health (e.g., migration of leukocytes during injury or infection). In addition, it has been recognized that mechanisms that allow chemotaxis in animals can be subverted during cancer metastasis. The aberrant chemotaxis of leukocytes and lymphocytes also contribute to inflammatory diseases such as atherosclerosis, asthma, and arthritis. "Positive" chemotaxis occurs if the movement is toward a higher concentration of the chemical in question; "negative" chemotaxis if the movement is in the opposite direction. Chemically prompted kinesis (randomly directed or nondirectional) can be called chemokinesis. Although migration of cells was detected from the early days of the development of microscopy by Leeuwenhoek, a Caltech lecture regarding chemotaxis propounds that 'erudite description of chemotaxis was only first made by T. W. Engelmann (1881) and W. F. Pfeffer (1884) in bacteria, and H. S. Jennings (1906) in ciliates'. The Nobel Prize laureate I. Metchnikoff also contributed to the study of the field during 1882 to 1886, with investigations of the process as an initial step of phagocytosis. The significance of chemotaxis in biology and clinical pathology was widely accepted in the 1930s, and the most fundamental definitions underlying the phenomenon were drafted by this time. The most important aspects in quality control of chemotaxis assays were described by H. Harris in the 1950s. In the 1960s and 1970s, the revolution of modern cell biology and biochemistry provided a series of novel techniques that became available to investigate the migratory responder cells and subcellular fractions responsible for chemotactic activity. The availability of this technology led to the discovery of C5a, a major chemotactic factor involved in acute inflammation. The pioneering works of J. Adler modernized Pfeffer's capillary assay and represented a significant turning point in understanding the whole process of intracellular signal transduction of bacteria. Some bacteria, such as "E. coli", have several flagella per cell (4–10 typically). These can rotate in two ways: The directions of rotation are given for an observer outside the cell looking down the flagella toward the cell. The overall movement of a bacterium is the result of alternating tumble and swim phases. As a result, the trajectory of a bacterium swimming in a uniform environment will form a random walk with relatively straight swims interrupted by random tumbles that reorient the bacterium. Bacteria such as "E. coli" are unable to choose the direction in which they swim, and are unable to swim in a straight line for more than a few seconds due to rotational diffusion; in other words, bacteria "forget" the direction in which they are going. By repeatedly evaluating their course, and adjusting if they are moving in the wrong direction, bacteria can direct their random walk motion toward favorable locations. In the presence of a chemical gradient bacteria will chemotax, or direct their overall motion based on the gradient. If the bacterium senses that it is moving in the correct direction (toward attractant/away from repellent), it will keep swimming in a straight line for a longer time before tumbling; however, if it is moving in the wrong direction, it will tumble sooner and try a new direction at random. In other words, bacteria like "E. coli" use temporal sensing to decide whether their situation is improving or not, and in this way, find the location with the highest concentration of attractant (usually the source) quite well. Even under very high concentrations, it can still distinguish very small differences in concentration, and fleeing from a repellent works with the same efficiency. This biased random walk is a result of simply choosing between two methods of random movement; namely tumbling and straight swimming. The helical nature of the individual flagellar filament is critical for this movement to occur. The protein structure that makes up the flagellar filament, flagellin, is conserved among all flagellated bacteria. Vertebrates seem to have taken advantage of this fact by possessing an immune receptor (TLR5) designed to recognize this conserved protein. As in many instances in biology, there are bacteria that do not follow this rule. Many bacteria, such as "Vibrio", are monoflagellated and have a single flagellum at one pole of the cell. Their method of chemotaxis is different. Others possess a single flagellum that is kept inside the cell wall. These bacteria move by spinning the whole cell, which is shaped like a corkscrew. Chemical gradients are sensed through multiple transmembrane receptors, called methyl-accepting chemotaxis proteins (MCPs), which vary in the molecules that they detect. Thousands of MCP receptors are known to be encoded across the bacterial kingdom. These receptors may bind attractants or repellents directly or indirectly through interaction with proteins of periplasmatic space. The signals from these receptors are transmitted across the plasma membrane into the cytosol, where "Che proteins" are activated. The Che proteins alter the tumbling frequency, and alter the receptors. The proteins CheW and CheA bind to the receptor. The absence of receptor activation results in autophosphorylation in the histidine kinase, CheA, at a single highly conserved histidine residue. CheA, in turn, transfers phosphoryl groups to conserved aspartate residues in the response regulators CheB and CheY; CheA is a histidine kinase and it does not actively transfer the phosphoryl group, rather, the response regulator CheB takes the phosphoryl group from CheA. This mechanism of signal transduction is called a two-component system, and it is a common form of signal transduction in bacteria. CheY induces tumbling by interacting with the flagellar switch protein FliM, inducing a change from counter-clockwise to clockwise rotation of the flagellum. Change in the rotation state of a single flagellum can disrupt the entire flagella bundle and cause a tumble. CheB, when activated by CheA, acts as a methylesterase, removing methyl groups from glutamate residues on the cytosolic side of the receptor; it works antagonistically with CheR, a methyltransferase, which adds methyl residues to the same glutamate residues. If the level of an attractant remains high, the level of phosphorylation of CheA (and, therefore, CheY and CheB) will remain low, the cell will swim smoothly, and the level of methylation of the MCPs will increase (because CheB-P is not present to demethylate). The MCPs no longer respond to the attractant when they are fully methylated; therefore, even though the level of attractant might remain high, the level of CheA-P (and CheB-P) increases and the cell begins to tumble. The MCPs can be demethylated by CheB-P, and, when this happens, the receptors can once again respond to attractants. The situation is the opposite with regard to repellents: fully methylated MCPs respond best to repellents, while least-methylated MCPs respond worst to repellents. This regulation allows the bacterium to 'remember' chemical concentrations from the recent past, a few seconds, and compare them to those it is currently experiencing, thus 'know' whether it is traveling up or down a gradient. Although the methylation system accounts for the wide range of sensitivity that bacteria have to chemical gradients, other mechanisms are involved in increasing the absolute value of the sensitivity on a given background. Well-established examples are the ultra-sensitive response of the motor to the CheY-P signal, and the clustering of chemoreceptors. Chemoattractants and chemorepellents are inorganic or organic substances possessing chemotaxis-inducer effect in motile cells. These chemotactic ligands create chemical concentration gradients that organisms, prokaryotic and eukaryotic, move toward or away from, respectively. Effects of chemoattractants are elicited via chemoreceptors such as methyl-accepting chemotaxis proteins (MCP). MCPs in E.coli include Tar, Tsr, Trg and Tap. Chemoattracttants to Trg include ribose and galactose with phenol as a chemorepellent. Tap and Tsr recognize dipeptides and serine as chemoattractants, respectively. Chemoattractants or chemorepellents bind MCPs at its extracellular domain; an intracellular signaling domain relays the changes in concentration of these chemotactic ligands to downstream proteins like that of CheA which then relays this signal to flagellar motors via phosphorylated CheY (CheY-P). CheY-P can then control flagellar rotation influencing the direction of cell motility. For "E.coli", "S. meliloti", and "R. spheroids," the binding of chemoattractants to MCPs inhibit CheA and therefore CheY-P activity, resulting in smooth runs, but for "B. substilis", CheA activity increases. Methylation events in "E.coli" cause MCPs to have lower affinity to chemoattractants which causes increased activity of CheA and CheY-P resulting in tumbles. In this way cells are able to adapt to the immediate chemoattractant concentration and detect further changes to modulate cell motility. Chemoattractants in eukaryotes are well characterized for immune cells. Formyl peptides, such as fMLF, attract leukocytes such as neutrophils and macrophages, causing movement toward infection sites. Non-acylated methioninyl peptides do not act as chemoattractants to neutrophils and macrophages. Leukocytes also move toward chemoattractants C5a, a complement component, and pathogen-specific ligands on bacteria. Mechanisms concerning chemorepellents are less known than chemoattractants. Although chemorepellents work to confer an avoidance response in organisms, "Tetrahymena thermophila" adapt to a chemorepellent, Netrin-1 peptide, within 10 minutes of exposure; however, exposure to chemorepellents such as GTP, PACAP-38, and nociceptin show no such adaptations. GTP and ATP are chemorepellents in micro-molar concentrations to both "Tetrahymena" and "Paramecium". These organisms avoid these molecules by producing avoiding reactions to re-orient themselves away from the gradient. The mechanism of chemotaxis that eukaryotic cells employ is quite different from that in bacteria; however, sensing of chemical gradients is still a crucial step in the process. Due to their small size, prokaryotes cannot directly detect a concentration gradient. Instead, prokaryotes sense their environments temporally, constantly swimming and redirecting themselves each time they sense a change in the gradient. Eukaryotic cells are much larger than prokaryotes and have receptors embedded uniformly throughout the cell membrane. Eukaryotic chemotaxis involves detecting a concentration gradient spatially by comparing the asymmetric activation of these receptors at the different ends of the cell. Activation of these receptors results in migration towards chemoattractants, or away from chemorepellants. It has also been shown that both prokaryotic and eukaryotic cells are capable of chemotactic memory. In prokaryotes, this mechanism involves the methylation of receptors called methyl-accepting chemotaxis proteins (MCPs). This results in their desensitization and allows prokaryotes to "remember" and adapt to a chemical gradient. In contrast, chemotactic memory in eukaryotes can be explained by the Local Excitation Global Inhibition (LEGI) model. LEGI involves the balance between a fast excitation and delayed inhibition which controls downstream signaling such as Ras activation and PIP3 production. Levels of receptors, intracellular signalling pathways and the effector mechanisms all represent diverse, eukaryotic-type components. In eukaryotic unicellular cells, amoeboid movement and cilium or the eukaryotic flagellum are the main effectors (e.g., Amoeba or Tetrahymena). Some eukaryotic cells of higher vertebrate origin, such as immune cells also move to where they need to be. Besides immune competent cells (granulocyte, monocyte, lymphocyte) a large group of cells—considered previously to be fixed into tissues—are also motile in special physiological (e.g., mast cell, fibroblast, endothelial cells) or pathological conditions (e.g., metastases). Chemotaxis has high significance in the early phases of embryogenesis as development of germ layers is guided by gradients of signal molecules. Unlike motility in bacterial chemotaxis, the mechanism by which eukaryotic cells physically move is unclear. There appear to be mechanisms by which an external chemotactic gradient is sensed and turned into an intracellular PIP3 gradient, which results in a gradient and the activation of a signaling pathway, culminating in the polymerisation of actin filaments. The growing distal end of actin filaments develops connections with the internal surface of the plasma membrane via different sets of peptides and results in the formation of anteriorpseudopods and posterior uropods. Cilia of eukaryotic cells can also produce chemotaxis; in this case, it is mainly a Ca2+-dependent induction of the microtubular system of the basal body and the beat of the 9 + 2 microtubules within cilia. The orchestrated beating of hundreds of cilia is synchronized by a submembranous system built between basal bodies. The details of the signaling pathways are still not totally clear. Chemotaxis refers to the directional migration of cells in response to chemical gradients; several variations of chemical-induced migration exist as listed below. In general, eukaryotic cells sense the presence of chemotactic stimuli through the use of 7-transmembrane (or serpentine) heterotrimeric G-protein-coupled receptors, a class representing a significant portion of the genome. Some members of this gene superfamily are used in eyesight (rhodopsins) as well as in olfaction (smelling). The main classes of chemotaxis receptors are triggered by: However, induction of a wide set of membrane receptors (e.g., cyclic nucleotides, amino acids, insulin, vasoactive peptides) also elicit migration of the cell. While some chemotaxis receptors are expressed in the surface membrane with long-term characteristics, as they are determined genetically, others have short-term dynamics, as they are assembled "ad hoc" in the presence of the ligand. The diverse features of the chemotaxis receptors and ligands allows for the possibility of selecting chemotactic responder cells with a simple chemotaxis assay. By chemotactic selection, we can determine whether a still-uncharacterized molecule acts via the long- or the short-term receptor pathway. The term "chemotactic selection" is also used to designate a technique that separates eukaryotic or prokaryotic cells according to their chemotactic responsiveness to selector ligands. The number of molecules capable of eliciting chemotactic responses is relatively high, and we can distinguish primary and secondary chemotactic molecules. The main groups of the primary ligands are as follows: Chemotactic responses elicited by the ligand-receptor interactions are, in general, distinguished upon the optimal effective concentration(s) of the ligand. Nevertheless, correlation of the amplitude elicited and ratio of the responder cells compared to the total number are also characteristic features of the chemotactic signaling. Investigations of ligand families (e.g., amino acids or oligo peptides) proved that there is a fitting of ranges (amplitudes; number of responder cells) and chemotactic activities: Chemoattractant moiety is accompanied by wide ranges, whereas chemorepellent character by narrow ranges. A changed migratory potential of cells has relatively high importance in the development of several clinical symptoms and syndromes. Altered chemotactic activity of extracellular (e.g., Escherichia coli) or intracellular (e.g., Listeria monocytogenes) pathogens itself represents a significant clinical target. Modification of endogenous chemotactic ability of these microorganisms by pharmaceutical agents can decrease or inhibit the ratio of infections or spreading of infectious diseases. Apart from infections, there are some other diseases wherein impaired chemotaxis is the primary etiological factor, as in Chédiak–Higashi syndrome, where giant intracellular vesicles inhibit normal migration of cells. Several mathematical models of chemotaxis were developed depending on the type of Although interactions of the factors listed above make the behavior of the solutions of mathematical models of chemotaxis rather complex, it is possible to describe the basic phenomenon of chemotaxis-driven motion in a straightforward way. Indeed, let us denote with formula_1 the spatially non-uniform concentration of the chemo-attractant and with formula_2 its gradient. Then the chemotactic cellular flow (also called current) formula_3 that is generated by the chemotaxis is linked to the above gradient by the law: formula_4, where formula_5 is the spatial density of the cells and formula_6 is the so-called ’Chemotactic coefficient’. However, note that in many cases formula_6 is not constant: It is, instead, a decreasing function of the concentration of the chemo-attractant formula_8: formula_9. Spatial ecology of soil microorganisms is a function of their chemotactic sensitivities towards substrate and fellow organisms. The chemotactic behavior of the bacteria was proven to lead to non-trivial population patterns even in the absence of environmental heterogeneities. The presence of structural pore scale heterogeneities has an extra impact on the emerging bacterial patterns. A wide range of techniques is available to evaluate chemotactic activity of cells or the chemoattractant and chemorepellent character of ligands. The basic requirements of the measurement are as follows: Despite the fact that an ideal chemotaxis assay is still not available, there are several protocols and pieces of equipment that offer good correspondence with the conditions described above. The most commonly used are summarised in the table below: "Chemical robots" that use artificial chemotaxis to navigate autonomously have been designed. Applications include targeted delivery of drugs in the body. More recently, enzyme molecules have also shown positive chemotactic behavior in the gradient of their substrates. The thermodynamically-favorable binding of enzymes to their specific substrates is recognized as the origin of enzymatic chemotaxis. Additionally, enzymes in cascades have also shown substrate-driven chemotactic aggregation. Apart from active enzymes, non-reacting molecules also show chemotactic behavior. This has been demonstrated by using dye molecules that move directionally in gradients of polymer solution through favorable hydrophobic interactions.
https://en.wikipedia.org/wiki?curid=7403
Cheshire Cheshire ( ; archaically the County Palatine of Chester) is a county in North West England, bordering Merseyside and Greater Manchester to the north, Derbyshire to the east, Staffordshire and Shropshire to the south, and Flintshire and Wrexham County Borough in Wales to the west. Cheshire's county town is the City of Chester (118,200); the largest town is Warrington (209,700). Other major towns include Crewe (71,722), Runcorn (61,789), Widnes (61,464), Ellesmere Port (55,715), Macclesfield (52,044), Winsford (32,610) and Northwich (19,924). The county covers and has a population of around 1 million. It is mostly rural, with a number of small towns and villages supporting the agricultural and other industries which produce Cheshire cheese, salt, chemicals and silk. Cheshire's name was originally derived from an early name for Chester, and was first recorded as "Legeceasterscir" in the "Anglo-Saxon Chronicle", meaning "the shire of the city of legions". Although the name first appears in 980, it is thought that the county was created by Edward the Elder around 920. In the Domesday Book, Chester was recorded as having the name "Cestrescir" (Chestershire), derived from the name for Chester at the time. A series of changes that occurred as English itself changed, together with some simplifications and elision, resulted in the name Cheshire, as it occurs today. Because of the historically close links with the land bordering Cheshire to the west, which became modern Wales, there is a history of interaction between Cheshire and North Wales. The Domesday Book records Cheshire as having two complete Hundreds (Atiscross and Exestan) that later became the principal part of Flintshire. Additionally, another large portion of the Duddestan Hundred later became known as Maelor Saesneg when it was transferred to North Wales. For this and other reasons, the Welsh language name for Cheshire ("Swydd Gaerlleon") is sometimes used. After the Norman conquest of 1066 by William I, dissent and resistance continued for many years after the invasion. In 1069 local resistance in Cheshire was finally put down using draconian measures as part of the Harrying of the North. The ferocity of the campaign against the English populace was enough to end all future resistance. Examples were made of major landowners such as Earl Edwin of Mercia, their properties confiscated and redistributed amongst Norman barons. William I made Cheshire a county palatine and gave Gerbod the Fleming the new title of Earl of Chester. When Gerbod returned to Normandy in about 1070, the king used his absence to declare the earldom forfeit and gave the title to Hugh d'Avranches (nicknamed Hugh Lupus, or "wolf"). Because of Cheshire's strategic location on the Welsh Marches, the Earl had complete autonomous powers to rule on behalf of the king in the county palatine. The earldom was sufficiently independent from the kingdom of England that the 13th-century Magna Carta did not apply to the shire of Chester, so the earl wrote up his own Chester Charter at the petition of his barons. Cheshire in the Domesday Book (1086) is recorded as a much larger county than it is today. It included two hundreds, Atiscross and Exestan, that later became part of North Wales. At the time of the Domesday Book, it also included as part of Duddestan Hundred the area of land later known as English Maelor (which used to be a detached part of Flintshire) in Wales. The area between the Mersey and Ribble (referred to in the Domesday Book as "Inter Ripam et Mersam") formed part of the returns for Cheshire. Although this has been interpreted to mean that at that time south Lancashire was part of Cheshire, more exhaustive research indicates that the boundary between Cheshire and what was to become Lancashire remained the River Mersey. With minor variations in spelling across sources, the complete list of hundreds of Cheshire at this time are: Atiscross, Bochelau, Chester, Dudestan, Exestan, Hamestan, Middlewich, Riseton, Roelau, Tunendune, Warmundestrou and Wilaveston. There were 8 feudal baronies in Chester, the barons of Kinderton, Halton, Malbank, Mold, Shipbrook, Dunham-Massey, and the honour of Chester itself. Feudal baronies or baronies by tenure were granted by the Earl as forms of feudal land tenure within the palatinate in a similar way to which the king granted English feudal baronies within England proper. An example is the barony of Halton. One of Hugh d'Avranche's barons has been identified as Robert Nicholls, Baron of Halton and Montebourg. In 1182, the land north of the Mersey became administered as part of the new county of Lancashire, thus resolving any uncertainty about the county in which the land "Inter Ripam et Mersam" was. Over the years, the ten hundreds consolidated and changed names to leave just seven—Broxton, Bucklow, Eddisbury, Macclesfield, Nantwich, Northwich and Wirral. In 1397 the county had lands in the march of Wales added to its territory, and was promoted to the rank of principality. This was because of the support the men of the county had given to King Richard II, in particular by his standing armed force of about 500 men called the "Cheshire Guard". As a result, the King's title was changed to "King of England and France, Lord of Ireland, and Prince of Chester". No other English county has been honoured in this way, although it lost the distinction on Richard's fall in 1399. Through the Local Government Act 1972, which came into effect on 1 April 1974, some areas in the north became part of the metropolitan counties of Greater Manchester and Merseyside. Stockport (previously a county borough), Altrincham, Hyde, Dukinfield and Stalybridge in the north-east became part of Greater Manchester. Much of the Wirral Peninsula in the north-west, including the county boroughs of Birkenhead and Wallasey, joined Merseyside as the Metropolitan Borough of Wirral. At the same time the Tintwistle Rural District was transferred to Derbyshire. The area of south Lancashire not included within either the Merseyside or Greater Manchester counties, including Widnes and the county borough of Warrington, was added to the new non-metropolitan county of Cheshire. Halton and Warrington became unitary authorities independent of Cheshire County Council on 1 April 1998, but remain part of Cheshire for ceremonial purposes and also for fire and policing. A referendum for a further local government reform connected with an elected regional assembly was planned for 2004, but was abandoned. As part of the local government restructuring in April 2009, Cheshire County Council and the Cheshire districts were abolished and replaced by two new unitary authorities, Cheshire East and Cheshire West and Chester. The existing unitary authorities of Halton and Warrington were not affected by the change. Prehistoric burial grounds have been discovered at The Bridestones near Congleton (Neolithic) and Robin Hood's Tump near Alpraham (Bronze Age). The remains of Iron Age hill forts are found on sandstone ridges at several locations in Cheshire. Examples include Maiden Castle on Bickerton Hill, Helsby Hillfort and Woodhouse Hillfort at Frodsham. The Roman fortress and walls of Chester, perhaps the earliest building works in Cheshire remaining above ground, are constructed from purple-grey sandstone. The distinctive local red sandstone has been used for many monumental and ecclesiastical buildings throughout the county: for example, the medieval Beeston Castle, Chester Cathedral and numerous parish churches. Occasional residential and industrial buildings, such as Helsby railway station (1849), are also in this sandstone. Many surviving buildings from the 15th to 17th centuries are timbered, particularly in the southern part of the county. Notable examples include the moated manor house Little Moreton Hall, dating from around 1450, and many commercial and residential buildings in Chester, Nantwich and surrounding villages. Early brick buildings include Peover Hall near Macclesfield (1585), Tattenhall Hall (pre-1622), and the Pied Bull Hotel in Chester (17th-century). From the 18th century, orange, red or brown brick became the predominant building material used in Cheshire, although earlier buildings are often faced or dressed with stone. Examples from the Victorian period onwards often employ distinctive brick detailing, such as brick patterning and ornate chimney stacks and gables. Notable examples include Arley Hall near Northwich, Willington Hall near Chester (both by Nantwich architect George Latham) and Overleigh Lodge, Chester. From the Victorian era, brick buildings often incorporate timberwork in a mock Tudor style, and this hybrid style has been used in some modern residential developments in the county. Industrial buildings, such as the Macclesfield silk mills (for example, Waters Green New Mill), are also usually in brick. Cheshire covers a boulder clay plain separating the hills of North Wales and the Peak District (the area is also known as the Cheshire Gap). This was formed following the retreat of ice age glaciers which left the area dotted with kettle holes, locally referred to as meres. The bedrock of this region is almost entirely Triassic sandstone, outcrops of which have long been quarried, notably at Runcorn, providing the distinctive red stone for Liverpool Cathedral and Chester Cathedral. The eastern half of the county is Upper Triassic Mercia Mudstone laid down with large salt deposits which were mined for hundreds of years around Winsford. Separating this area from Lower Triassic Sherwood Sandstone to the west is a prominent sandstone ridge known as the Mid Cheshire Ridge. A footpath, the Sandstone Trail, follows this ridge from Frodsham to Whitchurch passing Delamere Forest, Beeston Castle and earlier Iron Age forts. The highest point (county top) of the historic county of Cheshire is Black Hill () near Crowden in the Cheshire Panhandle, a long eastern projection of the county along the northern side of Longdendale, and on the border with the West Riding of Yorkshire (Black Hill is now the highest point in the ceremonial county of West Yorkshire). Within the current ceremonial county and the unitary authority of Cheshire East the highest point is Shining Tor on the Derbyshire/Cheshire border between Macclesfield and Buxton, at above sea level. After Shining Tor, the next highest point in Cheshire is Shutlingsloe, at above sea level. Shutlingshoe lies just to the south of Macclesfield Forest and is sometimes humorously referred to as the "Matterhorn of Cheshire" thanks to its distinctive steep profile. Cheshire contains portions of two green belt areas surrounding the large conurbations of Merseyside and Greater Manchester (North Cheshire Green Belt, part of the North West Green Belt) and Stoke-on-Trent (South Cheshire Green Belt, part of the Stoke-on-Trent Green Belt), these were first drawn up from the 1950s. Contained primarily within Cheshire East and Chester West & Chester, with small portions along the borders of the Halton and Warrington districts, towns and cities such as Chester, Macclesfield, Alsager, Congleton, Northwich, Ellesmere Port, Knutsford, Warrington, Poynton, Disley, Neston, Wilmslow, Runcorn, and Widnes are either surrounded wholly, partially enveloped by, or on the fringes of the belts. The North Cheshire Green Belt is contiguous with the Peak District Park boundary inside Cheshire. Based on the Census of 2001, the overall population of Cheshire East and Cheshire West and Chester is 673,781, of which 51.3% of the population were male and 48.7% were female. Of those aged between 0–14 years, 51.5% were male and 48.4% were female; and of those aged over 75 years, 62.9% were female and 37.1% were male. This increased to 699,735 at the 2011 Census. The population for 2021 is forecast to be 708,000. In 2001, the population density of Cheshire East and Cheshire West and Chester was 32 people per km2, lower than the North West average of 42 people/km2 and the England and Wales average of 38 people/km2. Ellesmere Port and Neston had a greater urban density than the rest of the county with 92 people/km2. In 2001, ethnic white groups accounted for 98% (662,794) of the population, and 10,994 (2%) in ethnic groups other than white. Of the 2% in non-white ethnic groups: File:Cheshire unitary labell.png|300px|thumb|The ceremonial county showing the four unitary authorities. Click on the map for more information poly 132 343 74 237 19 152 25 129 215 134 231 120 255 121 298 172 293 195 263 221 221 222 179 257 152 293 175 350 Cheshire West and Chester poly 168 293 241 241 285 239 325 225 382 264 383 272 316 308 312 348 225 368 Cheshire East poly 333 175 269 130 267 103 300 82 357 102 480 90 491 189 410 239 Cheshire East poly 313 228 334 183 419 219 354 293 Cheshire East poly 131 105 166 58 182 73 182 94 201 99 207 91 218 105 210 129 183 127 161 121 131 107 Halton poly 188 75 192 30 266 9 296 73 267 82 251 107 234 112 Warrington desc none Cheshire has no county-wide elected local council, but it does have a Lord Lieutenant under the Lieutenancies Act 1997 and a High Sheriff under the Sheriffs Act 1887. Local government functions apart from the Police and Fire/Rescue services are carried out by four smaller unitary authorities: Cheshire East, Cheshire West and Chester, Halton, and Warrington. All four unitary authority areas have borough status. Policing and fire and rescue services are still provided across the county as a whole. The Cheshire Fire Authority consist of members of the four councils, while governance of Cheshire Constabulary is performed by the elected Cheshire Police and Crime Commissioner. Winsford is a major administrative hub for Cheshire with the Police and Fire & Rescue Headquarters based in the town as well as a majority of Cheshire West and Chester Council. It was also home to the former Vale Royal Borough Council and Cheshire County Council. From 1 April 1974 the area under the control of the county council was divided into eight local government districts; Chester, Congleton, Crewe and Nantwich, Ellesmere Port and Neston, Halton, Macclesfield, Vale Royal and Warrington. Halton (which includes the towns of Runcorn and Widnes) and Warrington became unitary authorities in 1998. The remaining districts and the county were abolished as part of local government restructuring on 1 April 2009. The Halton and Warrington boroughs were not affected by the 2009 restructuring. On 25 July 2007, the Secretary of State Hazel Blears announced she was 'minded' to split Cheshire into two new unitary authorities, Cheshire West and Chester, and Cheshire East. She confirmed she had not changed her mind on 19 December 2007 and therefore the proposal to split two-tier Cheshire into two would proceed. Cheshire County Council leader Paul Findlow, who attempted High Court legal action against the proposal, claimed that splitting Cheshire would only disrupt excellent services while increasing living costs for all. A widespread sentiment that this decision was taken by the European Union long ago has often been portrayed via angered letters from Cheshire residents to local papers. On 31 January 2008 "The Standard", Cheshire and district's newspaper, announced that the legal action had been dropped. Members against the proposal were advised that they may be unable to persuade the court that the decision of Hazel Blears was "manifestly absurd". The Cheshire West and Chester unitary authority covers the area formerly occupied by the City of Chester and the boroughs of Ellesmere Port and Neston and Vale Royal; Cheshire East now covers the area formerly occupied by the boroughs of Congleton, Crewe and Nantwich, and Macclesfield. The changes were implemented on 1 April 2009. Congleton Borough Council pursued an appeal against the judicial review it lost in October 2007. The appeal was dismissed on 4 March 2008. The ceremonial county borders Merseyside, Greater Manchester, Derbyshire, Staffordshire and Shropshire in England along with Flintshire and Wrexham in Wales, arranged by compass directions as shown in the table. below. Cheshire also forms part of the North West England region. In the 2001 Census, 81% of the population (542,413) identified themselves as Christian; 124,677 (19%) did not identify with any religion or did not answer the question; 5,665 (1%) identified themselves as belonging to other major world religions; and 1,033 belonged to other religions. The boundary of the Church of England Diocese of Chester follows most closely the pre-1974 county boundary of Cheshire, so it includes all of Wirral, Stockport, and the Cheshire panhandle that included Tintwistle Rural District council area. In terms of Roman Catholic church administration, most of Cheshire falls into the Roman Catholic Diocese of Shrewsbury. Cheshire has a diverse economy with significant sectors including agriculture, automotive, bio-technology, chemical, financial services, food and drink, ICT, and tourism. The county is famous for the production of Cheshire cheese, salt and silk. The county has seen a number of inventions and firsts in its history. A mainly rural county, Cheshire has a high concentration of villages. Agriculture is generally based on the dairy trade, and cattle are the predominant livestock. Land use given to agriculture has fluctuated somewhat, and in 2005 totalled 1558 km2 over 4,609 holdings. Based on holdings by EC farm type in 2005, 8.51 km2 was allocated to dairy farming, with another 11.78 km2 allocated to cattle and sheep. The chemical industry in Cheshire was founded in Roman times, with the mining of salt in Middlewich and Northwich. Salt is still mined in the area by British Salt. The salt mining has led to a continued chemical industry around Northwich, with Brunner Mond based in the town. Other chemical companies, including Ineos (formerly ICI), have plants at Runcorn. The Essar Refinery (formerly Shell Stanlow Refinery) is at Ellesmere Port. The oil refinery has operated since 1924 and has a capacity of 12 million tonnes per year. Crewe was once the centre of the British railway industry, and remains a major railway junction. The Crewe railway works, built in 1840, employed 20,000 people at its peak, although the workforce is now less than 1,000. Crewe is also the home of Bentley cars. Also within Cheshire are manufacturing plants for Jaguar and Vauxhall Motors in Ellesmere Port. The county also has an aircraft industry, with the BAE Systems facility at Woodford Aerodrome, part of BAE System's Military Air Solutions division. The facility designed and constructed Avro Lancaster and Avro Vulcan bombers and the Hawker-Siddeley Nimrod. On the Cheshire border with Flintshire is the Broughton aircraft factory, more recently associated with Airbus. Tourism in Cheshire from within the UK and overseas continues to perform strongly. Over 8 million nights of accommodation (both UK and overseas) and over 2.8 million visits to Cheshire were recorded during 2003. At the start of 2003, there were 22,020 VAT-registered enterprises in Cheshire, an increase of 7% since 1998, many in the business services (31.9%) and wholesale/retail (21.7%) sectors. Between 2002 and 2003 the number of businesses grew in four sectors: public administration and other services (6.0%), hotels and restaurants (5.1%), construction (1.7%), and business services (1.0%). The county saw the largest proportional reduction between 2001 and 2002 in employment in the energy and water sector and there was also a significant reduction in the manufacturing sector. The largest growth during this period was in the other services and distribution, hotels and retail sectors. Cheshire is considered to be an affluent county. However, towns such as Crewe and Winsford have significant deprivation. The county's proximity to the cities of Manchester and Liverpool means counter urbanisation is common. Cheshire West has a fairly large proportion of residents who work in Liverpool and Manchester, while the town of Northwich and area of Cheshire East falls more within Manchester's sphere of influence. All four local education authorities in Cheshire operate only comprehensive state school systems. When Altrincham, Sale and Bebington were moved from Cheshire to Trafford and Merseyside in 1974, they took some former Cheshire selective schools. Today, there are three universities based in the county, the University of Chester, the Crewe campus of Manchester Metropolitan University and the Chester campus of The University of Law. Cheshire has two Football League teams: Crewe Alexandra and Macclesfield Town, who both play in League Two. Chester, a phoenix club formed in 2010 after ex-Football League club Chester City was dissolved, competes in the National League North. Northwich Victoria, another ex-League team who were founder members of the Football League Division Two in 1892/1893, now represent Cheshire in the Northern Premier League along with Nantwich Town, Warrington Town and Witton Albion. Warrington Wolves and the Widnes Vikings are the premier Rugby League teams in Cheshire and play in the Super League and Championship respectively. There are also numerous junior clubs in the county, including Chester Gladiators. Cheshire County Cricket Club is one of the clubs that make up the Minor counties of English and Welsh cricket. Cheshire also is represented in the highest level basketball league in the UK, the BBL, by Cheshire Phoenix (formerly Cheshire Jets). Each May, Europe's largest motorcycle event, the Thundersprint, is held in Northwich. The county has been home to many notable sports people and athletes. Many Premier League footballers have lived in Cheshire, including Dean Ashton, Seth Johnson, Michael Owen, Jesse Lingard and Wayne Rooney. Other local athletes include cricketer Ian Botham, marathon runner Paula Radcliffe, oarsman Matt Langridge, hurdler Shirley Strong, sailor Ben Ainslie, cyclist Sarah Storey and mountaineer George Mallory, who died in 1924 on Mount Everest. Cheshire has also produced a military hero in Norman Cyril Jones, a World War I flying ace who won the Distinguished Flying Cross. The county has produced several notable popular musicians, including Gary Barlow (Take That, born and raised in Frodsham), Harry Styles (singer with One Direction, raised in Holmes Chapel), John Mayall (John Mayall & the Bluesbreakers), Ian Astbury (The Cult), Tim Burgess (Charlatans), Ian Curtis (Joy Division) and Hooton Tennis Club. Matthew Healy, lead singer of The 1975, met his three bandmates at Wilmslow High School in Wilmslow. Concert pianist Stephen Hough, singer Thea Gilmore and her producer husband Nigel Stonier also reside in Cheshire. The county has also been home to several writers, including Hall Caine (1853–1931), popular romantic novelist and playwright; Alan Garner; Victorian novelist Elizabeth Gaskell, whose novel "Cranford" features her home town of Knutsford; and most famously Lewis Carroll, born and raised in Daresbury, hence the Cheshire Cat (a fictional cat popularised by Carroll in "Alice's Adventures in Wonderland" and known for its distinctive mischievous grin). Artists from the county include ceramic artist Emma Bossons and sculptor and photographer Andy Goldsworthy. Actors from Cheshire include Tim Curry, originated the role of Dr. Frank-N-Furter in "The Rocky Horror Show"; Daniel Craig, the sixth James Bond; Dame Wendy Hiller; and Lewis McGibbon, best known for his role in "Millions". Local radio stations in the county include Dee 106.3, Capital and Smooth Radio for Chester and West Cheshire, Silk FM for the east of the county, Signal 1 and The Cat 107.9 for the south, Wire FM for Warrington and Wish FM, which covers Widnes. Cheshire is one of the only counties (along with County Durham, Dorset and Rutland) that does not have its own designated BBC Radio station. The majority of the county (south and east) are covered by BBC Radio Stoke, while BBC Radio Merseyside tends to cover the west. The BBC directs readers to Stoke and Staffordshire when Cheshire is selected on their website. The BBC covers the west with BBC Radio Merseyside, the north and east with BBC Radio Manchester and the south with BBC Radio Stoke. There were plans to launch BBC Radio Cheshire, but those were shelved in 2007 after a lower than expected BBC licence fee settlement. The Royal Cheshire Show, an annual agricultural show, has taken place for the last 175 years and includes exhibitions, games and competitions. As part of a 2002 marketing campaign, the plant conservation charity Plantlife chose the cuckooflower as the county flower. Previously, a sheaf of golden wheat was the county emblem, a reference to the Earl of Chester's arms in use from the 12th century. The county is home to some of the most affluent areas of northern England, including Alderley Edge, Wilmslow, Prestbury, Tarporley and Knutsford, named in 2006 as the most expensive place to buy a house in the north of England. The former Cheshire town of Altrincham was in second place. The area is sometimes referred to as The Golden Triangle on account of the area in and around the aforementioned towns and villages. Parts of Cheshire away from the affluent areas, such as those around Blacon, Crewe, Ellesmere Port and Winsford, are among the top 5% most deprived areas in Northern England; the price of an average four-bedroom detached house in Knutsford is around £470,000 compared to £220,000 in Winsford and Ellesmere Port. The cities and towns in Cheshire are: Some settlements which were historically part of the county now fall under the counties of Derbyshire, Merseyside and Greater Manchester: Bus transport in Cheshire is provided by various operators. The major bus operator in the Cheshire area is Arriva North West. Other operators in Cheshire include Stagecoach Chester & Wirral, Halton Transport and Network Warrington. There are also several operators based outside of Cheshire, who either run services wholly within the area or services which start from outside the area. Companies include Arriva Buses Wales, BakerBus, High Peak, First Greater Manchester, D&G bus and Stagecoach Manchester. Some services are run under contract to Cheshire West and Chester, Cheshire East, Borough of Halton and Warrington Councils. The main railway line through the county is the West Coast Main Line. Trains on the main London to Scotland line call at Crewe (in the south of the county) and Warrington Bank Quay (in the north of the county). Trains stop at Crewe and Runcorn on the Liverpool branch of the WCML; Crewe and Macclesfield are each hourly stops on the two Manchester branches. The major interchanges are: In the east of Cheshire, Macclesfield station is served by Avanti West Coast, CrossCountry and Northern, on the Manchester–London line. Services from Manchester to the south coast frequently stop at Macclesfield. Cheshire has of roads, including of the M6, M62, M53 and M56 motorways; there are 23 interchanges and four service areas. The M6 motorway at the Thelwall Viaduct carries 140,000 vehicles every 24 hours. The Cheshire canal system includes several canals originally used to transport the county's industrial products (mostly chemicals). Nowadays they are mainly used for tourist traffic. The Cheshire Ring is formed from the Rochdale, Ashton, Peak Forest, Macclesfield, Trent and Mersey and Bridgewater canals. The Manchester Ship Canal is a wide, stretch of water opened in 1894. It consists of the rivers Irwell and Mersey made navigable to Manchester for seagoing ships leaving the Mersey estuary. The canal passes through the north of the county via Runcorn and Warrington.
https://en.wikipedia.org/wiki?curid=7406
Constitution of Canada The Constitution of Canada () is the supreme law in Canada. It outlines Canada's system of government and the civil and human rights of those who are citizens of Canada and non-citizens in Canada. Its contents are an amalgamation of various codified acts, treaties between the Crown and indigenous peoples (both historical and modern), uncodified traditions and conventions. Canada is one of the oldest constitutional democracies in the world. According to subsection 52(2) of the Constitution Act, 1982, the Canadian Constitution consists of the "Canada Act 1982" (which includes the "Constitution Act, 1982"), acts and orders referred to in its schedule (including in particular the "Constitution Act, 1867", formerly the British North America Act, 1867), and any amendments to these documents. The Supreme Court of Canada has held that the list is not exhaustive and also includes a number of pre-confederation acts and unwritten components as well. See list of Canadian constitutional documents for details. The first semblance of a constitution for Canada was the Royal Proclamation of 1763. The act renamed the northeasterly portion of the former French province of New France as Province of Quebec, roughly coextensive with the southern third of contemporary Quebec. The proclamation, which established an appointed colonial government, was the constitution of Quebec until 1774, when the British parliament passed the "Quebec Act", which expanded the province's boundaries to the Ohio and Mississippi Rivers (one of the grievances listed in the United States Declaration of Independence). Significantly, the Quebec Act also replaced the French criminal law presumption of guilty until proven innocent with the English criminal law presumption of innocent until proven guilty; but the French code or civil law system was retained for non-criminal matters. The Treaty of Paris of 1783 ended the American War of Independence and sent a wave of British loyalist refugees northward to Quebec and Nova Scotia. In 1784, the two provinces were divided; Nova Scotia was split into Nova Scotia, Cape Breton Island (rejoined to Nova Scotia in 1820), Prince Edward Island, and New Brunswick, while Quebec was split into Lower Canada (southern Quebec) and Upper Canada (southern through lower northern Ontario). The winter of 1837–38 saw rebellion in both Canadas, prompting their being rejoined as the Province of Canada in 1841. The British North America Act of 1867 established the Dominion of Canada as a federation of provinces. Initially, on 1 July 1867, four provinces entered into confederation as "One dominion under the name of Canada": Canada West (former Upper Canada, now Ontario), Canada East (former Lower Canada, now Quebec), Nova Scotia, and New Brunswick. Title to the Northwest Territories was transferred by the Hudson's Bay Company in 1870, out of which the province of Manitoba (the first to be established by the Parliament of Canada) was created. British Columbia joined confederation in 1871, followed by Prince Edward Island in 1873. The Yukon Territory was created by Parliament in 1898, followed by Alberta and Saskatchewan in 1905 (all out of parts of the Northwest Territories). Newfoundland, Britain's oldest colony in the Americas and by then also a Dominion, joined Confederation in 1949. Nunavut was created in 1999 from the Northwest Territories. An Imperial Conference in 1926 that included the leaders of all Dominions and representatives from India (which then included Burma, Bangladesh, and Pakistan), led to the eventual enactment of the Statute of Westminster 1931. The statute, an essential transitory step from the British Empire to the Commonwealth of Nations, provided that existing Dominions became fully sovereign of the United Kingdom and any new Dominions would be fully sovereign upon the grant of Dominion status. Although listed, Newfoundland never ratified the statute so was still subject to imperial authority when its entire system of government and economy collapsed in the mid-1930s. Canada did ratify the statute but with a requested exception—the Canadian federal and provincial governments could not agree on an amending formula for the Canadian constitution. It would be another 50 years before this was achieved. In the interim, the British parliament periodically passed enabling acts with respect to requested constitutional amendments as they arose; this was never anything but a rubber stamp. The patriation of the Canadian constitution was achieved in 1982 when the British parliament, with the request and assent of the Canadian parliament, passed the Canada Act 1982, which included in its schedules the Constitution Act, 1982. The United Kingdom thus formally absolving itself of any remaining responsibility for, or jurisdiction over, Canada. In a formal ceremony on Parliament Hill in Ottawa, Queen Elizabeth II proclaimed both acts as law on 17 April 1982. "Constitution Act, 1982", includes the Canadian Charter of Rights and Freedoms. Prior to the Charter, various statutes protected an assortment of civil rights and obligations but nothing was enshrined in the constitution until 1982. The Charter has thus placed a strong focus upon individual and collective rights of the people of Canada. Enactment of the Charter of Rights and Freedoms has fundamentally changed much of Canadian constitutional law. The act also codified many previously oral constitutional conventions and made amendment of the constitution in general significantly more difficult. Previously, the Canadian federal constitution could be amended by a solitary act of the Canadian or British parliament, by formal or informal agreement between the federal and provincial governments, or even simply by adoption as custom of an oral convention or performance that shows precedential but unwritten tradition. Since the act, textual amendments must now conform to certain specified provisions in the written portion of the Canadian constitution. This was an Act of the British parliament, originally called the British North America Act 1867. It outlined Canada's system of government, which combines Britain's Westminster model of parliamentary government with division of sovereignty (federalism). Although it is the first of 20 British North America Acts, it is the most famous as the primary document of Canadian Confederation. With the patriation of the Constitution in 1982, this Act was renamed "Constitution Act, 1867". In recent years, the 1867 document has mainly served as the basis on which the division of powers between the provinces and federal government are analyzed. Endorsed by all provincial governments except that of Quebec, this was the formal Act of Parliament that effected Canada's full legislative independence from the United Kingdom. Part V of this act established an amending formula for the Canadian constitution, the lack of which (due to more than 50 years of disagreement between the federal and provincial governments) meant Canada's constitutional amendments still required enactment by the British parliament after Statute of Westminster in 1931. The Act was enacted as a schedule to the Canada Act 1982, a British Act of Parliament which was introduced at the request of a joint address to the Queen by the Senate and House of Commons of Canada. As a bilingual act of parliament, the Canada Act 1982 has the distinction of being the only legislation in French that has been passed by an English or British parliament since Norman French (Law French) ceased to be the language of government in England. In addition to enacting the Constitution Act, 1982, the Canada Act 1982 provides that no further British Acts of Parliament will apply to Canada as part of its law, finalizing Canada's legislative independence. As noted above, this is Part I of the Constitution Act, 1982. The Charter is the constitutional guarantee of the civil rights and liberties of every citizen in Canada, such as freedom of expression, of religion, and of mobility. Part II addresses the rights of Aboriginal peoples in Canada. It is written in plain language to ensure accessibility to the average citizen. It applies only to government and government actions with the intention to prevent government from creating laws that are unconstitutional. Instead of the usual parliamentary procedure, that includes the monarch's formal royal assent for enacting legislation, amendments to the Constitution Act, 1982, must be done in accordance with Part V of the Constitution Act, 1982, which provides for five different amending formulae. Amendments can be brought forward under section 46(1) by any province or the federal legislature. The general formula set out in section 38(1), known as the "7/50 formula", requires: (a) assent from both the House of Commons and the Senate; (b) the approval of two thirds of the provincial legislatures (at least seven provinces) representing at least 50 per cent of the population (effectively, this would include at least Quebec or Ontario, as they are the most populous provinces). This formula specifically applies to amendments related to the proportionate representation in Parliament, powers, selection, and composition of the Senate, the Supreme Court and the addition of provinces or territories. The other amendment formulae are for particular cases as provided by in the act. An amendment related to the Office of the Queen, the use of either official language (subject to section 43), the amending formula itself, or the composition of the Supreme Court, must be adopted by unanimous consent of all the provinces in accordance with section 41. In the case of an amendment related to provincial boundaries or the use of an official language within a province alone, the amendment must be passed by the legislatures affected by the amendment (section 43). In the case of an amendment that affects the federal government only, the amendment does not need approval of the provinces (section 44). The same applies to amendments affecting the provincial government alone (section 45). Canada's constitution has roots going back to the thirteenth century, including England's Magna Carta and the first English parliament of 1275. It is one of the oldest working constitutions in the world (others are: the United Kingdom, United States, Sweden, Norway, Switzerland, Denmark). Canada's constitution is composed of several individual statutes. There are three general methods by which a statute becomes entrenched in the Constitution: The existence of unwritten constitutional components was reaffirmed in 1998 by the Supreme Court in "Reference re Secession of Quebec". The Constitution is more than a written text. It embraces the entire global system of rules and principles which govern the exercise of constitutional authority. A superficial reading of selected provisions of the written constitutional enactment, without more, may be misleading. In practice, there have been three sources of unwritten constitutional law: Unlike in most federations, Canadian provinces do not have written provincial constitutions. Provincial constitutions are instead a combination of uncodified constitution, provisions of the Constitution of Canada, and provincial statutes. Structures of the provincial government are described in parts of the Constitution. Part V of the "Constitution Act, 1867" describes the governmental structure of the original four provinces. The three colonies that joined Canada had existing UK legislation which described their governmental structure, and this was affirmed in the terms of union, which now form part of Canada's Constitution ("British Columbia Terms of Union", the schedules of "Prince Edward Island Terms of Union", and "Terms of Union of Newfoundland with Canada"). The remaining three provinces were created by statute, and their structures are described in those statutes, which now form part of Canada's Constitution ("Manitoba Act, 1870", "Alberta Act", and "Saskatchewan Act"). Section 45 of the "Constitution Act, 1982" allows each province to amend its own constitution. However, if the desired change would require an amendment to any documents that form part of the Constitution of Canada, it would require the consent of the federal government under section 43. This was done, for example, by the "Constitution Amendment, 1998", when Newfoundland asked the federal government to amend the "Terms of Union of Newfoundland" to allow it to end denominational quotas for religion classes. Provinces have used their power under section 45 to enact statutes related to elections and rules for its legislature. For examples of this, see Constitution of Quebec and Constitution Act (British Columbia). However, these acts are regular statutes, and are not enshrined as constitutional documents. They do not, generally, require a referendum or a super-majority of legislators to amend them. However, there are a small number of statutes that cannot be amended by a simple majority of the legislative assembly. For example, section 7 of the "Constitution of Alberta Amendment Act, 1990" requires plebiscites of Metis settlement members before that act can be amended. Courts have not yet ruled about whether this kind of language really would bind future legislatures, but it might do so if the higher bar was met when creating the law. One way that provinces have put some statutes 'above' others is through quasi-constitutionality. This is the practice of allowing some statutes to take precedence over all others. Quasi-constitutionality is often applied to human rights laws, allowing those laws to act as a "de facto" constitutional charter of rights. In 1983, Peter Greyson, an art student, entered Ottawa's National Archives (known today as Library and Archives Canada) and poured red paint mixed with glue over a copy of the proclamation of the 1982 constitutional amendment. He said he was displeased with the federal government's decision to allow United States missile testing in Canada and had wanted to "graphically illustrate to Canadians" how wrong he believed the government to be. Greyson was charged with public mischief and sentenced to 89 days in jail, 100 hours of community work, and two years of probation. A grapefruit-sized stain remains on the original document; restoration specialists opted to leave most of the paint intact, fearing that removal attempts would only cause further damage.
https://en.wikipedia.org/wiki?curid=7411
Crochet Crochet (; ) is a process of creating textiles by using a crochet hook to interlock loops of yarn, thread, or strands of other materials. The name is derived from the French term "crochet", meaning 'small hook'. Hooks can be made from a variety of materials, such as metal, wood, bamboo, or plastic. The key difference between crochet and knitting, beyond the implements used for their production, is that each stitch in crochet is completed before the next one is begun, while knitting keeps many stitches open at a time. Some variant forms of crochet, such as Tunisian crochet and broomstick lace, do keep multiple crochet stitches open at a time. The word crochet is derived from the Old French "crochet", a diminutive of "croche", in turn from the Germanic "croc", both meaning "hook". It was used in 17th-century French lace-making, where the term "crochetage" designated a stitch used to join separate pieces of lace. The word "crochet" subsequently came to describe both the specific type of textile, and the hooked needle used to produce it. Knitted textiles survive from as early as the 11th century CE, but the first substantive evidence of crocheted fabric emerges in Europe during the 19th century. Earlier work identified as crochet was commonly made by nålebinding, a different looped yarn technique. The first known published instructions for crochet explicitly using that term to describe the craft in its present sense appeared in the Dutch magazine "Penélopé" in 1823. This includes a colour plate showing five styles of purse, of which three were intended to be crocheted with silk thread. The first is "simple open crochet" ("crochet simple ajour"), a mesh of chain-stitch arches. The second (illustrated here) starts in a semi-open form ("demi jour"), where chain-stitch arches alternate with equally long segments of slip-stitch crochet, and closes with a star made with "double-crochet stitches" ("dubbelde hekelsteek": double-crochet in British terminology; single-crochet in US). The third purse is made entirely in double-crochet. The instructions prescribe the use of a tambour needle (as illustrated below) and introduce a number of decorative techniques. The earliest dated reference in English to garments made of cloth produced by looping yarn with a hook—"shepherd's knitting"—is in "The Memoirs of a Highland Lady" by Elizabeth Grant (1797–1830). The journal entry, itself, is dated 1812 but was not recorded in its subsequently published form until some time between 1845 and 1867, and the actual date of publication was first in 1898. Nonetheless, the 1833 volume of "Penélopé" describes and illustrates a shepherd's hook, and recommends its use for crochet with coarser yarn. In 1844, one of the numerous books discussing crochet that began to appear in the 1840s states: Two years later, the same author writes: An instruction book from 1846 describes "Shepherd or single crochet" as what in current British usage is either called single crochet or slip-stitch crochet, with U.S. American terminology always using the latter (reserving single crochet for use as noted above). It similarly equates "Double" and "French crochet". Notwithstanding the categorical assertion of a purely British origin, there is solid evidence of a connection between French tambour embroidery and crochet. French tambour embroidery was illustrated in detail in 1763 in Diderot's Encyclopedia. The tip of the needle shown there is indistinguishable from that of a present-day inline crochet hook and the chain stitch separated from a cloth support is a fundamental element of the latter technique. The 1823 "Penélopé" instructions unequivocally state that the tambour tool was used for crochet and the first of the 1840s instruction books uses the terms "tambour" and "crochet" as synonyms. This equivalence is retained in the 4th edition of that work, 1847. The strong taper of the shepherd's hook eases the production of slip-stitch crochet but is less amenable to stitches that require multiple loops on the hook at the same time. Early yarn hooks were also continuously tapered but gradually enough to accommodate multiple loops. The design with a cylindrical shaft that is commonplace today was largely reserved for tambour-style steel needles. Both types gradually merged into the modern form that appeared toward the end of the 19th century, including both tapered and cylindrical segments, and the continuously tapered bone hook remained in industrial production until World War II. The early instruction books make frequent reference to the alternative use of 'ivory, bone, or wooden hooks' and 'steel needles in a handle', as appropriate to the stitch being made. Taken with the synonymous labeling of shepherd's- and single crochet, and the similar equivalence of French- and double crochet, there is a strong suggestion that crochet is rooted both in tambour embroidery and shepherd's knitting, leading to thread and yarn crochet respectively; a distinction that is still made. The locus of the fusion of all these elements—the "invention" noted above—has yet to be determined, as does the origin of shepherd's knitting. Shepherd's hooks are still being made for local slip-stitch crochet traditions. The form in the accompanying photograph is typical for contemporary production. A longer continuously tapering design intermediate between it and the 19th-century tapered hook was also in earlier production, commonly being made from the handles of forks and spoons. In the 19th century, as Ireland was facing the Great Irish Famine (1845–1849), crochet lace work was introduced as a form of famine relief (the production of crocheted lace being an alternative way of making money for impoverished Irish workers). Men, women, children joined a cooperative in order to crochet and produce products to help with famine relief during the Great Irish Famine. Schools to teach crocheting were started. Teachers were trained and sent across Ireland to teach this craft. When the Irish immigrated to the Americas, they were able to take with them crocheting. Mademoiselle Riego de la Blanchardiere is generally credited with the invention of Irish Crochet, publishing the first book of patterns in 1846. Irish lace became popular in Europe and America, and was made in quantity until the first World War. Fashions in crochet changed with the end of the Victorian era in the 1890s. Crocheted laces in the new Edwardian era, peaking between 1910 and 1920, became even more elaborate in texture and complicated stitching. The strong Victorian colours disappeared, though, and new publications called for white or pale threads, except for fancy purses, which were often crocheted of brightly colored silk and elaborately beaded. After World War I, far fewer crochet patterns were published, and most of them were simplified versions of the early 20th-century patterns. After World War II, from the late 1940s until the early 1960s, there was a resurgence in interest in home crafts, particularly in the United States, with many new and imaginative crochet designs published for colorful doilies, potholders, and other home items, along with updates of earlier publications. These patterns called for thicker threads and yarns than in earlier patterns and included wonderful variegated colors. The craft remained primarily a homemaker's art until the late 1960s and early 1970s, when the new generation picked up on crochet and popularized granny squares, a motif worked in the round and incorporating bright colors. Although crochet underwent a subsequent decline in popularity, the early 21st century has seen a revival of interest in handcrafts and DIY, as well as great strides in improvement of the quality and varieties of yarn. There are many more new pattern books with modern patterns being printed, and most yarn stores now offer crochet lessons in addition to the traditional knitting lessons. There are many books you can purchase from local book stores to teach yourself how to crochet whether it be as a beginner or intermediate. There are also many books for children and teenagers who are hoping to take up the hobby. Filet crochet, Tunisian crochet, tapestry crochet, broomstick lace, hairpin lace, cro-hooking, and Irish crochet are all variants of the basic crochet method. Crochet has experienced a revival on the catwalk as well. Christopher Kane's Fall 2011 Ready-to-Wear collection makes intensive use of the granny square, one of the most basic of crochet motifs. In addition, crochet has been utilized many times by designers on the popular reality show "Project Runway". Websites such as Etsy and Ravelry have made it easier for individual hobbyists to sell and distribute their patterns or projects across the internet. Laneya Wiles released a music video titled "Straight Hookin'" which makes a play on the word "hookers," which has a double meaning for both "one who crochets" and "a prostitute." Basic materials required for crochet are a hook and some type of material that will be crocheted, most commonly yarn or thread. Yarn, one of the most commonly used materials for crocheting, has varying weights which need to be taken into consideration when following patterns. Additional tools are convenient for keeping stitches counted, measuring crocheted fabric, or making related accessories. Examples include cardboard cutouts, which can be used to make tassels, fringe, and many other items; a pom-pom circle, used to make pom-poms; a tape measure and a gauge measure, both used for measuring crocheted work and counting stitches; a row counter; and occasionally plastic rings, which are used for special projects. In recent years, yarn selections have moved beyond synthetic and plant and animal-based fibers to include bamboo, qiviut, hemp, and banana stalks, to name a few. Many advanced crocheters have also incorporated recycled materials into their work in an effort to "go green" and experiment with new textures by using items such as plastic bags, old t-shirts or sheets, VCR or Cassette tape, and ribbon. The crochet hook comes in many sizes and materials, such as bone, bamboo, aluminium, plastic, and steel. Because sizing is categorized by the diameter of the hook's shaft, a crafter aims to create stitches of a certain size in order to reach a particular gauge specified in a given pattern. If gauge is not reached with one hook, another is used until the stitches made are the needed size. Crafters may have a preference for one type of hook material over another due to aesthetic appeal, yarn glide, or hand disorders such as arthritis, where bamboo or wood hooks are favored over metal for the perceived warmth and flexibility during use. Hook grips and ergonomic hook handles are also available to assist crafters. Steel crochet hooks range in size from 0.4 to 3.5 millimeters, or from 00 to 16 in American sizing. These hooks are used for fine crochet work such as doilies and lace. Aluminium, bamboo, and plastic crochet hooks are available from 2.5 to 19 millimeters in size, or from B to S in American sizing. Artisan-made hooks are often made of hand-turned woods, sometimes decorated with semi-precious stones or beads. Crochet hooks used for Tunisian crochet are elongated and have a stopper at the end of the handle, while double-ended crochet hooks have a hook on both ends of the handle. There is also a double hooked apparatus called a Cro-hook that has become popular. A hairpin loom is often used to create lacy and long stitches, known as hairpin lace. While this is not in itself a hook, it is a device used in conjunction with a crochet hook to produce stitches. See : List of United States standard crochet hook and knitting needle sizes Yarn for crochet is usually sold as balls or skeins (hanks), although it may also be wound on spools or cones. Skeins and balls are generally sold with a "yarn band", a label that describes the yarn's weight, length, dye lot, fiber content, washing instructions, suggested needle size, likely gauge, etc. It is a common practice to save the yarn band for future reference, especially if additional skeins must be purchased. Crocheters generally ensure that the yarn for a project comes from a single dye lot. The dye lot specifies a group of skeins that were dyed together and thus have precisely the same color; skeins from different dye lots, even if very similar in color, are usually slightly different and may produce a visible stripe when added onto existing work. If insufficient yarn of a single dye lot is bought to complete a project, additional skeins of the same dye lot can sometimes be obtained from other yarn stores or online. The thickness or weight of the yarn is a significant factor in determining how many stitches and rows are required to cover a given area for a given stitch pattern. This is also termed the gauge. Thicker yarns generally require large-diameter crochet hooks, whereas thinner yarns may be crocheted with thick or thin hooks. Hence, thicker yarns generally require fewer stitches, and therefore less time, to work up a given project. The recommended gauge for a given ball of yarn can be found on the label that surrounds the skein when buying in stores. Patterns and motifs are coarser with thicker yarns and produce bold visual effects, whereas thinner yarns are best for refined or delicate pattern-work. Yarns are standardly grouped by thickness into six categories: superfine, fine, light, medium, bulky and superbulky. Quantitatively, thickness is measured by the number of wraps per inch (WPI). The related "weight per unit length" is usually measured in tex or denier. Before use, hanks are wound into balls in which the yarn emerges from the center, making crocheting easier by preventing the yarn from becoming easily tangled. The winding process may be performed by hand or done with a ballwinder and swift. A yarn's usefulness is judged by several factors, such as its "loft" (its ability to trap air), its "resilience" (elasticity under tension), its washability and colorfastness, its "hand" (its feel, particularly softness vs. scratchiness), its durability against abrasion, its resistance to pilling, its "hairiness" (fuzziness), its tendency to twist or untwist, its overall weight and drape, its blocking and felting qualities, its comfort (breathability, moisture absorption, wicking properties) and its appearance, which includes its color, sheen, smoothness and ornamental features. Other factors include allergenicity, speed of drying, resistance to chemicals, moths, and mildew, melting point and flammability, retention of static electricity, and the propensity to accept dyes. Desirable properties may vary for different projects, so there is no one "best" yarn. Although crochet may be done with ribbons, metal wire or more exotic filaments, most yarns are made by spinning fibers. In spinning, the fibers are twisted so that the yarn resists breaking under tension; the twisting may be done in either direction, resulting in a Z-twist or S-twist yarn. If the fibers are first aligned by combing them and the spinner uses a worsted type drafting method such as the short forward draw, the yarn is smoother and called a "worsted"; by contrast, if the fibers are carded but not combed and the spinner uses a woolen drafting method such as the long backward draw, the yarn is fuzzier and called "woolen-spun". The fibers making up a yarn may be continuous "filament" fibers such as silk and many synthetics, or they may be "staples" (fibers of an average length, typically a few inches); naturally filament fibers are sometimes cut up into staples before spinning. The strength of the spun yarn against breaking is determined by the amount of twist, the length of the fibers and the thickness of the yarn. In general, yarns become stronger with more twist (also called "worst"), longer fibers and thicker yarns (more fibers); for example, thinner yarns require more twist than do thicker yarns to resist breaking under tension. The thickness of the yarn may vary along its length; a "slub" is a much thicker section in which a mass of fibers is incorporated into the yarn. The spun fibers are generally divided into animal fibers, plant and synthetic fibers. These fiber types are chemically different, corresponding to proteins, carbohydrates and synthetic polymers, respectively. Animal fibers include silk, but generally are long hairs of animals such as sheep (wool), goat (angora, or cashmere goat), rabbit (angora), llama, alpaca, dog, cat, camel, yak, and muskox (qiviut). Plants used for fibers include cotton, flax (for linen), bamboo, ramie, hemp, jute, nettle, raffia, yucca, coconut husk, banana trees, soy and corn. Rayon and acetate fibers are also produced from cellulose mainly derived from trees. Common synthetic fibers include acrylics, polyesters such as dacron and ingeo, nylon and other polyamides, and olefins such as polypropylene. Of these types, wool is generally favored for crochet, chiefly owing to its superior elasticity, warmth and (sometimes) felting; however, wool is generally less convenient to clean and some people are allergic to it. It is also common to blend different fibers in the yarn, e.g., 85% alpaca and 15% silk. Even within a type of fiber, there can be great variety in the length and thickness of the fibers; for example, Merino wool and Egyptian cotton are favored because they produce exceptionally long, thin (fine) fibers for their type. A single spun yarn may be crochet as is, or braided or plied with another. In plying, two or more yarns are spun together, almost always in the opposite sense from which they were spun individually; for example, two Z-twist yarns are usually plied with an S-twist. The opposing twist relieves some of the yarns' tendency to curl up and produces a thicker, "balanced" yarn. Plied yarns may themselves be plied together, producing "cabled yarns" or "multi-stranded yarns". Sometimes, the yarns being plied are fed at different rates, so that one yarn loops around the other, as in bouclé. The single yarns may be dyed separately before plying, or afterwards to give the yarn a uniform look. The dyeing of yarns is a complex art. Yarns need not be dyed; or they may be dyed one color, or a great variety of colors. Dyeing may be done industrially, by hand or even hand-painted onto the yarn. A great variety of synthetic dyes have been developed since the synthesis of indigo dye in the mid-19th century; however, natural dyes are also possible, although they are generally less brilliant. The color-scheme of a yarn is sometimes called its colorway. Variegated yarns can produce interesting visual effects, such as diagonal stripes. Crocheted fabric is begun by placing a slip-knot loop on the hook (though other methods, such as a magic ring or simple folding over of the yarn may be used), pulling another loop through the first loop, and repeating this process to create a chain of a suitable length. The chain is either turned and worked in rows, or joined to the beginning of the row with a slip stitch and worked in rounds. Rounds can also be created by working many stitches into a single loop. Stitches are made by pulling one or more loops through each loop of the chain. At any one time at the end of a stitch, there is only one loop left on the hook. Tunisian crochet, however, draws all of the loops for an entire row onto a long hook before working them off one at a time. Like knitting, crochet can be worked either flat (back and forth in rows) or in the round (in spirals, such as when making tubular pieces). There are five main types of basic stitches (the following description uses US crochet terminology which differs from the terminology used in the UK and Europe). While the horizontal distance covered by these basic stitches is the same, they differ in height and thickness. The more advanced stitches are often combinations of these basic stitches, or are made by inserting the hook into the work in unusual locations. More advanced stitches include the "shell stitch", "V stitch", "spike stitch", "Afghan stitch", "butterfly stitch", "popcorn stitch", "cluster stitch", and "crocodile stitch". In the English-speaking crochet world, basic stitches have different names that vary by country. The differences are usually referred to as UK/US or British/American. Crochet is traditionally worked off a written pattern in which stitches and placement are communicated using textual abbreviations. To help counter confusion when reading patterns, a diagramming system using a standard international notation has come into use (illustration, left). Another terminological difference is known as "tension" (UK) and "gauge" (US). Individual crocheters work yarn with a loose or a tight hold and, if unmeasured, these differences can lead to significant size changes in finished garments that have the same number of stitches. In order to control for this inconsistency, printed crochet instructions include a standard for the number of stitches across a standard swatch of fabric. An individual crocheter begins work by producing a test swatch and compensating for any discrepancy by changing to a smaller or larger hook. North Americans call this "gauge", referring to the end result of these adjustments; British crocheters speak of "tension", which refers to the crafter's grip on the yarn while producing stitches. One of the more obvious differences is that crochet uses one hook while much knitting uses two needles. In most crochet, the artisan usually has only one live stitch on the hook (with the exception being Tunisian crochet), while a knitter keeps an entire row of stitches active simultaneously. Dropped stitches, which can unravel a knitted fabric, rarely interfere with crochet work, due to a second structural difference between knitting and crochet. In knitting, each stitch is supported by the corresponding stitch in the row above and it supports the corresponding stitch in the row below, whereas crochet stitches are only supported by and support the stitches on either side of it. If a stitch in a finished crocheted item breaks, the stitches above and below remain intact, and because of the complex looping of each stitch, the stitches on either side are unlikely to come loose unless heavily stressed. Round or cylindrical patterns are simple to produce with a regular crochet hook, but cylindrical knitting requires either a set of circular needles or three to five special double-ended needles. Many crocheted items are composed of individual motifs which are then joined together, either by sewing or crocheting, whereas knitting is usually composed of one fabric, such as entrelac. Freeform crochet is a technique that can create interesting shapes in three dimensions because new stitches can be made independently of previous stitches almost anywhere in the crocheted piece. It is generally accomplished by building shapes or structural elements onto existing crocheted fabric at any place the crafter desires. Knitting can be accomplished by machine, while many crochet stitches can only be crafted by hand. The height of knitted and crocheted stitches is also different: a single crochet stitch is twice the height of a knit stitch in the same yarn size and comparable diameter tools, and a double crochet stitch is about four times the height of a knit stitch. While most crochet is made with a hook, there is also a method of crocheting with a knitting loom. This is called "loomchet". Slip stitch crochet is very similar to knitting. Each stitch in slip stitch crochet is formed the same way as a knit or purl stitch which is then bound off. A person working in slip stitch crochet can follow a knitted pattern with knits, purls, and cables, and get a similar result. It is a common perception that crochet produces a thicker fabric than knitting, tends to have less "give" than knitted fabric, and uses approximately a third more yarn for a comparable project than knitted items. Although this is true when comparing a single crochet swatch with a stockinette swatch, both made with the same size yarn and needle/hook, it is not necessarily true for crochet in general. Most crochet uses far less than 1/3 more yarn than knitting for comparable pieces, and a crocheter can get similar feel and drape to knitting by using a larger hook or thinner yarn. Tunisian crochet and slip stitch crochet can in some cases use less yarn than knitting for comparable pieces. According to sources claiming to have tested the 1/3 more yarn assertion, a single crochet stitch (sc) uses approximately the same amount of yarn as knit garter stitch, but more yarn than stockinette stitch. Any stitch using yarnovers uses less yarn than single crochet to produce the same amount of fabric. Cluster stitches, which are in fact multiple stitches worked together, will use the most length. Standard crochet stitches like sc and dc also produce a thicker fabric, more like knit garter stitch. This is part of why they use more yarn. Slip stitch can produce a fabric much like stockinette that is thinner and therefore uses less yarn. Any yarn can be either knitted or crocheted, provided needles or hooks of the correct size are used, but the cord's properties should be taken into account. For example, lofty, thick woolen yarns tend to function better when knitted, which does not crush their airy structure, while thin and tightly spun yarn helps to achieve the firm texture required for Amigurumi crochet. It has been very common for people and groups to crochet clothing and other garments and then donate them to soldiers during war. People have also crocheted clothing and then donated it to hospitals, for sick patients and also for newborn babies. Sometimes groups will crochet for a specific charity purpose, such as crocheting for homeless shelters, nursing homes, etc. It is becoming increasingly popular to crochet hats (commonly referred to as "chemo caps") and donate them to cancer treatment centers, for those undergoing chemotherapy and therefore losing hair. During October pink hats and scarves are made and proceeds are donated to breast cancer funds. Organizations dedicated to using crochet as a way to help others include Knots of Love, Crochet for Cancer, and Soldiers' Angels. These organizations offer warm useful items for people in need. Crocheting and other forms of needlecraft or handicraft are proven to have several health benefits including reducing stress and anxiety. Studies show that the constant repetition of hand movement helps keep the mind calm and distracts the brain from the stresses of life. When crocheting, the brain releases serotonin that helps calm the mind and improve mood. Crocheting and similar activities are shown to improve Alzheimer's disease, insomnia, and depression. The colors and textures of the yarn are usually pleasing to the senses, and the finished product can give a person a sense of accomplishment. Some elementary school teachers have used crocheting to enhance their students' literacy skills. Crocheting can help young children with fine motor skills (such as handwriting). Learning to crochet in a literary setting encourages young children to stay focused, multitask better, and be more excited about school. This fun yet challenging skill can in the process teach students to have better social skills in their community, with their peers, and with their families. Crochet has been used to illustrate shapes in hyperbolic space that are difficult to reproduce using other media or are difficult to understand when viewed two-dimensionally. Mathematician Daina Taimiņa first used crochet in 1997 to create strong, durable models of hyperbolic space after finding paper models were delicate and hard to create. These models enable one to turn, fold, and otherwise manipulate space to more fully grasp ideas such as how a line can appear curved in hyperbolic space yet actually be straight. Her work received an exhibition by the Institute For Figuring. Examples in nature of organisms that show hyperbolic structures include lettuces, sea slugs, flatworms and coral. Margaret Wertheim and Christine Wertheim of the Institute For Figuring created a travelling art installation of a coral reef using Taimina's method. Local artists are encouraged to create their own "satellite reefs" to be included alongside the original display. As hyperbolic and mathematics-based crochet has continued to become more popular, there have been several events highlighting work from various fiber artists. Two such shows include Sant Ocean Hall at the Smithsonian in Washington D.C. and Sticks, Hooks, and the Mobius: Knit and Crochet Go Cerebral at Lafayette College in Pennsylvania. In "Style in the technical arts", Gottfried Semper looks at the textile with great promise and historical precedent. In Section 53, he writes of the "loop stitch, or Noeud Coulant: a knot that, if untied, causes the whole system to unravel." In the same section, Semper confesses his ignorance of the subject of crochet but believes strongly that it is a technique of great value as a textile technique and possibly something more. There are a small number of architects currently interested in the subject of crochet as it relates to architecture. The following publications, explorations and thesis projects can be used as a resource to see how crochet is being used within the capacity of architecture. In the past few years, a practice called yarn bombing, or the use of knitted or crocheted cloth to modify and beautify one's (usually outdoor) surroundings, emerged in the US and spread worldwide. Yarn bombers sometimes target existing pieces of graffiti for beautification. In 2010, an entity dubbed "the Midnight Knitter" hit West Cape May. Residents awoke to find knit cozies hugging tree branches and sign poles. In September 2015, Grace Brett was named "The World's Oldest Yarn Bomber". She is part of a group of yarn graffiti-artists called the Souter Stormers, who beautify their local town in Scotland.
https://en.wikipedia.org/wiki?curid=7424
Electromagnetic coil An electromagnetic coil is an electrical conductor such as a wire in the shape of a coil, spiral or helix. Electromagnetic coils are used in electrical engineering, in applications where electric currents interact with magnetic fields, in devices such as electric motors, generators, inductors, electromagnets, transformers, and sensor coils. Either an electric current is passed through the wire of the coil to generate a magnetic field, or conversely an external "time-varying" magnetic field through the interior of the coil generates an EMF (voltage) in the conductor. A current through any conductor creates a circular magnetic field around the conductor due to Ampere's law. The advantage of using the coil shape is that it increases the strength of magnetic field produced by a given current. The magnetic fields generated by the separate turns of wire all pass through the center of the coil and add (superpose) to produce a strong field there. The more turns of wire, the stronger the field produced. Conversely, a "changing" external magnetic flux induces a voltage in a conductor such as a wire, due to Faraday's law of induction. The induced voltage can be increased by winding the wire into a coil, because the field lines intersect the circuit multiple times. The direction of the magnetic field produced by a coil can be determined by the right hand grip rule. If the fingers of the right hand are wrapped around the magnetic core of a coil in the direction of conventional current through the wire, the thumb will point in the direction the magnetic field lines pass through the coil. The end of a magnetic core from which the field lines emerge is defined to be the North pole. There are many different types of coils used in electric and electronic equipment. The wire or conductor which constitutes the coil is called the winding. The hole in the center of the coil is called the core area or "magnetic axis". Each loop of wire is called a turn. In windings in which the turns touch, the wire must be insulated with a coating of nonconductive insulation such as plastic or enamel to prevent the current from passing between the wire turns. The winding is often wrapped around a "coil form" made of plastic or other material to hold it in place. The ends of the wire are brought out and attached to an external circuit. Windings may have additional electrical connections along their length; these are called taps. A winding which has a single tap in the center of its length is called center-tapped. Coils can have more than one winding, insulated electrically from each other. When there are two or more windings around a common magnetic axis, the windings are said to be inductively coupled or magnetically coupled. A time-varying current through one winding will create a time-varying magnetic field which passes through the other winding, which will induce a time-varying voltage in the other windings. This is called a transformer. The winding to which current is applied, which creates the magnetic field, is called the "primary winding". The other windings are called "secondary windings". Many electromagnetic coils have a magnetic core, a piece of ferromagnetic material like iron in the center to increase the magnetic field. The current through the coil magnetizes the iron, and the field of the magnetized material adds to the field produced by the wire. This is called a ferromagnetic-core or iron-core coil. A ferromagnetic core can increase the magnetic field and inductance of a coil by hundreds or thousands of times over what it would be without the core. A ferrite core coil is a variety of coil with a core made of ferrite, a ferrimagnetic ceramic compound. Ferrite coils have lower core losses at high frequencies. A coil without a ferromagnetic core is called an air-core coil. This includes coils wound on plastic or other nonmagnetic forms, as well as coils which actually have empty air space inside their windings. Coils can be classified by the frequency of the current they are designed to operate with: Coils can be classified by their function: Electromagnets are coils that generate a magnetic field for some external use, often to exert a mechanical force on something. A few specific types: Inductors or reactors are coils which generate a magnetic field which interacts with the coil itself, to induce a back EMF which opposes changes in current through the coil. Inductors are used as circuit elements in electrical circuits, to temporarily store energy or resist changes in current. A few types: A transformer is a device with two or more magnetically coupled windings (or sections of a single winding). A time varying current in one coil (called the primary winding) generates a magnetic field which induces a voltage in the other coil (called the secondary winding). A few types: Electric machines such as motors and generators have one or more windings which interact with moving magnetic fields to convert electrical energy to mechanical energy. Often a machine will have one winding through which passes most of the power of the machine (the "armature"), and a second winding which provides the magnetic field of the rotating element ( the "field winding") which may be connected by brushes or slip rings to an external source of electric current. In an induction motor, the "field" winding of the rotor is energized by the slow relative motion between the rotating winding and the rotating magnetic field produced by the stator winding, which induces the necessary exciting current in the rotor. These are coils used to translate time-varying magnetic fields to electric signals, and vice versa. A few types: There are also types of coil which don't fit into these categories.
https://en.wikipedia.org/wiki?curid=7425
Charles I of England Charles I (19 November 1600 – 30 January 1649) was King of England, King of Scotland, and King of Ireland from 27 March 1625 until his execution in 1649. Charles was born into the House of Stuart as the second son of King James VI of Scotland, but after his father inherited the English throne in 1603 (as James I), he moved to England, where he spent much of the rest of his life. He became heir apparent to the three kingdoms of England, Scotland and Ireland in 1612 on the death of his elder brother Henry Frederick, Prince of Wales. An unsuccessful and unpopular attempt to marry him to the Spanish Habsburg princess Maria Anna culminated in an eight-month visit to Spain in 1623 that demonstrated the futility of the marriage negotiations. Two years later he married the Bourbon princess Henrietta Maria of France. After his succession in 1625, Charles quarrelled with the Parliament of England, which sought to curb his royal prerogative. Charles believed in the divine right of kings, and was determined to govern according to his own conscience. Many of his subjects opposed his policies, in particular the levying of taxes without parliamentary consent, and perceived his actions as those of a tyrannical absolute monarch. His religious policies, coupled with his marriage to a Roman Catholic, generated antipathy and mistrust from Reformed religious groups such as the English Puritans and Scottish Covenanters, who thought his views were too Catholic. He supported high church Anglican ecclesiastics such as Richard Montagu and William Laud, and failed to aid continental Protestant forces successfully during the Thirty Years' War. His attempts to force the Church of Scotland to adopt high Anglican practices led to the Bishops' Wars, strengthened the position of the English and Scottish parliaments, and helped precipitate his own downfall. From 1642, Charles fought the armies of the English and Scottish parliaments in the English Civil War. After his defeat in 1645, he surrendered to a Scottish force that eventually handed him over to the English Parliament. Charles refused to accept his captors' demands for a constitutional monarchy, and temporarily escaped captivity in November 1647. Re-imprisoned on the Isle of Wight, Charles forged an alliance with Scotland, but by the end of 1648 Oliver Cromwell's New Model Army had consolidated its control over England. Charles was tried, convicted, and executed for high treason in January 1649. The monarchy was abolished and the Commonwealth of England was established as a republic. The monarchy would be restored to Charles's son, Charles II, in 1660. The second son of King James VI of Scotland and Anne of Denmark, Charles was born in Dunfermline Palace, Fife, on 19 November 1600. At a Protestant ceremony in the Chapel Royal of Holyrood Palace in Edinburgh on 23 December 1600, he was baptised by David Lindsay, Bishop of Ross, and created Duke of Albany, the traditional title of the second son of the King of Scotland, with the subsidiary titles of Marquess of Ormond, Earl of Ross and Lord Ardmannoch. James VI was the first cousin twice removed of Queen Elizabeth I of England, and when she died childless in March 1603, he became King of England as James I. Charles was a weak and sickly infant, and while his parents and older siblings left for England in April and early June that year, due to his fragile health, he remained in Scotland with his father's friend Lord Fyvie, appointed as his guardian. By 1604, when Charles was three-and-a-half, he was able to walk the length of the great hall at Dunfermline Palace without assistance, and it was decided that he was strong enough to make the journey to England to be reunited with his family. In mid-July 1604, Charles left Dunfermline for England where he was to spend most of the rest of his life. In England, Charles was placed under the charge of Elizabeth, Lady Carey, the wife of courtier Sir Robert Carey, who put him in boots made of Spanish leather and brass to help strengthen his weak ankles. His speech development was also slow, and he retained a stammer, or hesitant speech, for the rest of his life. In January 1605, Charles was created Duke of York, as is customary in the case of the English sovereign's second son, and made a Knight of the Bath. Thomas Murray, a presbyterian Scot, was appointed as a tutor. Charles learnt the usual subjects of classics, languages, mathematics and religion. In 1611, he was made a Knight of the Garter. Eventually, Charles apparently conquered his physical infirmity, which might have been caused by rickets. He became an adept horseman and marksman, and took up fencing. Even so, his public profile remained low in contrast to that of his physically stronger and taller elder brother, Henry Frederick, Prince of Wales, whom Charles adored and attempted to emulate. However, in early November 1612, Henry died at the age of 18 of what is suspected to have been typhoid (or possibly porphyria). Charles, who turned 12 two weeks later, became heir apparent. As the eldest surviving son of the sovereign, Charles automatically gained several titles (including Duke of Cornwall and Duke of Rothesay). Four years later, in November 1616, he was created Prince of Wales and Earl of Chester. In 1613, Charles's sister Elizabeth married Frederick V, Elector Palatine, and moved to Heidelberg. In 1617, the Habsburg Archduke Ferdinand of Austria, a Catholic, was elected king of Bohemia. The following year, the Bohemians rebelled, defenestrating the Catholic governors. In August 1619, the Bohemian diet chose as their monarch Frederick V, who was leader of the Protestant Union, while Ferdinand was elected Holy Roman Emperor in the imperial election. Frederick's acceptance of the Bohemian crown in defiance of the emperor marked the beginning of the turmoil that would develop into the Thirty Years' War. The conflict, originally confined to Bohemia, spiralled into a wider European war, which the English Parliament and public quickly grew to see as a polarised continental struggle between Catholics and Protestants. In 1620, Charles's brother-in-law, Frederick V, was defeated at the Battle of White Mountain near Prague and his hereditary lands in the Electoral Palatinate were invaded by a Habsburg force from the Spanish Netherlands. James, however, had been seeking marriage between the new Prince of Wales and Ferdinand's niece, Habsburg princess Maria Anna of Spain, and began to see the Spanish match as a possible diplomatic means of achieving peace in Europe. Unfortunately for James, negotiation with Spain proved generally unpopular, both with the public and with James's court. The English Parliament was actively hostile towards Spain and Catholicism, and thus, when called by James in 1621, the members hoped for an enforcement of recusancy laws, a naval campaign against Spain, and a Protestant marriage for the Prince of Wales. James's Lord Chancellor, Francis Bacon, was impeached before the House of Lords for corruption. The impeachment was the first since 1459 without the king's official sanction in the form of a bill of attainder. The incident set an important precedent as the process of impeachment would later be used against Charles and his supporters: the Duke of Buckingham, Archbishop William Laud, and the Earl of Strafford. James insisted that the House of Commons be concerned exclusively with domestic affairs, while the members protested that they had the privilege of free speech within the Commons' walls, demanding war with Spain and a Protestant Princess of Wales. Charles, like his father, considered the discussion of his marriage in the Commons impertinent and an infringement of his father's royal prerogative. In January 1622, James dissolved Parliament, angry at what he perceived as the members' impudence and intransigence. Charles and Buckingham, James's favourite and a man who had great influence over the prince, travelled incognito to Spain in February 1623 to try to reach agreement on the long-pending Spanish match. In the end, however, the trip was an embarrassing failure. The Infanta thought Charles to be little more than an infidel, and the Spanish at first demanded that he convert to Roman Catholicism as a condition of the match. The Spanish insisted on toleration of Catholics in England and the repeal of the penal laws, which Charles knew would never be agreed by Parliament, and that the Infanta remain in Spain for a year after any wedding to ensure that England complied with all the terms of the treaty. A personal quarrel erupted between Buckingham and the Count of Olivares, the Spanish chief minister, and so Charles conducted the ultimately futile negotiations personally. When Charles returned to London in October, without a bride and to a rapturous and relieved public welcome, he and Buckingham pushed a reluctant King James to declare war on Spain. With the encouragement of his Protestant advisers, James summoned the English Parliament in 1624 so that he could request subsidies for a war. Charles and Buckingham supported the impeachment of the Lord Treasurer, Lionel Cranfield, 1st Earl of Middlesex, who opposed war on grounds of cost and who quickly fell in much the same manner as Bacon had. James told Buckingham he was a fool, and presciently warned his son Charles that he would live to regret the revival of impeachment as a parliamentary tool. An under-funded makeshift army under Ernst von Mansfeld set off to recover the Palatinate, but it was so poorly provisioned that it never advanced beyond the Dutch coast. By 1624, an increasingly ill James was finding it difficult to control Parliament. By the time of his death in March 1625, Charles and the Duke of Buckingham had already assumed "de facto" control of the kingdom. With the failure of the Spanish match, Charles and Buckingham turned their attention to France. On 1 May 1625 Charles was married by proxy to the fifteen-year-old French princess Henrietta Maria in front of the doors of Notre Dame de Paris. Charles had seen Henrietta Maria in Paris while en route to Spain. The married couple met in person on 13 June 1625 in Canterbury. Charles delayed the opening of his first Parliament until after the marriage was consummated, to forestall any opposition. Many members of the Commons were opposed to the king's marriage to a Roman Catholic, fearing that Charles would lift restrictions on Catholic recusants and undermine the official establishment of the reformed Church of England. Although he told Parliament that he would not relax religious restrictions, he promised to do exactly that in a secret marriage treaty with his brother-in-law Louis XIII of France. Moreover, the treaty loaned to the French seven English naval ships that would be used to suppress the Protestant Huguenots at La Rochelle in September 1625. Charles was crowned on 2 February 1626 at Westminster Abbey, but without his wife at his side because she refused to participate in a Protestant religious ceremony. Distrust of Charles's religious policies increased with his support of a controversial anti-Calvinist ecclesiastic, Richard Montagu, who was in disrepute among the Puritans. In his pamphlet "A New Gag for an Old Goose" (1624), a reply to the Catholic pamphlet "A New Gag for the New Gospel", Montagu argued against Calvinist predestination, the doctrine that salvation and damnation were preordained by God. Anti-Calvinistsknown as Arminiansbelieved that human beings could influence their own fate through the exercise of free will. Arminian divines had been one of the few sources of support for Charles's proposed Spanish marriage. With the support of King James, Montagu produced another pamphlet, entitled "Appello Caesarem", in 1625 shortly after the old king's death and Charles's accession. To protect Montagu from the stricture of Puritan members of Parliament, Charles made the cleric one of his royal chaplains, increasing many Puritans' suspicions that Charles favoured Arminianism as a clandestine attempt to aid the resurgence of Catholicism. Rather than direct involvement in the European land war, the English Parliament preferred a relatively inexpensive naval attack on Spanish colonies in the New World, hoping for the capture of the Spanish treasure fleets. Parliament voted to grant a subsidy of £140,000, which was an insufficient sum for Charles's war plans. Moreover, the House of Commons limited its authorisation for royal collection of tonnage and poundage (two varieties of customs duties) to a period of one year, although previous sovereigns since Henry VI had been granted the right for life. In this manner, Parliament could delay approval of the rates until after a full-scale review of customs revenue. The bill made no progress in the House of Lords past its first reading. Although no Parliamentary Act for the levy of tonnage and poundage was obtained, Charles continued to collect the duties. A poorly conceived and executed naval expedition against Spain under the leadership of Buckingham went badly, and the House of Commons began proceedings for the impeachment of the duke. In May 1626, Charles nominated Buckingham as Chancellor of Cambridge University in a show of support, and had two members who had spoken against BuckinghamDudley Digges and Sir John Eliotarrested at the door of the House. The Commons was outraged by the imprisonment of two of their members, and after about a week in custody, both were released. On 12 June 1626, the Commons launched a direct protestation attacking Buckingham, stating, "We protest before your Majesty and the whole world that until this great person be removed from intermeddling with the great affairs of state, we are out of hope of any good success; and do fear that any money we shall or can give will, through his misemployment, be turned rather to the hurt and prejudice of this your kingdom than otherwise, as by lamentable experience we have found those large supplies formerly and lately given." Despite Parliament's protests, however, Charles refused to dismiss his friend, dismissing Parliament instead. Meanwhile, domestic quarrels between Charles and Henrietta Maria were souring the early years of their marriage. Disputes over her jointure, appointments to her household, and the practice of her religion culminated in the king expelling the vast majority of her French attendants in August 1626. Despite Charles's agreement to provide the French with English ships as a condition of marrying Henrietta Maria, in 1627 he launched an attack on the French coast to defend the Huguenots at La Rochelle. The action, led by Buckingham, was ultimately unsuccessful. Buckingham's failure to protect the Huguenotsand his retreat from Saint-Martin-de-Réspurred Louis XIII's siege of La Rochelle and furthered the English Parliament's and people's detestation of the duke. Charles provoked further unrest by trying to raise money for the war through a "forced loan": a tax levied without parliamentary consent. In November 1627, the test case in the King's Bench, the "Five Knights' Case", found that the king had a prerogative right to imprison without trial those who refused to pay the forced loan. Summoned again in March 1628, on 26 May Parliament adopted a Petition of Right, calling upon the king to acknowledge that he could not levy taxes without Parliament's consent, not impose martial law on civilians, not imprison them without due process, and not quarter troops in their homes. Charles assented to the petition on 7 June, but by the end of the month he had prorogued Parliament and re-asserted his right to collect customs duties without authorisation from Parliament. On 23 August 1628, Buckingham was assassinated. Charles was deeply distressed. According to Edward Hyde, 1st Earl of Clarendon, he "threw himself upon his bed, lamenting with much passion and with abundance of tears". He remained grieving in his room for two days. In contrast, the public rejoiced at Buckingham's death, which accentuated the gulf between the court and the nation, and between the Crown and the Commons. Although the death of Buckingham effectively ended the war with Spain and eliminated his leadership as an issue, it did not end the conflicts between Charles and Parliament. It did, however, coincide with an improvement in Charles's relationship with his wife, and by November 1628 their old quarrels were at an end. Perhaps Charles's emotional ties were transferred from Buckingham to Henrietta Maria. She became pregnant for the first time, and the bond between them grew stronger. Together, they embodied an image of virtue and family life, and their court became a model of formality and morality. In January 1629, Charles opened the second session of the English Parliament, which had been prorogued in June 1628, with a moderate speech on the tonnage and poundage issue. Members of the House of Commons began to voice opposition to Charles's policies in light of the case of John Rolle, a Member of Parliament whose goods had been confiscated for failing to pay tonnage and poundage. Many MPs viewed the imposition of the tax as a breach of the Petition of Right. When Charles ordered a parliamentary adjournment on 2 March, members held the Speaker, Sir John Finch, down in his chair so that the ending of the session could be delayed long enough for resolutions against Catholicism, Arminianism and tonnage and poundage to be read out and acclaimed by the chamber. The provocation was too much for Charles, who dissolved Parliament and had nine parliamentary leaders, including Sir John Eliot, imprisoned over the matter, thereby turning the men into martyrs, and giving popular cause to their protest. Personal rule necessitated peace. Without the means in the foreseeable future to raise funds from Parliament for a European war, or the help of Buckingham, Charles made peace with France and Spain. The following eleven years, during which Charles ruled England without a Parliament, are referred to as the personal rule or the "eleven years' tyranny". Ruling without Parliament was not exceptional, and was supported by precedent. Only Parliament, however, could legally raise taxes, and without it Charles's capacity to acquire funds for his treasury was limited to his customary rights and prerogatives. A large fiscal deficit had arisen in the reigns of Elizabeth I and James I. Notwithstanding Buckingham's short-lived campaigns against both Spain and France, there was little financial capacity for Charles to wage wars overseas. Throughout his reign Charles was obliged to rely primarily on volunteer forces for defence and on diplomatic efforts to support his sister, Elizabeth, and his foreign policy objective for the restoration of the Palatinate. England was still the least taxed country in Europe, with no official excise and no regular direct taxation. To raise revenue without reconvening Parliament, Charles resurrected an all-but-forgotten law called the "Distraint of Knighthood", in abeyance for over a century, which required any man who earned £40 or more from land each year to present himself at the king's coronation to be knighted. Relying on this old statute, Charles fined individuals who had failed to attend his coronation in 1626. The chief tax imposed by Charles was a feudal levy known as ship money, which proved even more unpopular, and lucrative, than tonnage and poundage before it. Previously, collection of ship money had been authorised only during wars, and only on coastal regions. Charles, however, argued that there was no legal bar to collecting the tax for defence during peacetime and throughout the whole of the kingdom. Ship money, paid directly to the Treasury of the Navy, provided between £150,000 to £200,000 annually between 1634 and 1638, after which yields declined. Opposition to ship money steadily grew, but the 12 common law judges of England declared that the tax was within the king's prerogative, though some of them had reservations. The prosecution of John Hampden for non-payment in 1637–38 provided a platform for popular protest, and the judges found against Hampden only by the narrow margin of 7–5. The king also derived money through the granting of monopolies, despite a statute forbidding such action, which, though inefficient, raised an estimated £100,000 a year in the late 1630s. One such monopoly was for soap, pejoratively referred to as "popish soap" because some of its backers were Catholics. Charles also raised funds from the Scottish nobility, at the price of considerable acrimony, by the Act of Revocation (1625), whereby all gifts of royal or church land made to the nobility since 1540 were revoked, with continued ownership being subject to an annual rent. In addition, the boundaries of the royal forests in England were restored to their ancient limits as part of a scheme to maximise income by exploiting the land and fining land users within the reasserted boundaries for encroachment. The focus of the programme was disafforestation and sale of forest lands for conversion to pasture and arable farming, or in the case of the Forest of Dean, development for the iron industry. Disafforestation frequently caused riots and disturbances including those known as the Western Rising. Against the background of this unrest, Charles faced bankruptcy in mid-1640. The City of London, preoccupied with its own grievances, refused to make any loans to the king, as did foreign powers. In this extremity, in July Charles seized silver bullion worth £130,000 held in trust at the mint in the Tower of London, promising its later return at 8% interest to its owners. In August, after the East India Company refused to grant a loan, Lord Cottington seized the company's stock of pepper and spices and sold it for £60,000 (far below its market value), promising to refund the money with interest later. Throughout Charles's reign, the issue of how far the English Reformation should progress was constantly in the forefront of political debate. Arminian theology emphasised clerical authority and the individual's ability to reject or accept salvation, which opponents viewed as heretical and a potential vehicle for the reintroduction of Roman Catholicism. Puritan reformers thought Charles was too sympathetic to the teachings of Arminianism, which they considered irreligious, and opposed his desire to move the Church of England in a more traditional and sacramental direction. In addition, his Protestant subjects followed the European war closely and grew increasingly dismayed by Charles's diplomacy with Spain and his failure to support the Protestant cause abroad effectively. In 1633, Charles appointed William Laud Archbishop of Canterbury. They initiated a series of reforms aimed at ensuring religious uniformity by restricting non-conformist preachers, insisting the liturgy be celebrated as prescribed by the Book of Common Prayer, organising the internal architecture of English churches to emphasise the sacrament of the altar, and re-issuing King James's Declaration of Sports, which permitted secular activities on the sabbath. The Feoffees for Impropriations, an organisation that bought benefices and advowsons so that Puritans could be appointed to them, was dissolved. Laud prosecuted those who opposed his reforms in the Court of High Commission and the Star Chamber, the two most powerful courts in the land. The courts became feared for their censorship of opposing religious views and unpopular among the propertied classes for inflicting degrading punishments on gentlemen. For example, in 1637 William Prynne, Henry Burton and John Bastwick were pilloried, whipped and mutilated by cropping and imprisoned indefinitely for publishing anti-episcopal pamphlets. When Charles attempted to impose his religious policies in Scotland he faced numerous difficulties. Although born in Scotland, Charles had become estranged from his northern kingdom; his first visit since early childhood was for his Scottish coronation in 1633. To the dismay of the Scots, who had removed many traditional rituals from their liturgical practice, Charles insisted that the coronation be conducted using the Anglican rite. In 1637, the king ordered the use of a new prayer book in Scotland that was almost identical to the English Book of Common Prayer, without consulting either the Scottish Parliament or the Kirk. Although it had been written, under Charles's direction, by Scottish bishops, many Scots resisted it, seeing the new prayer book as a vehicle for introducing Anglicanism to Scotland. On 23 July, riots erupted in Edinburgh upon the first Sunday of the prayer book's usage, and unrest spread throughout the Kirk. The public began to mobilise around a reaffirmation of the National Covenant, whose signatories pledged to uphold the reformed religion of Scotland and reject any innovations that were not authorised by Kirk and Parliament. When the General Assembly of the Church of Scotland met in November 1638, it condemned the new prayer book, abolished episcopal church government by bishops, and adopted presbyterian government by elders and deacons. Charles perceived the unrest in Scotland as a rebellion against his authority, precipitating the First Bishops' War in 1639. Charles did not seek subsidies from the English Parliament to wage war, but instead raised an army without parliamentary aid and marched to Berwick-upon-Tweed, on the border of Scotland. Charles's army did not engage the Covenanters as the king feared the defeat of his forces, whom he believed to be significantly outnumbered by the Scots. In the Treaty of Berwick, Charles regained custody of his Scottish fortresses and secured the dissolution of the Covenanters' interim government, albeit at the decisive concession that both the Scottish Parliament and General Assembly of the Scottish Church were called. The military failure in the First Bishops' War caused a financial and diplomatic crisis for Charles that deepened when his efforts to raise funds from Spain, while simultaneously continuing his support for his Palatine relatives, led to the public humiliation of the Battle of the Downs, where the Dutch destroyed a Spanish bullion fleet off the coast of Kent in sight of the impotent English navy. Charles continued peace negotiations with the Scots in a bid to gain time before launching a new military campaign. Because of his financial weakness, he was forced to call Parliament into session in an attempt to raise funds for such a venture. Both English and Irish parliaments were summoned in the early months of 1640. In March 1640, the Irish Parliament duly voted in a subsidy of £180,000 with the promise to raise an army 9,000 strong by the end of May. In the English general election in March, however, court candidates fared badly, and Charles's dealings with the English Parliament in April quickly reached stalemate. The earls of Northumberland and Strafford attempted to broker a compromise whereby the king would agree to forfeit ship money in exchange for £650,000 (although the cost of the coming war was estimated at around £1 million). Nevertheless, this alone was insufficient to produce consensus in the Commons. The Parliamentarians' calls for further reforms were ignored by Charles, who still retained the support of the House of Lords. Despite the protests of Northumberland, the Short Parliament (as it came to be known) was dissolved in May 1640, less than a month after it assembled. By this stage Strafford, Lord Deputy of Ireland since 1632, had emerged as Charles's right-hand man and together with Laud, pursued a policy of "Thorough" that aimed to make central royal authority more efficient and effective at the expense of local or anti-government interests. Although originally a critic of the king, Strafford defected to royal service in 1628 (due in part to Buckingham's persuasion), and had since emerged, alongside Laud, as the most influential of Charles's ministers. Bolstered by the failure of the English Short Parliament, the Scottish Parliament declared itself capable of governing without the king's consent, and in August 1640 the Covenanter army moved into the English county of Northumberland. Following the illness of the earl of Northumberland, who was the king's commander-in-chief, Charles and Strafford went north to command the English forces, despite Strafford being ill himself with a combination of gout and dysentery. The Scottish soldiery, many of whom were veterans of the Thirty Years' War, had far greater morale and training compared to their English counterparts. They met virtually no resistance until reaching Newcastle upon Tyne, where they defeated the English forces at the Battle of Newburn and occupied the city, as well as the neighbouring county of Durham. As demands for a parliament grew, Charles took the unusual step of summoning a great council of peers. By the time it met, on 24 September at York, Charles had resolved to follow the almost universal advice to call a parliament. After informing the peers that a parliament would convene in November, he asked them to consider how he could acquire funds to maintain his army against the Scots in the meantime. They recommended making peace. A cessation of arms, although not a final settlement, was negotiated in the humiliating Treaty of Ripon, signed in October 1640. The treaty stated that the Scots would continue to occupy Northumberland and Durham and be paid £850 per day until peace was restored and the English Parliament recalled, which would be required to raise sufficient funds to pay the Scottish forces. Consequently, Charles summoned what later became known as the Long Parliament. Once again, Charles's supporters fared badly at the polls. Of the 493 members of the Commons returned in November, over 350 were opposed to the king. The Long Parliament proved just as difficult for Charles as had the Short Parliament. It assembled on 3 November 1640 and quickly began proceedings to impeach the king's leading counsellors of high treason. Strafford was taken into custody on 10 November; Laud was impeached on 18 December; John Finch, now Lord Keeper of the Great Seal, was impeached the following day, and he consequently fled to the Hague with Charles's permission on 21 December. To prevent the king from dissolving it at will, Parliament passed the Triennial Act, which required Parliament to be summoned at least once every three years, and permitted the Lord Keeper and 12 peers to summon Parliament if the king failed to do so. The Act was coupled with a subsidy bill, and so to secure the latter, Charles grudgingly granted royal assent in February 1641. Strafford had become the principal target of the Parliamentarians, particularly John Pym, and he went on trial for high treason on 22 March 1641. However, the key allegation by Sir Henry Vane that Strafford had threatened to use the Irish army to subdue England was not corroborated and on 10 April Pym's case collapsed. Pym and his allies immediately launched a bill of attainder, which simply declared Strafford guilty and pronounced the sentence of death. Charles assured Strafford that "upon the word of a king you shall not suffer in life, honour or fortune", and the attainder could not succeed if Charles withheld assent. Furthermore, many members and most peers were opposed to the attainder, not wishing, in the words of one, to "commit murder with the sword of justice". However, increased tensions and an attempted coup by royalist army officers in support of Strafford and in which Charles was involved began to sway the issue. The Commons passed the bill on 20 April by a large margin (204 in favour, 59 opposed, and 230 abstained), and the Lords acquiesced (by 26 votes to 19, with 79 absent) in May. On 3 May, Parliament's Protestation attacked the "wicked counsels" of Charles's "arbitrary and tyrannical government". While those who signed the petition undertook to defend the king's "person, honour and estate", they also swore to preserve "the true reformed religion", parliament, and the "rights and liberties of the subjects". Charles, fearing for the safety of his family in the face of unrest, assented reluctantly to Strafford's attainder on 9 May after consulting his judges and bishops. Strafford was beheaded three days later. Additionally in early May, Charles assented to an unprecedented Act that forbade the dissolution of the English Parliament without its consent. In the following months, ship money, fines in distraint of knighthood and excise without parliamentary consent were declared unlawful, and the Courts of Star Chamber and High Commission were abolished. All remaining forms of taxation were legalised and regulated by the Tonnage and Poundage Act. The House of Commons also launched bills attacking bishops and episcopacy, but these failed in the Lords. Charles had made important concessions in England, and temporarily improved his position in Scotland by securing the favour of the Scots on a visit from August to November 1641 during which he conceded to the official establishment of presbyterianism. However, following an attempted royalist coup in Scotland, known as "The Incident", Charles's credibility was significantly undermined. In Ireland, the population was split into three main socio-political groups: the Gaelic Irish, who were Catholic; the Old English, who were descended from medieval Normans and were also predominantly Catholic; and the New English, who were Protestant settlers from England and Scotland aligned with the English Parliament and the Covenanters. Strafford's administration had improved the Irish economy and boosted tax revenue, but had done so by heavy-handedly imposing order. He had trained up a large Catholic army in support of the king and had weakened the authority of the Irish Parliament, while continuing to confiscate land from Catholics for Protestant settlement at the same time as promoting a Laudian Anglicanism that was anathema to presbyterians. As a result, all three groups had become disaffected. Strafford's impeachment provided a new departure for Irish politics whereby all sides joined together to present evidence against him. In a similar manner to the English Parliament, the Old English members of the Irish Parliament argued that while opposed to Strafford they remained loyal to Charles. They argued that the king had been led astray by malign counsellors, and that, moreover, a viceroy such as Strafford could emerge as a despotic figure instead of ensuring that the king was directly involved in governance. Strafford's fall from power weakened Charles's influence in Ireland. The dissolution of the Irish army was unsuccessfully demanded three times by the English Commons during Strafford's imprisonment, until Charles was eventually forced through lack of money to disband the army at the end of Strafford's trial. Disputes concerning the transfer of land ownership from native Catholic to settler Protestant, particularly in relation to the plantation of Ulster, coupled with resentment at moves to ensure the Irish Parliament was subordinate to the Parliament of England, sowed the seeds of rebellion. When armed conflict arose between the Gaelic Irish and New English, in late October 1641, the Old English sided with the Gaelic Irish while simultaneously professing their loyalty to the king. In November 1641, the House of Commons passed the Grand Remonstrance, a long list of grievances against actions by Charles's ministers committed since the beginning of his reign (that were asserted to be part of a grand Catholic conspiracy of which the king was an unwitting member), but it was in many ways a step too far by Pym and passed by only 11 votes – 159 to 148. Furthermore, the Remonstrance had very little support in the House of Lords, which the Remonstrance attacked. The tension was heightened by news of the Irish rebellion, coupled with inaccurate rumours of Charles's complicity. Throughout November, a series of alarmist pamphlets published stories of atrocities in Ireland, which included massacres of New English settlers by the native Irish who could not be controlled by the Old English lords. Rumours of "papist" conspiracies circulated in England, and English anti-Catholic opinion was strengthened, damaging Charles's reputation and authority. The English Parliament distrusted Charles's motivations when he called for funds to put down the Irish rebellion; many members of the Commons suspected that forces raised by Charles might later be used against Parliament itself. Pym's Militia Bill was intended to wrest control of the army from the king, but it did not have the support of the Lords, let alone Charles. Instead, the Commons passed the bill as an ordinance, which they claimed did not require royal assent. The Militia Ordinance appears to have prompted more members of the Lords to support the king. In an attempt to strengthen his position, Charles generated great antipathy in London, which was already fast falling into lawlessness, when he placed the Tower of London under the command of Colonel Thomas Lunsford, an infamous, albeit efficient, career officer. When rumours reached Charles that Parliament intended to impeach his wife for supposedly conspiring with the Irish rebels, the king decided to take drastic action. Charles suspected, probably correctly, that some members of the English Parliament had colluded with the invading Scots. On 3 January 1642, Charles directed Parliament to give up five members of the Commons – Pym, John Hampden, Denzil Holles, William Strode and Sir Arthur Haselrig – and one peer – Lord Mandeville – on the grounds of high treason. When Parliament refused, it was possibly Henrietta Maria who persuaded Charles to arrest the five members by force, which Charles intended to carry out personally. However, news of the warrant reached Parliament ahead of him, and the wanted men slipped away by boat shortly before Charles entered the House of Commons with an armed guard on 4 January. Having displaced the Speaker, William Lenthall, from his chair, the king asked him where the MPs had fled. Lenthall, on his knees, famously replied, "May it please your Majesty, I have neither eyes to see nor tongue to speak in this place but as the House is pleased to direct me, whose servant I am here." Charles abjectly declared "all my birds have flown", and was forced to retire, empty-handed. The botched arrest attempt was politically disastrous for Charles. No English sovereign had ever entered the House of Commons, and his unprecedented invasion of the chamber to arrest its members was considered a grave breach of parliamentary privilege. In one stroke Charles destroyed his supporters' efforts to portray him as a defence against innovation and disorder. Parliament quickly seized London, and Charles fled the capital for Hampton Court Palace on 10 January, moving two days later to Windsor Castle. After sending his wife and eldest daughter to safety abroad in February, he travelled northwards, hoping to seize the military arsenal at Hull. To his dismay, he was rebuffed by the town's Parliamentary governor, Sir John Hotham, who refused him entry in April, and Charles was forced to withdraw. In mid-1642, both sides began to arm. Charles raised an army using the medieval method of commission of array, and Parliament called for volunteers for its militia. Following futile negotiations, Charles raised the royal standard in Nottingham on 22 August 1642. At the start of the First English Civil War, Charles's forces controlled roughly the Midlands, Wales, the West Country and northern England. He set up his court at Oxford. Parliament controlled London, the south-east and East Anglia, as well as the English navy. After a few skirmishes, the opposing forces met in earnest at Edgehill, on 23 October 1642. Charles's nephew Prince Rupert of the Rhine disagreed with the battle strategy of the royalist commander Lord Lindsey, and Charles sided with Rupert. Lindsey resigned, leaving Charles to assume overall command assisted by Lord Forth. Rupert's cavalry successfully charged through the parliamentary ranks, but instead of swiftly returning to the field, rode off to plunder the parliamentary baggage train. Lindsey, acting as a colonel, was wounded and bled to death without medical attention. The battle ended inconclusively as the daylight faded. In his own words, the experience of battle had left Charles "exceedingly and deeply grieved". He regrouped at Oxford, turning down Rupert's suggestion of an immediate attack on London. After a week, he set out for the capital on 3 November, capturing Brentford on the way while simultaneously continuing to negotiate with civic and parliamentary delegations. At Turnham Green on the outskirts of London, the royalist army met resistance from the city militia, and faced with a numerically superior force, Charles ordered a retreat. He overwintered in Oxford, strengthening the city's defences and preparing for the next season's campaign. Peace talks between the two sides collapsed in April. The war continued indecisively over the next couple of years, and Henrietta Maria returned to Britain for 17 months from February 1643. After Rupert captured Bristol in July 1643, Charles visited the port city and laid siege to Gloucester, further up the river Severn. His plan to undermine the city walls failed due to heavy rain, and on the approach of a parliamentary relief force, Charles lifted the siege and withdrew to Sudeley Castle. The parliamentary army turned back towards London, and Charles set off in pursuit. The two armies met at Newbury, Berkshire, on 20 September. Just as at Edgehill, the battle stalemated at nightfall, and the armies disengaged. In January 1644, Charles summoned a Parliament at Oxford, which was attended by about 40 peers and 118 members of the Commons; all told, the Oxford Parliament, which sat until March 1645, was supported by the majority of peers and about a third of the Commons. Charles became disillusioned by the assembly's ineffectiveness, calling it a "mongrel" in private letters to his wife. In 1644, Charles remained in the southern half of England while Rupert rode north to relieve Newark and York, which were under threat from parliamentary and Scottish Covenanter armies. Charles was victorious at the battle of Cropredy Bridge in late June, but the royalists in the north were defeated at the battle of Marston Moor just a few days later. The king continued his campaign in the south, encircling and disarming the parliamentary army of the Earl of Essex. Returning northwards to his base at Oxford, he fought at Newbury for a second time before the winter closed in; the battle ended indecisively. Attempts to negotiate a settlement over the winter, while both sides re-armed and re-organised, were again unsuccessful. At the battle of Naseby on 14 June 1645, Rupert's horsemen again mounted a successful charge against the flank of Parliament's New Model Army, but Charles's troops elsewhere on the field were pushed back by the opposing forces. Charles, attempting to rally his men, rode forward but as he did so, Lord Carnwath seized his bridle and pulled him back, fearing for the king's safety. Carnwath's action was misinterpreted by the royalist soldiers as a signal to move back, leading to a collapse of their position. The military balance tipped decisively in favour of Parliament. There followed a series of defeats for the royalists, and then the Siege of Oxford, from which Charles escaped (disguised as a servant) in April 1646. He put himself into the hands of the Scottish presbyterian army besieging Newark, and was taken northwards to Newcastle upon Tyne. After nine months of negotiations, the Scots finally arrived at an agreement with the English Parliament: in exchange for £100,000, and the promise of more money in the future, the Scots withdrew from Newcastle and delivered Charles to the parliamentary commissioners in January 1647. Parliament held Charles under house arrest at Holdenby House in Northamptonshire until Cornet George Joyce took him by threat of force from Holdenby on 3 June in the name of the New Model Army. By this time, mutual suspicion had developed between Parliament, which favoured army disbandment and presbyterianism, and the New Model Army, which was primarily officered by congregationalist Independents, who sought a greater political role. Charles was eager to exploit the widening divisions, and apparently viewed Joyce's actions as an opportunity rather than a threat. He was taken first to Newmarket, at his own suggestion, and then transferred to Oatlands and subsequently Hampton Court, while more ultimately fruitless negotiations took place. By November, he determined that it would be in his best interests to escape – perhaps to France, Southern England or to Berwick-upon-Tweed, near the Scottish border. He fled Hampton Court on 11 November, and from the shores of Southampton Water made contact with Colonel Robert Hammond, Parliamentary Governor of the Isle of Wight, whom he apparently believed to be sympathetic. Hammond, however, confined Charles in Carisbrooke Castle and informed Parliament that Charles was in his custody. From Carisbrooke, Charles continued to try to bargain with the various parties. In direct contrast to his previous conflict with the Scottish Kirk, on 26 December 1647 he signed a secret treaty with the Scots. Under the agreement, called the "Engagement", the Scots undertook to invade England on Charles's behalf and restore him to the throne on condition that presbyterianism be established in England for three years. The royalists rose in May 1648, igniting the Second Civil War, and as agreed with Charles, the Scots invaded England. Uprisings in Kent, Essex, and Cumberland, and a rebellion in South Wales, were put down by the New Model Army, and with the defeat of the Scots at the Battle of Preston in August 1648, the royalists lost any chance of winning the war. Charles's only recourse was to return to negotiations, which were held at Newport on the Isle of Wight. On 5 December 1648, Parliament voted by 129 to 83 to continue negotiating with the king, but Oliver Cromwell and the army opposed any further talks with someone they viewed as a bloody tyrant and were already taking action to consolidate their power. Hammond was replaced as Governor of the Isle of Wight on 27 November, and placed in the custody of the army the following day. In Pride's Purge on 6 and 7 December, the members of Parliament out of sympathy with the military were arrested or excluded by Colonel Thomas Pride, while others stayed away voluntarily. The remaining members formed the Rump Parliament. It was effectively a military coup. Charles was moved to Hurst Castle at the end of 1648, and thereafter to Windsor Castle. In January 1649, the Rump House of Commons indicted him on a charge of treason, which was rejected by the House of Lords. The idea of trying a king was a novel one. The Chief Justices of the three common law courts of England – Henry Rolle, Oliver St John and John Wilde – all opposed the indictment as unlawful. The Rump Commons declared itself capable of legislating alone, passed a bill creating a separate court for Charles's trial, and declared the bill an act without the need for royal assent. The High Court of Justice established by the Act consisted of 135 commissioners, but many either refused to serve or chose to stay away. Only 68 (all firm Parliamentarians) attended Charles's trial on charges of high treason and "other high crimes" that began on 20 January 1649 in Westminster Hall. John Bradshaw acted as President of the Court, and the prosecution was led by the Solicitor General, John Cook. Charles was accused of treason against England by using his power to pursue his personal interest rather than the good of the country. The charge stated that he, "for accomplishment of such his designs, and for the protecting of himself and his adherents in his and their wicked practices, to the same ends hath traitorously and maliciously levied war against the present Parliament, and the people therein represented", and that the "wicked designs, wars, and evil practices of him, the said Charles Stuart, have been, and are carried on for the advancement and upholding of a personal interest of will, power, and pretended prerogative to himself and his family, against the public interest, common right, liberty, justice, and peace of the people of this nation." Presaging the modern concept of command responsibility, the indictment held him "guilty of all the treasons, murders, rapines, burnings, spoils, desolations, damages and mischiefs to this nation, acted and committed in the said wars, or occasioned thereby." An estimated 300,000 people, or 6% of the population, died during the war. Over the first three days of the trial, whenever Charles was asked to plead, he refused, stating his objection with the words: "I would know by what power I am called hither, by what lawful authority...?" He claimed that no court had jurisdiction over a monarch, that his own authority to rule had been given to him by God and by the traditional laws of England, and that the power wielded by those trying him was only that of force of arms. Charles insisted that the trial was illegal, explaining that, The court, by contrast, challenged the doctrine of sovereign immunity and proposed that "the King of England was not a person, but an office whose every occupant was entrusted with a limited power to govern 'by and according to the laws of the land and not otherwise'." At the end of the third day, Charles was removed from the court, which then heard over 30 witnesses against the king in his absence over the next two days, and on 26 January condemned him to death. The following day, the king was brought before a public session of the commission, declared guilty, and sentenced. Fifty-nine of the commissioners signed Charles's death warrant. Charles's beheading was scheduled for Tuesday, 30 January 1649. Two of his children remained in England under the control of the Parliamentarians: Elizabeth and Henry. They were permitted to visit him on 29 January, and he bade them a tearful farewell. The following morning, he called for two shirts to prevent the cold weather causing any noticeable shivers that the crowd could have mistaken for fear: "the season is so sharp as probably may make me shake, which some observers may imagine proceeds from fear. I would have no such imputation." He walked under guard from St James's Palace, where he had been confined, to the Palace of Whitehall, where an execution scaffold had been erected in front of the Banqueting House. Charles was separated from spectators by large ranks of soldiers, and his last speech reached only those with him on the scaffold. He blamed his fate on his failure to prevent the execution of his loyal servant Strafford: "An unjust sentence that I suffered to take effect, is punished now by an unjust sentence on me." He declared that he had desired the liberty and freedom of the people as much as any, "but I must tell you that their liberty and freedom consists in having government ... It is not their having a share in the government; that is nothing appertaining unto them. A subject and a sovereign are clean different things." He continued, "I shall go from a corruptible to an incorruptible Crown, where no disturbance can be." At about 2:00 p.m., Charles put his head on the block after saying a prayer and signalled the executioner when he was ready by stretching out his hands; he was then beheaded with one clean stroke. According to observer Philip Henry, a moan "as I never heard before and desire I may never hear again" rose from the assembled crowd, some of whom then dipped their handkerchiefs in the king's blood as a memento. The executioner was masked and disguised, and there is debate over his identity. The commissioners approached Richard Brandon, the common hangman of London, but he refused, at least at first, despite being offered £200. It is possible he relented and undertook the commission after being threatened with death, but there are others who have been named as potential candidates, including George Joyce, William Hulet and Hugh Peters. The clean strike, confirmed by an examination of the king's body at Windsor in 1813, suggests that the execution was carried out by an experienced headsman. It was common practice for the severed head of a traitor to be held up and exhibited to the crowd with the words "Behold the head of a traitor!" Although Charles's head was exhibited, the words were not used, possibly because the executioner did not want his voice recognised. On the day after the execution, the king's head was sewn back onto his body, which was then embalmed and placed in a lead coffin. The commission refused to allow Charles's burial at Westminster Abbey, so his body was conveyed to Windsor on the night of 7 February. He was buried in private on 9 February 1649 in the Henry VIII vault in the chapel's quire, alongside the coffins of Henry VIII and Henry's third wife, Jane Seymour, in St George's Chapel, Windsor Castle. The king's son, Charles II, later planned for an elaborate royal mausoleum to be erected in Hyde Park, London, but it was never built. Ten days after Charles's execution, on the day of his interment, a memoir purporting to be written by the king appeared for sale. This book, the "Eikon Basilike" (Greek for the "Royal Portrait"), contained an "apologia" for royal policies, and it proved an effective piece of royalist propaganda. John Milton wrote a Parliamentary rejoinder, the "Eikonoklastes" ("The Iconoclast"), but the response made little headway against the pathos of the royalist book. Anglicans and royalists fashioned an image of martyrdom, and in the Convocations of Canterbury and York of 1660 King Charles the Martyr was added to the Church of England's liturgical calendar. High church Anglicans held special services on the anniversary of his death. Churches, such as those at Falmouth and Tunbridge Wells, and Anglican devotional societies such as the Society of King Charles the Martyr, were founded in his honour. With the monarchy overthrown, England became a republic or "Commonwealth". The House of Lords was abolished by the Rump Commons, and executive power was assumed by a Council of State. All significant military opposition in Britain and Ireland was extinguished by the forces of Oliver Cromwell in the Third English Civil War and the Cromwellian conquest of Ireland. Cromwell forcibly disbanded the Rump Parliament in 1653, thereby establishing the Protectorate with himself as Lord Protector. Upon his death in 1658, he was briefly succeeded by his ineffective son, Richard. Parliament was reinstated, and the monarchy was restored to Charles I's eldest son, Charles II, in 1660. Partly inspired by his visit to the Spanish court in 1623, Charles became a passionate and knowledgeable art collector, amassing one of the finest art collections ever assembled. In Spain, he sat for a sketch by Velázquez, and acquired works by Titian and Correggio, among others. In England, his commissions included the ceiling of the Banqueting House, Whitehall, by Rubens and paintings by other artists from the Low Countries such as van Honthorst, Mytens, and van Dyck. His close associates, including the Duke of Buckingham and the Earl of Arundel, shared his interest and have been dubbed the Whitehall Group. In 1627 and 1628, Charles purchased the entire collection of the Duke of Mantua, which included work by Titian, Correggio, Raphael, Caravaggio, del Sarto and Mantegna. His collection grew further to encompass Bernini, Bruegel, da Vinci, Holbein, Hollar, Tintoretto and Veronese, and self-portraits by both Dürer and Rembrandt. By Charles's death, there were an estimated 1,760 paintings, most of which were sold and dispersed by Parliament. In the words of John Philipps Kenyon, "Charles Stuart is a man of contradictions and controversy". Revered by high Tories who considered him a saintly martyr, he was condemned by Whig historians, such as Samuel Rawson Gardiner, who thought him duplicitous and delusional. In recent decades, most historians have criticised him, the main exception being Kevin Sharpe who offered a more sympathetic view of Charles that has not been widely adopted. While Sharpe argued that the king was a dynamic man of conscience, Professor Barry Coward thought Charles "was the most incompetent monarch of England since Henry VI", a view shared by Ronald Hutton, who called him "the worst king we have had since the Middle Ages". Archbishop William Laud, who was beheaded by Parliament during the war, described Charles as "A mild and gracious prince who knew not how to be, or how to be made, great." Charles was more sober and refined than his father, but he was intransigent. He deliberately pursued unpopular policies that ultimately brought ruin on himself. Both Charles and James were advocates of the divine right of kings, but while James's ambitions concerning absolute prerogative were tempered by compromise and consensus with his subjects, Charles believed that he had no need to compromise or even to explain his actions. He thought he was answerable only to God. "Princes are not bound to give account of their actions," he wrote, "but to God alone". The official style of Charles I as king in England was "Charles, by the Grace of God, King of England, Scotland, France and Ireland, Defender of the Faith, etc." The style "of France" was only nominal, and was used by every English monarch from Edward III to George III, regardless of the amount of French territory actually controlled. The authors of his death warrant referred to him as "Charles Stuart, King of England". As Duke of York, Charles bore the royal arms of the kingdom differenced by a label Argent of three points, each bearing three torteaux Gules. As the Prince of Wales, he bore the royal arms differenced by a plain label Argent of three points. As king, Charles bore the royal arms undifferenced: Quarterly, I and IV Grandquarterly, Azure three fleurs-de-lis Or (for France) and Gules three lions passant guardant in pale Or (for England); II Or a lion rampant within a tressure flory-counter-flory Gules (for Scotland); III Azure a harp Or stringed Argent (for Ireland). In Scotland, the Scottish arms were placed in the first and fourth quarters with the English and French arms in the second quarter. Charles had nine children, two of whom eventually succeeded as king, and two of whom died at or shortly after birth.
https://en.wikipedia.org/wiki?curid=7426
Counter-Strike (video game) Counter-Strike (also known as Half-Life: Counter-Strike) is a first-person shooter video game developed by Valve. It was initially developed and released as a "Half-Life" modification by Minh "Gooseman" Le and Jess Cliffe in 1999, before Le and Cliffe were hired and the game's intellectual property acquired. "Counter-Strike" was released by Valve for Microsoft Windows in 2000. The game is the first installment in the "Counter-Strike" series. Several remakes and ports have been released on Xbox, as well as OS X and Linux. Set in various locations around the globe, players assume the roles of counter-terrorist forces and terrorist militants opposing them. During each round of gameplay, the two teams are tasked with defeating the other by the means of either achieving the map's objectives or eliminating all of the enemy combatants. Each player may customize their arsenal of weapons and accessories at the beginning of every match, with currency being earned after the end of each round. "Counter-Strike" is a first-person shooter game in which players join either the terrorist team, the counter-terrorist team, or become spectators. Each team attempts to complete their mission objective and/or eliminate the opposing team. Each round starts with the two teams spawning simultaneously. All players have only one life by default and start with a pistol as well as a knife. The objectives vary depending on the type of map, and these are the most usual ones: A player can choose to play as one of eight different default character models (four for each side, although "" added two extra models, bringing the total to ten). Players are generally given a few seconds before the round begins (known as "freeze time") to prepare and buy equipment, during which they cannot attack or move. They can return to the buy area within a set amount of time to buy more equipment (some custom maps included neutral "buy zones" that could be used by both teams). Once the round has ended, surviving players retain their equipment for use in the next round; players who were killed begin the next round with the basic default starting equipment. Standard monetary bonuses are awarded for winning a round, losing a round, killing an enemy, being the first to instruct a hostage to follow, rescuing a hostage, planting the bomb (Terrorist) or defusing the bomb (Counter-Terrorist). The scoreboard displays team scores in addition to statistics for each player: name, kills, deaths, and ping (in milliseconds). The scoreboard also indicates whether a player is dead, carrying the bomb (on bomb maps), or is the VIP (on assassination maps), although information on players on the opposing team is hidden from a player until their death, as this information can be important. Killed players become "spectators" for the duration of the round; they cannot change their names before their next spawn, text chat cannot be sent to or received from live players, and voice chat can only be received from live players and not sent to them. Spectators are generally able to watch the rest of the round from multiple selectable views, although some servers disable some of these views to prevent dead players from relaying information about living players to their teammates through alternative media (most notably voice in the case of Internet cafes and Voice over IP programs such as TeamSpeak or Ventrilo). This form of cheating is known as "ghosting." "Counter-Strike" began as a mod of "Half-Life"s engine GoldSrc. Minh Le, the mod's co-creator, had started his last semester at university, and wanted to do something in game development to help give him better job prospects. Throughout university, Le had worked on mods with the "Quake" engine, and on looking for this latest project, wanted to try something new and opted for GoldSrc. At the onset, Valve had not yet released the software development kit (SDK) for GoldSrc but affirmed it would be available in a few months, allowing Le to work on the character models in the interim. Once the GoldSrc SDK was available, Le estimated it took him about a month and a half to complete the programming and integrate his models for "Beta One" of "Counter-Strike". To assist, Le had help from Jess Cliffe who managed the game's website and community, and had contacts within level map making community to help build some of the levels for the game. The theme of countering terrorists was inspired by Le's own interest in guns and the military, and from games like "Rainbow Six" and "Spec Ops". Le and Cliffe continued to release Betas on a frequent basis for feedback. The initial few Betas, released starting in June 1999, had limited audiences but by the fifth one, interest in the project dramatically grew. The interest in the game drew numerous players to the website, which helped Le and Cliffe to make revenue from ads hosted on the site. Around 2000 at the time of Beta 5's release, the two were approached by Valve, offering to buy the "Counter-Strike" intellectual property and offering both jobs to continue its development. Both accepted the offer, and by September 2000, Valve released the first non-beta version of the game. While Cliffe stayed with Valve, Le did some additional work towards a "Counter-Strike 2.0" based on Valve's upcoming Source engine, but left to start his own studio after Valve opted to shelve the sequel. "Counter-Strike" itself is a mod, and it has developed its own community of script writers and mod creators. Some mods add bots, while others remove features of the game, and others create different modes of play. Some mods, often called "admin plugins", give server administrators more flexible and efficient control over their servers. There are some mods which affect gameplay heavily, such as Gun Game, where players start with a basic pistol and must score kills to receive better weapons, and Zombie Mod, where one team consists of zombies and must "spread the infection" by killing the other team (using only the knife). There are also Superhero mods which mix the first-person gameplay of "Counter-Strike" with an experience system, allowing a player to become more powerful as they continue to play. The game is highly customizable on the player's end, allowing the user to install or even create their own custom skins, HUDs, spray graphics, sprites, and sound effects, given the proper tools. "Counter-Strike" has been a target for cheating in online games since its release. In-game, cheating is often referred to as "hacking" in reference to programs or "hacks" executed by the client. Valve has implemented an anti-cheat system called Valve Anti-Cheat (VAC). Players cheating on a VAC-enabled server risk having their account permanently banned from all VAC-secured servers. With the first version of VAC, a ban took hold almost instantly after being detected and the cheater had to wait two years to have the account unbanned. Since VAC's second version, cheaters are not banned automatically. With the second version, Valve instituted a policy of 'delayed bans,' the theory being that if a new hack is developed which circumvents the VAC system, it will spread amongst the 'cheating' community. By delaying the initial ban, Valve hopes to identify and ban as many cheaters as possible. Like any software detection system, some cheats are not detected by VAC. To remedy this, some servers implement a voting system, in which case players can call for a vote to kick or ban the accused cheater. VAC's success at identifying cheats and banning those who use them has also provided a boost in the purchasing of private cheats. These cheats are updated frequently to minimize the risk of detection, and are generally only available to a trusted list of recipients who collectively promise not to reveal the underlying design. Even with private cheats however, some servers have alternative anticheats to coincide with VAC itself. This can help with detecting some cheaters, but most paid for cheats are designed to bypass these alternative server-based anticheats. When "Counter-Strike" was published by Sierra Studios, it was bundled with "Team Fortress Classic", "" multiplayer, and the "Wanted", "Half-Life: Absolute Redemption" and "Firearms" mods. On March 24, 1999, Planet Half-Life opened its "Counter-Strike" section. Within two weeks, the site had received 10,000 hits. On June 19, 1999, the first public beta of "Counter-Strike" was released, followed by numerous further "beta" releases. On April 12, 2000, Valve announced that the "Counter-Strike" developers and Valve had teamed up. In January 2013, Valve began testing a version of "Counter-Strike" for OS X and Linux, eventually releasing the update to all users in April 2013. Upon its retail release, "Counter-Strike" received highly favorable reviews. "The New York Times" reported that E-Sports Entertainment ESEA League started the first professional fantasy e-sports league in 2004 with the game "Counter-Strike". Some credit the move into professional competitive team play with prizes as a major factor in "Counter-Strike" longevity and success. Global retail sales of "Counter-Strike" surpassed 250,000 units by July 2001, and 1.5 million by December 2003. In the United States, its retail version sold 550,000 copies and earned $15.7 million by August 2006, after its release in November 2000. It was the country's 22nd best-selling PC game between January 2000 and August 2006. The Xbox version sold 1.5 million copies in total. On January 17, 2008, a Brazilian federal court order prohibiting all sales of "Counter-Strike" and "EverQuest" began to be enforced. The federal Brazilian judge Carlos Alberto Simões de Tomaz ordered the ban in October 2007 because, as argued by the judge, the games "bring imminent stimulus to the subversion of the social order, attempting against the democratic state and the law and against public security." As of June 18, 2009, a regional federal court order lifting the prohibition on the sale of "Counter-Strike" was published. The game is now being sold again in Brazil. The original "Counter-Strike" has been played in tournaments since 2000 with the first major being hosted in 2001 at the Cyberathlete Professional League Winter Championship. The first official sequel was "" released on November 1, 2004. The game was criticized by the competitive community, who believed the game's skill ceiling was significantly lower than that of CS 1.6. This caused a divide in the competitive community as to which game to play competitively. Following the success of the first "Counter-Strike", Valve went on to make multiple sequels to the game. "", a game using "Counter-Strike"s GoldSrc engine, was released in 2004. "", a remake of the original "Counter-Strike", was the first in the series to use Valve's Source engine and was also released in 2004, eight months after the release of "Counter-Strike: Condition Zero". The next game in the "Counter-Strike" series to be developed primarily by Valve was "", released for Windows, OS X, Linux, PlayStation 3, and Xbox 360 in 2012. The game spawned multiple spin-offs in the form of arcade games developed by Nexon Corporation and targeted primarily at Asian gaming markets. Four "Counter-Strike" games have been developed and released by Nexon Corporation thus far, "Counter-Strike Neo", "Counter-Strike Online", "", and "Counter-Strike Online 2".
https://en.wikipedia.org/wiki?curid=7431
Camille Pissarro Camille Pissarro ( , ; 10 July 1830 – 13 November 1903) was a Danish-French Impressionist and Neo-Impressionist painter born on the island of St Thomas (now in the US Virgin Islands, but then in the Danish West Indies). His importance resides in his contributions to both Impressionism and Post-Impressionism. Pissarro studied from great forerunners, including Gustave Courbet and Jean-Baptiste-Camille Corot. He later studied and worked alongside Georges Seurat and Paul Signac when he took on the Neo-Impressionist style at the age of 54. In 1873 he helped establish a collective society of fifteen aspiring artists, becoming the "pivotal" figure in holding the group together and encouraging the other members. Art historian John Rewald called Pissarro the "dean of the Impressionist painters", not only because he was the oldest of the group, but also "by virtue of his wisdom and his balanced, kind, and warmhearted personality". Paul Cézanne said "he was a father for me. A man to consult and a little like the good Lord", and he was also one of Paul Gauguin's masters. Pierre-Auguste Renoir referred to his work as "revolutionary", through his artistic portrayals of the "common man", as Pissarro insisted on painting individuals in natural settings without "artifice or grandeur". Pissarro is the only artist to have shown his work at all eight Paris Impressionist exhibitions, from 1874 to 1886. He "acted as a father figure not only to the Impressionists" but to all four of the major Post-Impressionists, Georges Seurat, Paul Cézanne, Vincent van Gogh, and Paul Gauguin. Jacob Abraham Camille Pissarro was born on 10 July 1830 on the island of St. Thomas to Frederick and Rachel Manzano de Pissarro. His father was of Portuguese Jewish descent and held French nationality. His mother was from a French-Jewish family from the island of St. Thomas. His father was a merchant who came to the island from France to deal with the hardware store of a deceased uncle, Isaac Petit, and married his widow. The marriage caused a stir within St. Thomas' small Jewish community because she was previously married to Frederick's uncle and according to Jewish law a man is forbidden from marrying his aunt. In subsequent years his four children attended the all-black primary school. Upon his death, his will specified that his estate be split equally between the synagogue and St. Thomas' Protestant church. When Camille was twelve his father sent him to boarding school in France. He studied at the Savary Academy in Passy near Paris. While a young student, he developed an early appreciation of the French art masters. Monsieur Savary himself gave him a strong grounding in drawing and painting and suggested he draw from nature when he returned to St. Thomas, which he did when he was seventeen. However, his father preferred he work in his business, giving him a job working as a cargo clerk. He took every opportunity during those next five years at the job to practise drawing during breaks and after work. When Pissarro turned twenty-one, Danish artist Fritz Melbye, then living on St. Thomas, inspired him to take on painting as a full-time profession, becoming his teacher and friend. Pissarro then chose to leave his family and job and live in Venezuela, where he and Melbye spent the next two years working as artists in Caracas and La Guaira. He drew everything he could, including landscapes, village scenes, and numerous sketches, enough to fill up multiple sketchbooks. In 1855 he moved back to Paris where he began working as assistant to Anton Melbye, Fritz Melbye's brother. In Paris he worked as assistant to Danish painter Anton Melbye. He also studied paintings by other artists whose style impressed him: Courbet, Charles-François Daubigny, Jean-François Millet, and Corot. He also enrolled in various classes taught by masters, at schools such as École des Beaux-Arts and Académie Suisse. But Pissarro eventually found their teaching methods "stifling," states art historian John Rewald. This prompted him to search for alternative instruction, which he requested and received from Corot. His initial paintings were in accord with the standards at the time to be displayed at the Paris Salon, the official body whose academic traditions dictated the kind of art that was acceptable. The Salon's annual exhibition was essentially the only marketplace for young artists to gain exposure. As a result, Pissarro worked in the traditional and prescribed manner to satisfy the tastes of its official committee. In 1859 his first painting was accepted and exhibited. His other paintings during that period were influenced by Camille Corot, who tutored him. He and Corot both shared a love of rural scenes painted from nature. It was by Corot that Pissarro was inspired to paint outdoors, also called "plein air" painting. Pissarro found Corot, along with the work of Gustave Courbet, to be "statements of pictorial truth," writes Rewald. He discussed their work often. Jean-François Millet was another whose work he admired, especially his "sentimental renditions of rural life". During this period Pissarro began to understand and appreciate the importance of expressing on canvas the beauties of nature without adulteration. After a year in Paris, he therefore began to leave the city and paint scenes in the countryside to capture the daily reality of village life. He found the French countryside to be "picturesque," and worthy of being painted. It was still mostly agricultural and sometimes called the "golden age of the peasantry". Pissarro later explained the technique of painting outdoors to a student: Corot would complete his paintings back in his studio, often revising them according to his preconceptions. Pissarro, however, preferred to finish his paintings outdoors, often at one sitting, which gave his work a more realistic feel. As a result, his art was sometimes criticised as being "vulgar," because he painted what he saw: "rutted and edged hodgepodge of bushes, mounds of earth, and trees in various stages of development." According to one source, such details were equivalent to today's art showing garbage cans or beer bottles on the side of a street. This difference in style created disagreements between Pissarro and Corot. In 1859, while attending the free school, the Académie Suisse, Pissarro became friends with a number of younger artists who likewise chose to paint in the more realistic style. Among them were Claude Monet, Armand Guillaumin and Paul Cézanne. What they shared in common was their dissatisfaction with the dictates of the Salon. Cézanne's work had been mocked at the time by the others in the school, and, writes Rewald, in his later years Cézanne "never forgot the sympathy and understanding with which Pissarro encouraged him." As a part of the group, Pissarro was comforted from knowing he was not alone, and that others similarly struggled with their art. Pissarro agreed with the group about the importance of portraying individuals in natural settings, and expressed his dislike of any artifice or grandeur in his works, despite what the Salon demanded for its exhibits. In 1863 almost all of the group's paintings were rejected by the Salon, and French Emperor Napoleon III instead decided to place their paintings in a separate exhibit hall, the Salon des Refusés. However, only works of Pissarro and Cézanne were included, and the separate exhibit brought a hostile response from both the officials of the Salon and the public. In subsequent Salon exhibits of 1865 and 1866, Pissarro acknowledged his influences from Melbye and Corot, whom he listed as his masters in the catalogue. But in the exhibition of 1868 he no longer credited other artists as an influence, in effect declaring his independence as a painter. This was noted at the time by art critic and author Émile Zola, who offered his opinion: Another writer tries to describe elements of Pissarro's style: And though, on orders from the hanging Committee and the Marquis de Chennevières, Pissarro's paintings of Pontoise for example had been skyed, hung near the ceiling, this did not prevent Jules-Antoine Castagnary from noting that the qualities of his paintings had been observed by art lovers. At the age of thirty-eight, Pissarro had begun to win himself a reputation as a landscapist to rival Corot and Daubigny. In the late 1860s or early 1870s, Pissarro became fascinated with Japanese prints, which influenced his desire to experiment in new compositions. He described the art to his son Lucien: In 1871 in Croydon he married his mother's maid, Julie Vellay, a vineyard grower's daughter, with whom he would later have seven children. They lived outside Paris in Pontoise and later in Louveciennes, both of which places inspired many of his paintings including scenes of village life, along with rivers, woods, and people at work. He also kept in touch with the other artists of his earlier group, especially Monet, Renoir, Cézanne, and Frédéric Bazille. After the outbreak of the Franco-Prussian War of 1870–71, having only Danish nationality and being unable to join the army, he moved his family to Norwood, then a village on the edge of London. However, his style of painting, which was a forerunner of what was later called "Impressionism", did not do well. He wrote to his friend, Théodore Duret, that "my painting doesn't catch on, not at all ..." Pissarro met the Paris art dealer Paul Durand-Ruel, in London, who became the dealer who helped sell his art for most of his life. Durand-Ruel put him in touch with Monet who was likewise in London during this period. They both viewed the work of British landscape artists John Constable and J. M. W. Turner, which confirmed their belief that their style of open air painting gave the truest depiction of light and atmosphere, an effect that they felt could not be achieved in the studio alone. Pissarro's paintings also began to take on a more spontaneous look, with loosely blended brushstrokes and areas of impasto, giving more depth to the work. Through the paintings Pissarro completed at this time, he records Sydenham and the Norwoods at a time when they were just recently connected by railways, but prior to the expansion of suburbia. One of the largest of these paintings is a view of "St. Bartholomew's Church" at Lawrie Park Avenue, commonly known as "The Avenue, Sydenham", in the collection of the National Gallery in London. Twelve oil paintings date from his stay in Upper Norwood and are listed and illustrated in the catalogue raisonné prepared jointly by his fifth child Ludovic-Rodolphe Pissarro and Lionello Venturi and published in 1939. These paintings include "Norwood Under the Snow", and "Lordship Lane Station", views of The Crystal Palace relocated from Hyde Park, "Dulwich College", "Sydenham Hill", "All Saints Church Upper Norwood", and a lost painting of St. Stephen's Church. Returning to France, Pissarro lived in Pontoise from 1872 to 1884. In 1890 he again visited England and painted some ten scenes of central London. He came back again in 1892, painting in Kew Gardens and Kew Green, and also in 1897, when he produced several oils described as being of Bedford Park, Chiswick, but in fact all being of the nearby Stamford Brook area except for one of Bath Road, which runs from Stamford Brook along the south edge of Bedford Park. When Pissarro returned to his home in France after the war, he discovered that of the 1,500 paintings he had done over 20 years, which he was forced to leave behind when he moved to London, only 40 remained. The rest had been damaged or destroyed by the soldiers, who often used them as floor mats outside in the mud to keep their boots clean. It is assumed that many of those lost were done in the Impressionist style he was then developing, thereby "documenting the birth of Impressionism." Armand Silvestre, a critic, went so far as to call Pissarro "basically the inventor of this [Impressionist] painting"; however, Pissarro's role in the Impressionist movement was "less that of the great man of ideas than that of the good counselor and appeaser ..." "Monet ... could be seen as the guiding force." He soon reestablished his friendships with the other Impressionist artists of his earlier group, including Cézanne, Monet, Manet, Renoir, and Degas. Pissarro now expressed his opinion to the group that he wanted an alternative to the Salon so their group could display their own unique styles. To assist in that endeavour, in 1873 he helped establish a separate collective, called the "Société Anonyme des Artistes, Peintres, Sculpteurs et Graveurs," which included fifteen artists. Pissarro created the group's first charter and became the "pivotal" figure in establishing and holding the group together. One writer noted that with his prematurely grey beard, the forty-three-year-old Pissarro was regarded as a "wise elder and father figure" by the group. Yet he was able to work alongside the other artists on equal terms due to his youthful temperament and creativity. Another writer said of him that "he has unchanging spiritual youth and the look of an ancestor who remained a young man". The following year, in 1874, the group held their first 'Impressionist' Exhibition, which shocked and "horrified" the critics, who primarily appreciated only scenes portraying religious, historical, or mythological settings. They found fault with the Impressionist paintings on many grounds: Pissarro showed five of his paintings, all landscapes, at the exhibit, and again Émile Zola praised his art and that of the others. In the Impressionist exhibit of 1876; however, art critic Albert Wolff complained in his review, "Try to make M. Pissarro understand that trees are not violet, that sky is not the color of fresh butter ..." Journalist and art critic Octave Mirbeau on the other hand, writes, "Camille Pissarro has been a revolutionary through the revitalized working methods with which he has endowed painting". According to Rewald, Pissarro had taken on an attitude more simple and natural than the other artists. He writes: In later years, Cézanne also recalled this period and referred to Pissarro as "the first Impressionist". In 1906, a few years after Pissarro's death, Cézanne, then 67 and a role model for the new generation of artists, paid Pissarro a debt of gratitude by having himself listed in an exhibition catalogue as "Paul Cézanne, pupil of Pissarro". Pissarro, Degas, and American impressionist Mary Cassatt planned a journal of their original prints in the late 1870s, a project that nevertheless came to nothing when Degas withdrew. Art historian and the artist's great-grandson Joachim Pissarro notes that they "professed a passionate disdain for the Salons and refused to exhibit at them." Together they shared an "almost militant resolution" against the Salon, and through their later correspondences it is clear that their mutual admiration "was based on a kinship of ethical as well as aesthetic concerns". Cassatt had befriended Degas and Pissarro years earlier when she joined Pissarro's newly formed French Impressionist group and gave up opportunities to exhibit in the United States. She and Pissarro were often treated as "two outsiders" by the Salon since neither were French or had become French citizens. However, she was "fired up with the cause" of promoting Impressionism and looked forward to exhibiting "out of solidarity with her new friends". Towards the end of the 1890s she began to distance herself from the Impressionists, avoiding Degas at times as she did not have the strength to defend herself against his "wicked tongue". Instead, she came to prefer the company of "the gentle Camille Pissarro", with whom she could speak frankly about the changing attitudes toward art. She once described him as a teacher "that could have taught the stones to draw correctly." By the 1880s, Pissarro began to explore new themes and methods of painting to break out of what he felt was an artistic "mire". As a result, Pissarro went back to his earlier themes by painting the life of country people, which he had done in Venezuela in his youth. Degas described Pissarro's subjects as "peasants working to make a living". However, this period also marked the end of the Impressionist period due to Pissarro's leaving the movement. As Joachim Pissarro points out, "Once such a die-hard Impressionist as Pissarro had turned his back on Impressionism, it was apparent that Impressionism had no chance of surviving ..." It was Pissarro's intention during this period to help "educate the public" by painting people at work or at home in realistic settings, without idealising their lives. Pierre-Auguste Renoir, in 1882, referred to Pissarro's work during this period as "revolutionary," in his attempt to portray the "common man." Pissarro himself did not use his art to overtly preach any kind of political message, however, although his preference for painting humble subjects was intended to be seen and purchased by his upper class clientele. He also began painting with a more unified brushwork along with pure strokes of color. In 1885 he met Georges Seurat and Paul Signac, both of whom relied on a more "scientific" theory of painting by using very small patches of pure colours to create the illusion of blended colours and shading when viewed from a distance. Pissarro then spent the years from 1885 to 1888 practising this more time-consuming and laborious technique, referred to as pointillism. The paintings that resulted were distinctly different from his Impressionist works, and were on display in the 1886 Impressionist Exhibition, but under a separate section, along with works by Seurat, Signac, and his son Lucien. All four works were considered an "exception" to the eighth exhibition. Joachim Pissarro notes that virtually every reviewer who commented on Pissarro's work noted "his extraordinary capacity to change his art, revise his position and take on new challenges." One critic writes: Pissarro explained the new art form as a "phase in the logical march of Impressionism", but he was alone among the other Impressionists with this attitude, however. Joachim Pissarro states that Pissarro thereby became the "only artist who went from Impressionism to Neo-Impressionism". In 1884, art dealer Theo van Gogh asked Pissarro if he would take in his older brother, Vincent, as a boarder in his home. Lucien Pissarro wrote that his father was impressed by Van Gogh's work and had "foreseen the power of this artist", who was 23 years younger. Although Van Gogh never boarded with him, Pissarro did explain to him the various ways of finding and expressing light and color, ideas which he later used in his paintings, notes Lucien. Pissarro eventually turned away from Neo-Impressionism, claiming its system was too artificial. He explains in a letter to a friend: However, after reverting to his earlier style, his work became, according to Rewald, "more subtle, his color scheme more refined, his drawing firmer ... So it was that Pissarro approached old age with an increased mastery." But the change also added to Pissarro's continual financial hardship which he felt until his 60s. His "headstrong courage and a tenacity to undertake and sustain the career of an artist", writes Joachim Pissarro, was due to his "lack of fear of the immediate repercussions" of his stylistic decisions. In addition, his work was strong enough to "bolster his morale and keep him going", he writes. His Impressionist contemporaries, however, continued to view his independence as a "mark of integrity", and they turned to him for advice, referring to him as "Père Pissarro" (father Pissarro). In his older age Pissarro suffered from a recurring eye infection that prevented him from working outdoors except in warm weather. As a result of this disability, he began painting outdoor scenes while sitting by the window of hotel rooms. He often chose hotel rooms on upper levels to get a broader view. He moved around northern France and painted from hotels in Rouen, Paris, Le Havre and Dieppe. On his visits to London, he would do the same. Pissarro died in Paris on 13 November 1903 and was buried in Père Lachaise Cemetery. During the period Pissarro exhibited his works, art critic Armand Silvestre had called Pissarro the "most real and most naive member" of the Impressionist group. His work has also been described by art historian Diane Kelder as expressing "the same quiet dignity, sincerity, and durability that distinguished his person." She adds that "no member of the group did more to mediate the internecine disputes that threatened at times to break it apart, and no one was a more diligent proselytizer of the new painting." According to Pissarro's son, Lucien, his father painted regularly with Cézanne beginning in 1872. He recalls that Cézanne walked a few miles to join Pissarro at various settings in Pontoise. While they shared ideas during their work, the younger Cézanne wanted to study the countryside through Pissarro's eyes, as he admired Pissarro's landscapes from the 1860s. Cézanne, although only nine years younger than Pissarro, said that "he was a father for me. A man to consult and a little like the good Lord." Lucien Pissarro was taught painting by his father, and described him as a "splendid teacher, never imposing his personality on his pupil." Gauguin, who also studied under him, referred to Pissarro "as a force with which future artists would have to reckon". Art historian Diane Kelder notes that it was Pissarro who introduced Gauguin, who was then a young stockbroker studying to become an artist, to Degas and Cézanne. Gauguin, near the end of his career, wrote a letter to a friend in 1902, shortly before Pissarro's death: The American impressionist Mary Cassatt, who at one point lived in Paris to study art, and joined his Impressionist group, noted that he was "such a teacher that he could have taught the stones to draw correctly." Caribbean author and scholar Derek Walcott based his book-length poem, "Tiepolo's Hound" (2000), on Pissarro's life. During the early 1930s throughout Europe, Jewish owners of numerous fine art masterpieces found themselves forced to give up or sell off their collections for minimal prices due to anti-Jewish laws created by the new Nazi regime. Many Jews were forced to flee Germany. When those forced into exile owned valuables, including artwork, they were often seized by officials for personal gain. In the decades after World War II, many art masterpieces were found on display in various galleries and museums in Europe and the United States. Some, as a result of legal action, were later returned to the families of the original owners. Many of the recovered paintings were then donated to the same or other museums as a gift. One such lost piece, Pissarro's 1897 oil painting, "Rue St. Honoré, Apres Midi, Effet de Pluie", was discovered hanging at Madrid's government-owned museum, the Museo Thyssen-Bornemisza. In January 2011 the Spanish government denied a request by the US ambassador to return the painting. At the subsequent trial in Los Angeles, the court ruled that the Thyssen-Bornemisza Collection Foundation was the rightful owner. Pissarro's "Le Quai Malaquais, Printemps" is said to have been similarly stolen, while in 1999, Pissarro's 1897 "Le Boulevard de Montmartre, Matinée de Printemps" appeared in the Israel Museum in Jerusalem, its donor having been unaware of its pre-war provenance. In January 2012, "Le Marché aux Poissons" (The Fish Market), a color monotype, was returned after 30 years. During his lifetime, Camille Pissarro sold few of his paintings. By the 21st century, however, his paintings were selling for millions. An auction record for the artist was set on 6 November 2007 at Christie's in New York, where a group of four paintings, "Les Quatre Saisons" (the Four Seasons), sold for $14,601,000 (estimate $12,000,000 – $18,000,000). In November 2009 "Le Pont Boieldieu et la Gare d'Orléans, Rouen, Soleil" sold for $7,026,500 at Sotheby's in New York. In February 2014 the 1897 "Le Boulevard de Montmartre, Matinée de Printemps", originally owned by the German industrialist and Holocaust victim Max Silberberg (), sold at Sotheby's in London for £19.9M, nearly five times the previous record. Camille's son Lucien was an Impressionist and Neo-impressionist painter as were his second and third sons Georges Henri Manzana Pissarro and Félix Pissarro. Lucien's daughter Orovida Pissarro was also a painter. Camille's great-grandson, Joachim Pissarro, became Head Curator of Drawing and Painting at the Museum of Modern Art in New York City and a professor in Hunter College's Art Department. Camille's great-granddaughter, Lélia Pissarro, has had her work exhibited alongside her great-grandfather. From the only daughter of Camille, Jeanne Pissarro, other painters include Henri Bonin-Pissarro (1918–2003) and Claude Bonin-Pissarro (born 1921), who is the father of the Abstract artist Frédéric Bonin-Pissarro (born 1964). In June 2006 publishers Skira/Wildenstein released "Pissarro: Critical Catalogue of Paintings", compiled by Joachim Pissarro (descendant of the painter) and Claire Durand-Ruel Snollaerts (descendant of the French art dealer Paul Durand-Ruel). The 1,500-page, three-volume work is the most comprehensive collection of Pissarro paintings to date, and contains accompanying images of drawings and studies, as well as photographs of Pissarro and his family that had not previously been published.
https://en.wikipedia.org/wiki?curid=7434
Cardiology diagnostic tests and procedures The diagnostic tests in cardiology are methods of identifying heart conditions associated with healthy vs. unhealthy, pathologic heart function. Obtaining a medical history is always the first "test", part of understanding the likelihood of significant disease, as detectable within the current limitations of clinical medicine. Yet heart problems often produce no symptoms until very advanced, and many symptoms, such as palpitations and sensations of extra or missing heart beats correlate poorly with relative heart health "vs" disease. Hence, a history alone is rarely sufficient to diagnose a heart condition. "Auscultation" employs a stethoscope to more easily hear various normal and abnormal sounds, such as normal heart beat sounds and the usual heart beat sound changes associated with breathing versus heart murmurs. A variety of "blood tests" are available for analyzing cholesterol transport behavior, HDL, LDL, triglycerides, lipoprotein little a, homocysteine, C-reactive protein, blood sugar control: fasting, after eating or averages using glycosylated albumen or hemoglobin, myoglobin, creatine kinase, troponin, brain-type natriuretic peptide, etc. to assess the evolution of coronary artery disease and evidence of existing damage. A great many more physiologic markers related to atherosclerosis and heart function are used and being developed and evaluated in research. (*) due to the high cost, LDL is usually calculated instead of being measured directly source: Beyond Cholesterol, Julius Torelli MD, 2005 "Electrocardiography" (ECG/EKG in German vernacular. Elektrokardiogram) monitors electrical activity of the heart, primarily as recorded from the skin surface. A 12 lead recording, recording the electrical activity in three planes, anterior, posterior, and lateral is the most commonly used form. The ECG allows observation of the heart electrical activity by visualizing waveform beat origin (typically from the sinoatrial or SA node) following down the bundle of HIS and ultimately stimulating the ventricles to contract forcing blood through the body. Much can be learned by observing the QRS morphology (named for the respective portions of the polarization/repolarization waveform of the wave, P,Q,R,S,T wave). Rhythm abnormalities can also be visualized as in slow heart rate bradycardia, or fast heart rate tachycardia. A "Holter monitor" records a continuous EKG rhythm pattern (rarely a full EKG) for 24 hours or more. These monitors are used for suspected frequent rhythm abnormalities, especially ones the wearer may not recognize by symptoms. They are more expensive than event monitors. An "event monitor" records short term EKG rhythm patterns, generally storing the last 2 to 5 minutes, adding in new and discarding old data, for 1 to 2 weeks or more. There are several different types with different capabilities. When the wearer presses a button on the monitor, it quits discarding old and continues recording for a short additional period. The wearer then plays the recording, via a standard phone connection, to a center with compatible receiving and rhythm printing equipment, after which the monitor is ready to record again. These monitors are used for suspected infrequent rhythm abnormalities, especially ones the wearer does recognize by symptoms. They are less expensive than Holter monitors. "Cardiac stress testing" is used to determine to assess cardiac function and to disclose evidence of exertion-related cardiac hypoxia. Radionuclide testing using thallium or technetium can be used to demonstrate areas of perfusion abnormalities. With a maximal stress test the level of exercise is increased until the person's heart rate will not increase any higher, despite increased exercise. A fairly accurate estimate of the target heart rate, based on extensive clinical research, can be estimated by the formula 220 beats per minute minus patient's age. This linear relation is accurate up to about age 30, after which it mildly underestimates typical maximum attainable heart rates achievable by healthy individuals. Other formulas exist, such as that by Miller (217 - (0.85 × Age)) and others. Achieving a high enough heart rate at the end of exercise is critical to improving the sensitivity of the test to detect high grade heart artery stenosis. High frequency analysis of the QRS complex may be useful for detection of coronary artery disease during an exercise stress test. The electrophysiology study or EP study is the end all of electrophysiological tests of the heart. It involves a catheter with electrodes probing the endocardium, the inside of the heart, and testing the conduction pathways and electrical activity of individual areas of the heart. Cardiac imaging techniques include coronary catheterization, echocardiogram, intravascular ultrasound, retinal vessel analysis and the coronary calcium scan.
https://en.wikipedia.org/wiki?curid=7435
Carlo Collodi Carlo Lorenzini (24 November 1826 – 26 October 1890), better known by the pen name Carlo Collodi (), was an Italian author, humorist, and journalist, widely known for his fairy tale novel "The Adventures of Pinocchio". Collodi was born in Florence on 24 November 1826. His mother Angiolina Orzali was a seamstress from the town of Collodi from which he took the pen name and his father Domenico Lorenzini was a cook; and both worked for the "marchese" Ginori Lisci. Carlo was the eldest child in the family and he had ten siblings but seven died at a young age. He spent most of his childhood in the town of Collodi where his mother was born. He lived there with his maternal grandmother. After attending primary school, he was sent to study at a theological seminary called Colle Val d’Elsa. An account at the seminary shows that the "marchese" had offered financial aid, but the boy found that he did not want to be a priest so he continued his education at the College of the Scolopi Fathers in Florence. In 1844 he started working at the Florentine bookstore Libreria Piatti, where he assisted Giuseppe Aiazzi, a prominent Italian manuscript specialist. During the Italian Wars of Independence in 1848 and 1860 Collodi served as a volunteer with the Tuscan army. His active interest in political matters may be seen in his earliest literary works as well as in the founding of the satirical newspaper "Il Lampione" in 1853. This newspaper was censored by order of the Grand Duke of Tuscany. In 1854 he published his second newspaper, "Lo scaramuccia" ("The Controversy"). Lorenzini's first publications were in his periodicals. A debut came in 1856 with the play "Gli amici di casa" and parodic guidebook "Un romanzo in vapore", both in 1856. By 1860, he published his first notable work called "Il signor Alberi ha ragione!" (Mr. Alberi Is Right!), which outlined his political and cultural vision of Italy. This is the text where Lorenzini started using the Collodi pseudonym, which was taken from his mother's hometown. Collodi had also begun intense activity on other political newspapers such as "Il Fanfulla"; at the same time he was employed by the Censorship Commission for the Theatre. During this period he composed various satirical sketches and stories (sometimes simply by collating earlier articles), including "Macchiette" (1880), "Occhi e nasi" (1881), "Storie allegre" (1887). Collodi became disenchanted with Italian politics afterward so he turned to children's literature and his first works involved translating French fairy tales into Italian. In 1875, for instance, he completed "Racconti delle fate", a translation of French fairy tales by Charles Perrault. In 1876 Lorenzini wrote "Giannettino" (inspired by Alessandro Luigi Parravicini's "Giannetto"), the "Minuzzolo", and "Il viaggio per l'Italia di Giannettino", a pedagogic series which explored the unification of Italy through the ironic thoughts and actions of the character Giannettino. Lorenzini became fascinated by the idea of using an amiable, rascally character as a means of expressing his own convictions through allegory. In 1880 he began writing "Storia di un burattino" ("Story of a Marionette"), also called "Le avventure di Pinocchio", which was published weekly in "Il Giornale per i Bambini", the first Italian newspaper for children. Pinocchio was adapted into a 1940 film by Disney that is considered to be one of Disney's greatest films ever made. Collodi died suddenly in Florence on 26 October 1890 and lies buried in Cimitero Monumentale Delle Porte Sante, Firenze, Toscana, Italy. The National Carlo Collodi Foundation was established to promote education and the works of Carlo Lorenzini, and the Park of Pinocchio attracts many visitors each year.
https://en.wikipedia.org/wiki?curid=7437
Constructible number In geometry and algebra, a real number is constructible if and only if, given a line segment of unit length, a line segment of length || can be constructed with compass and straightedge in a finite number of steps. Equivalently, is constructible if and only if there is a closed-form expression for using only the integers 0 and 1 and the operations for addition, subtraction, multiplication, division, and square roots. The geometric definition of constructible numbers motivates a corresponding definition of constructible points, which can again be described either geometrically or algebraically. A point is constructible if it can be produced (as an endpoint of a line segment or crossing point of two lines or circles) as one of the points of a compass and straight edge construction, starting from a given unit length segment. Alternatively and equivalently, taking the two endpoints of the segments to be the points (0,0) and (1,0) of a Cartesian coordinate system, a point is constructible if and only if its Cartesian coordinates are both constructible numbers. The set of constructible numbers forms a field: applying any of the four basic arithmetic operations to members of this set produces another constructible number. This field is a field extension of the rational numbers and in turn is contained in the field of algebraic numbers. It is the real quadratic closure of the rational numbers, the smallest field extension of the rationals that includes the square roots of all of its positive numbers. The proof of the equivalence between the algebraic and geometric definitions of constructible numbers has the effect of transforming geometric questions about compass and straightedge constructions into algebra. This transformation leads to the solutions of many famous mathematical problems, which defied centuries of attack. Let and be two given distinct points in the Euclidean plane, and define "S" to be the set of points that can be constructed with compass and straightedge starting with and . Then the points of "S" are called constructible points. and are, by definition, elements of . To more precisely describe the remaining elements of , make the following two definitions: Then, the points of , besides and are: As an example, the midpoint of constructed segment is a constructible point. One construction for it is to construct two circles with as radius, and the line through the two crossing points of these two circles. Then the midpoint of segment is the point where this segment is crossed by the constructed line. This geometric formulation can be used to define a Cartesian coordinate system in which the point is associated to the origin having coordinates and in which the point is associated with the coordinates . The points of may now be used to link the geometry and algebra by defining a constructible number to be a coordinate of a constructible point. An equivalent definition is that a constructible number is the length of a constructible line segment. If a constructible number is represented as the -coordinate of a constructible point , then the segment from to the perpendicular projection of onto line is a constructible line segment with length . And conversely, if is the length of a constructible line segment, then the intersection of line and a circle centered at with radius equal to the length of this segment gives a point whose first Cartesian coordinate is . Given any two constructible numbers and , one can construct the points and as above, as the points at distances and from along line and its perpendicular axis through . Then, the point can be constructed as the intersection of two lines perpendicular to the axes through and . Therefore, the constructible points are exactly the points whose Cartesian coordinates are constructible numbers. If and are the non-zero lengths of constructed segments then elementary compass and straightedge constructions can be used to obtain constructed segments of lengths , (if ), and . The latter two can be done with a construction based on the intercept theorem. A slightly less elementary construction using these tools is based on the geometric mean theorem and will construct a segment of length from a constructed segment of length . It follows from these considerations that the numbers defined as constructible in this geometric way include all of the numbers in the quadratic closure of the rationals. If and are constructible numbers with , then , , , and , for non-negative , are constructible. Thus, the set of constructible real numbers form a field. Furthermore, since 1 is a constructible number, all rational numbers are constructible and is a (proper) subfield of the field of constructible numbers. Also, any constructible number is an algebraic number. More precisely, if is a constructible real number and , then there is a finite sequence of real numbers such that is an extension of of degree 2. In particular, for some integer . Using slightly different terminology, a real number is constructible if and only if it lies in a field at the top of a finite tower of quadratic extensions, starting with the rational field . More precisely, is constructible if and only if there exists a tower of fields where is in and for all , . For yet another formulation of this result, this time using the geometric definition of a constructible point, let be a non-empty set of points in and the subfield of generated by all the coordinates of points in . If a point is constructible from the points of , then the degrees and are powers of 2. Using the natural correspondence between points of and complex numbers (namely, ), some authors prefer to phrase results in the complex setting by defining:a complex number to be constructible if and only if its real and imaginary parts are constructible real numbers. It can then be shown, in a manner analogous to the real case, that a complex number is constructible if and only if it lies in a field at the top of a finite tower of complex quadratic extensions, starting with the field . More precisely, is constructible if and only if there exists a tower of complex fields where is in , and for all , . Consequently, if a complex number is constructible, then is a power of two. This algebraic characterization of constructible numbers provides an important condition for constructibility: if is constructible, then it is algebraic, and its minimal irreducible polynomial has degree a power of 2, which is equivalent to the statement that the field extension has dimension a power of 2. However, the converse is false – this is not a condition for constructibility as there are non-constructible numbers with . Trigonometric numbers are irrational cosines or sines of angles that are rational multiples of . Such a number is constructible if and only if the denominator of the fully reduced multiple is a power of or the product of a power of with the product of one or more distinct Fermat primes. Thus, for example, Is constructible because is the product of two Fermat primes, and . See here a list of trigonometric numbers expressed in terms of square roots. The ancient Greeks thought that certain problems of straightedge and compass construction they could not solve were simply obstinate, not unsolvable. However, the non-constructibility of certain numbers proves them to be logically impossible to perform. (The problems themselves, however, are solvable using methods that go beyond the constraint of working only with straightedge and compass, and the Greeks knew how to solve them in this way.) In the following chart, each table represents a specific ancient construction problem. The left column gives the name of the problem. The second column gives an equivalent algebraic formulation of the problem. In other words, the solution to the problem is affirmative if and only if each number in the given set of numbers is constructible. Finally, the last column provides a simple counterexample. In other words, the number in the last column is an element of the set in the same row, but is not constructible. The birth of the concept of constructible numbers is inextricably linked with the history of the three impossible compass and straightedge constructions: duplicating the cube, trisecting an angle, and squaring the circle. The restriction of using only compass and straightedge in geometric constructions is often credited to Plato due to a passage in Plutarch. According to Plutarch, Plato gave the duplication of the cube (Delian) problem to Eudoxus and Archytas and Menaechmus, who solved the problem using mechanical means, earning a rebuke from Plato for not solving the problem using pure geometry (Plut., "Quaestiones convivales" VIII.ii, 718ef). However, this attribution is challenged, due, in part, to the existence of another version of the story (attributed to Eratosthenes by Eutocius of Ascalon) that says that all three found solutions but they were too abstract to be of practical value. Since Oenopides (circa 450 BCE) is credited with two ruler and compass constructions, by Proclus– citing Eudemus (circa 370 - 300 BCE)–when other methods were available to him, has led some authors to hypothesize that Oenopides originated the restriction. The restriction to compass and straightedge is essential in making these constructions impossible. Angle trisection, for instance, can be done in many ways, several known to the ancient Greeks. The Quadratrix of Hippias of Elis, the conics of Menaechmus, or the marked straightedge (neusis) construction of Archimedes have all been used and we can add a more modern approach via paper folding to the list. Although not one of the classic three construction problems, the problem of constructing regular polygons with straightedge and compass is usually treated alongside them. The Greeks knew how to construct regular -gons with (for any integer ) or the product of any two or three of these numbers, but other regular -gons eluded them. Then, in 1796, an eighteen-year-old student named Carl Friedrich Gauss announced in a newspaper that he had constructed a regular 17-gon with straightedge and compass. Gauss's treatment was algebraic rather than geometric; in fact, he did not actually construct the polygon, but rather showed that the cosine of a central angle was a constructible number. The argument was generalized in his 1801 book "Disquisitiones Arithmeticae" giving the "sufficient" condition for the construction of a regular -gon. Gauss claimed, but did not prove, that the condition was also necessary and several authors, notably Felix Klein, attributed this part of the proof to him as well. are impossible to solve if one uses only compass and straightedge. In the same paper he also solved the problem of determining which regular polygons are constructible: a regular polygon is constructible if and only if the number of its sides is the product of a power of two and any number of distinct Fermat primes (i.e., the sufficient conditions given by Gauss are also necessary) An attempted proof of the impossibility of squaring the circle was given by James Gregory in "Vera Circuli et Hyperbolae Quadratura" (The True Squaring of the Circle and of the Hyperbola) in 1667. Although his proof was faulty, it was the first paper to attempt to solve the problem using algebraic properties of . It was not until 1882 that Ferdinand von Lindemann rigorously proved its impossibility, by extending the work of Charles Hermite and proving that is a transcendental number. The study of constructible numbers, per se, was initiated by René Descartes in La Géométrie, an appendix to his book "Discourse on the Method" published in 1637. Descartes associated numbers to geometrical line segments in order to display the power of his philosophical method by solving an ancient straightedge and compass construction problem put forth by Pappus.
https://en.wikipedia.org/wiki?curid=7439
Carson City, Nevada Carson City, officially the Consolidated Municipality of Carson City, is an independent city and the capital of the U.S. state of Nevada, named after the mountain man Kit Carson. As of the 2010 census, the population was 55,274, making it the sixth largest city in Nevada. The majority of the town's population lives in Eagle Valley, on the eastern edge of the Carson Range, a branch of the Sierra Nevada, about south of Reno. The town began as a stopover for California-bound emigrants, but developed into a city with the Comstock Lode, a silver strike in the mountains to the northeast. The city has served as Nevada's capital since statehood in 1864; for much of its history it was a hub for the Virginia and Truckee Railroad, although the tracks were removed in 1950. Before 1969, Carson City was the county seat of Ormsby County. The county was abolished that year, and its territory merged with Carson City. With the consolidation, the city limits extend west across the Sierra Nevada to the California state line in the middle of Lake Tahoe. Like other independent cities in the United States, it is treated as a county-equivalent for census purposes. The Washoe people have inhabited the valley and surrounding areas for about 6,000 years. The first European Americans to arrive in what is now known as Eagle Valley were John C. Frémont and his exploration party in January 1843. Fremont named the river flowing through the valley Carson River in honor of Kit Carson, the mountain man and scout he had hired for his expedition. Later, settlers named the area Washoe in reference to the indigenous people. By 1851 the Eagle Station ranch along the Carson River was a trading post and stopover for travelers on the California Trail's Carson Branch which ran through Eagle Valley. The valley and trading post received their name from a bald eagle that was hunted and killed by one of the early settlers and was featured on a wall inside the post. As the area was part of the Utah Territory, it was governed from Salt Lake City, where the territorial government was headquartered. Early settlers bristled at the control by Mormon-influenced officials and desired the creation of the Nevada territory. A vigilante group of influential settlers, headed by Abraham Curry, sought a site for a capital city for the envisioned territory. In 1858, Abraham Curry bought Eagle Station and the settlement was thereafter renamed Carson City. Curry and several other partners had Eagle Valley surveyed for development. Curry decided Carson City would someday serve as the capital city and left a plot in the center of town for a capitol building. After gold and silver were discovered in 1859 on nearby Comstock Lode, Carson City's population began to grow. Curry built the Warm Springs Hotel a mile to the east of the city center. When territorial governor James W. Nye traveled to Nevada, he chose Carson City as the territorial capital, influenced by Carson City lawyer William Stewart, who escorted him from San Francisco to Nevada. As such, Carson City bested Virginia City and American Flat. Curry loaned the Warm Springs Hotel to the territorial Legislature as a meeting hall. The Legislature named Carson City to be the seat of Ormsby County and selected the hotel as the territorial prison with Curry serving as its first warden. Today the property is still part of the state prison. When Nevada became a state in 1864 during the American Civil War, Carson City was confirmed as Nevada's permanent capital. Carson City's development was no longer dependent on the mining industry and instead became a thriving commercial center. The Virginia and Truckee Railroad was built between Virginia City and Carson City. A log flume was also built from the Sierra Nevadas into Carson City. The current capitol building was constructed from 1870 to 1871. The United States Mint operated the Carson City Mint between the years 1870 and 1893, which struck gold and silver coins. People came from China during that time, many to work on the railroad. Some of them owned businesses and taught school. By 1880, almost a thousand Chinese people, "one for every five Caucasians", lived in Carson City. Carson City's population and transportation traffic decreased when the Central Pacific Railroad built a line through Donner Pass, too far to the north to benefit Carson City. The city was slightly revitalized with the mining booms in Tonopah and Goldfield. The US federal building (now renamed the Paul Laxalt Building) was completed in 1890 as was the Stewart Indian School. Even these developments could not prevent the city's population from dropping to just over 1,500 people by 1930. Carson City resigned itself to small city status, advertising itself as "America's smallest capital". The city slowly grew after World War II; by 1960 it had reached its 1880 boom-time population. As early as the late 1940s, discussions began about merging Ormsby County and Carson City. By this time, the county was little more than Carson City and a few hamlets to the west. However, the effort did not pay off until 1966, when a statewide referendum approved the merger. The required constitutional amendment was passed in 1968. On April 1, 1969, Ormsby County and Carson City officially merged as the Consolidated Municipality of Carson City. With this consolidation, Carson City absorbed former town sites such as Empire City, which had grown up in the 1860s as a milling center along the Carson River and current U.S. Route 50. Carson City could now advertise itself as one of America's largest state capitals with its of city limits. In 1991, the city adopted a downtown master plan, specifying no building within of the capitol would surpass it in height. This plan effectively prohibited future high-rise development in the center of downtown. The Ormsby House is the tallest building in downtown Carson City, at a height of . The structure was completed in 1972. Carson City features a semi-arid climate with cold winters and hot summers. The city is in a high desert river valley approximately above sea level. There are four fairly distinct seasons. Winters see typically light to moderate snowfall, with a median of . Most precipitation occurs in winter and spring, with summer and fall being fairly dry, drier than neighboring California. There are 37 days of + highs annually, with + temperatures occurring in some years. The average temperature in Carson City increased by between 1984 and 2014, a greater change than in any other city in the United States. The Carson River flows from Douglas County through the southwestern edge of Carson City. Carson City is the smallest of the United States' 366 metropolitan statistical areas. As of the 2010 census, there were 55,274 people, 20,171 households, and 13,252 families residing in the city. The population density was 366 people per square mile (141/km2). There were 21,283 housing units at an average density of 148/sq mi (57/km2). The racial makeup of the city was 81.1% White, 1.9% Black or African American, 2.4% Native American, 2.1% Asian, 0.2% Pacific Islander, 9.4% from other races, and 2.9% from two or more races. 21% of the population were Hispanic or Latino of any race. As of the 2000 census, there were 20,171 households, out of which 29.8% had children under the age of 18 living with them, 50.0% were married couples living together, 11.0% had a female householder with no husband present, and 34.3% were non-families. 27.8% of all households were made up of individuals and 11.00% had someone living alone who was 65 years of age or older. The average household size was 2.44 and the average family size was 2.97. The city's age distribution was: 23.4% under the age of 18, 7.9% from 18 to 24, 28.9% from 25 to 44, 24.9% from 45 to 64, and 14.9% who were 65 years of age or older. The median age was 39 years. For every 100 females, there were 106.9 males. For every 100 females age 18 and over, there were 108.2 males. Data from the 2000 census indicates the median income for a household in the city was $41,809, and the median income for a family was $49,570. Males had a median income of $35,296 versus $27,418 for females. The per capita income for the city was $20,943. 10.0% of the population and 6.9% of families were below the poverty line. Out of the total population, 13.7% of those under the age of 18 and 5.8% of those 65 and older were living below the poverty line. As of 2010, 82.3% (42,697) of Carson City residents age 5 and older spoke English at home as a first language, while 14.1% (7,325) spoke Spanish, 0.6% (318) French, and numerous Indo-Aryan languages were spoken as a main language by 0.5% (261) of the population over the age of five. In total, 17.7% (9,174) of Carson City's population age 5 and older spoke a first language other than English. Ormsby County consolidated with Carson City in 1969, and the county simultaneously dissolved. The city is now governed by a five-member board of supervisors, consisting of a mayor and four supervisors. All members are elected at-large, but each of the four supervisors must reside in respective wards, numbered 1 through 4. The mayor and supervisors serve four year terms. Elections are staggered so the mayor and the supervisors from Wards 2 and Ward 4 are elected in presidential election years, and the supervisors from Ward 1 and 3 are elected in the even-numbered years in between (i.e., the same year as gubernatorial elections). Nevada's capital is generally considered a Republican stronghold, often voting for Republicans by wide margins. In 2004, George Bush defeated John Kerry 57–40%. In 2008, however, Barack Obama became the first Democrat since 1964 to win Ormsby County/Carson City, defeating John McCain 49% to 48%, by 204 votes, a margin of under 1%. Carson City, being the state capital, is home to many political protests and demonstrations at any given time. In an attempt to either make proposed spent nuclear fuel storage facility at Yucca Mountain prohibitively expensive (by raising property tax rates to the maximum allowed) or to allow the state to collect the potential federal payments of property taxes on the facility, the state government in 1987 carved Yucca Mountain out of Nye County and created a new county with no residents out of the area surrounding Yucca called Bullfrog County. Carson City became the county seat of Bullfrog County, even though it is not in Bullfrog County and is more than from Yucca Mountain. A state judge found the process unconstitutional in 1989, and Bullfrog County's territory was retroceded to Nye County. Carson City has never hosted any professional team sports. However, a variety of sports are offered at parks and recreation. Many neighborhood parks offer a wide variety of features including picnic tables, beaches, restrooms, fishing, softball, basketball hoops, ponds, tennis, and volleyball. The largest park is Mills Park, which has a total land area of and includes the narrow-gauge Carson & Mills Park Railroad. While there are no ski slopes within Carson City, the city is near the Heavenly Mountain Resort, Diamond Peak and Mount Rose Ski Tahoe skiing areas. Carson City has served as one of the state's centers for politics and business. Every state governor since Denver S. Dickerson has resided in the Governor's Mansion in Carson City. "See also: List of Governors of Nevada." The following personalities took up a residence in Carson City at some point in their lives. The following is a list of the top employers in Carson City from the fourth quarter of 2012: 1,500–1,999 employees 1,000–1,499 Employees 500–999 employees 200–499 employees 100–199 employees There are four highways in the city: Nevada State Route 28, U.S. Route 395, U.S. Route 50, and Interstate 580, its only freeway. Phase 1 of the Carson City Freeway Project from US 395, just north of the city, to US 50 was completed in February 2006 and Phase 2A, extending from Rt. 50 to Fairview Drive, was officially opened on September 24, 2009. Phase 2B, Fairview Drive to Rt. 50, was completed in August 2017. Prior to 2012, Carson City was one of only five state capitals not directly served by an Interstate highway; the city lost this distinction when I-580 was extended into the city limits. Carson City's first modern bus system, Jump Around Carson, or JAC, opened to the public in October 2005. JAC uses a smaller urban bus ideal for Carson City. However, there is virtually no ground public transportation to other destinations. Passenger trains haven't served Carson City since 1950, when the Virginia and Truckee Railroad was shut down. Greyhound Lines stopped their bus services to the town in 2006 and Amtrak discontinued their connecting thruway bus to Sacramento, California in 2008. There is now only a limited Monday – Friday RTC bus service to Reno which is still served by both Greyhound and Amtrak, as well as Eastern Sierra Transit Authority service from Lone Pine to Reno. Carson City is also served by the Carson Airport, which is a regional airport in the northern part of the city. Reno–Tahoe International Airport, which is away, handles domestic commercial flights. The Carson City School District operates ten schools in Carson City. The six elementary schools are Bordewich-Bray Elementary School, Empire Elementary School, Fremont Elementary School, Fritsch Elementary School, Mark Twain Elementary School, and Al Seeliger Elementary School. The two middle schools are Carson Middle School and Eagle Valley Middle School. Carson High School and the alternative Pioneer High School serve high school students. Carson High is on Saliman Road. Western Nevada College (WNC) is a regionally accredited, two-year and four-year institution which is part of the Nevada System of Higher Education. The college offers many programs including education, arts and science. Carson City has a public library, the Carson City Library.
https://en.wikipedia.org/wiki?curid=7441
Classification of finite simple groups In mathematics, the classification of the finite simple groups is a theorem stating that every finite simple group is either cyclic, or alternating, or it belongs to a broad infinite class called the groups of Lie type, or else it is one of twenty-six or twenty-seven exceptions, called sporadic. Group theory is central to many areas of pure and applied mathematics and the classification theorem has been called one of the great intellectual achievements of humanity. The proof consists of tens of thousands of pages in several hundred journal articles written by about 100 authors, published mostly between 1955 and 2004. Simple groups can be seen as the basic building blocks of all finite groups, reminiscent of the way the prime numbers are the basic building blocks of the natural numbers. The Jordan–Hölder theorem is a more precise way of stating this fact about finite groups. However, a significant difference from integer factorization is that such "building blocks" do not necessarily determine a unique group, since there might be many non-isomorphic groups with the same composition series or, put in another way, the extension problem does not have a unique solution. Gorenstein (d.1992), Lyons, and Solomon are gradually publishing a simplified and revised version of the proof. The classification theorem has applications in many branches of mathematics, as questions about the structure of finite groups (and their action on other mathematical objects) can sometimes be reduced to questions about finite simple groups. Thanks to the classification theorem, such questions can sometimes be answered by checking each family of simple groups and each sporadic group. Daniel Gorenstein announced in 1983 that the finite simple groups had all been classified, but this was premature as he had been misinformed about the proof of the classification of quasithin groups. The completed proof of the classification was announced by after Aschbacher and Smith published a 1221-page proof for the missing quasithin case. wrote two volumes outlining the low rank and odd characteristic part of the proof, and wrote a 3rd volume covering the remaining characteristic 2 case. The proof can be broken up into several major pieces as follows: The simple groups of low 2-rank are mostly groups of Lie type of small rank over fields of odd characteristic, together with five alternating and seven characteristic 2 type and nine sporadic groups. The simple groups of small 2-rank include: The classification of groups of small 2-rank, especially ranks at most 2, makes heavy use of ordinary and modular character theory, which is almost never directly used elsewhere in the classification. All groups not of small 2 rank can be split into two major classes: groups of component type and groups of characteristic 2 type. This is because if a group has sectional 2-rank at least 5 then MacWilliams showed that its Sylow 2-subgroups are connected, and the balance theorem implies that any simple group with connected Sylow 2-subgroups is either of component type or characteristic 2 type. (For groups of low 2-rank the proof of this breaks down, because theorems such as the signalizer functor theorem only work for groups with elementary abelian subgroups of rank at least 3.) A group is said to be of component type if for some centralizer "C" of an involution, "C"/"O"("C") has a component (where "O"("C") is the core of "C", the maximal normal subgroup of odd order). These are more or less the groups of Lie type of odd characteristic of large rank, and alternating groups, together with some sporadic groups. A major step in this case is to eliminate the obstruction of the core of an involution. This is accomplished by the B-theorem, which states that every component of "C"/"O"("C") is the image of a component of "C". The idea is that these groups have a centralizer of an involution with a component that is a smaller quasisimple group, which can be assumed to be already known by induction. So to classify these groups one takes every central extension of every known finite simple group, and finds all simple groups with a centralizer of involution with this as a component. This gives a rather large number of different cases to check: there are not only 26 sporadic groups and 16 families of groups of Lie type and the alternating groups, but also many of the groups of small rank or over small fields behave differently from the general case and have to be treated separately, and the groups of Lie type of even and odd characteristic are also quite different. A group is of characteristic 2 type if the generalized Fitting subgroup "F"*("Y") of every 2-local subgroup "Y" is a 2-group. As the name suggests these are roughly the groups of Lie type over fields of characteristic 2, plus a handful of others that are alternating or sporadic or of odd characteristic. Their classification is divided into the small and large rank cases, where the rank is the largest rank of an odd abelian subgroup normalizing a nontrivial 2-subgroup, which is often (but not always) the same as the rank of a Cartan subalgebra when the group is a group of Lie type in characteristic 2. The rank 1 groups are the thin groups, classified by Aschbacher, and the rank 2 ones are the notorious quasithin groups, classified by Aschbacher and Smith. These correspond roughly to groups of Lie type of ranks 1 or 2 over fields of characteristic 2. Groups of rank at least 3 are further subdivided into 3 classes by the trichotomy theorem, proved by Aschbacher for rank 3 and by Gorenstein and Lyons for rank at least 4. The three classes are groups of GF(2) type (classified mainly by Timmesfeld), groups of "standard type" for some odd prime (classified by the Gilman–Griess theorem and work by several others), and groups of uniqueness type, where a result of Aschbacher implies that there are no simple groups. The general higher rank case consists mostly of the groups of Lie type over fields of characteristic 2 of rank at least 3 or 4. The main part of the classification produces a characterization of each simple group. It is then necessary to check that there exists a simple group for each characterization and that it is unique. This gives a large number of separate problems; for example, the original proofs of existence and uniqueness of the monster group totaled about 200 pages, and the identification of the Ree groups by Thompson and Bombieri was one of the hardest parts of the classification. Many of the existence proofs and some of the uniqueness proofs for the sporadic groups originally used computer calculations, most of which have since been replaced by shorter hand proofs. In 1972 announced a program for completing the classification of finite simple groups, consisting of the following 16 steps: Many of the items in the list below are taken from . The date given is usually the publication date of the complete proof of a result, which is sometimes several years later than the proof or first announcement of the result, so some of the items appear in the "wrong" order. The proof of the theorem, as it stood around 1985 or so, can be called "first generation". Because of the extreme length of the first generation proof, much effort has been devoted to finding a simpler proof, called a second-generation classification proof. This effort, called "revisionism", was originally led by Daniel Gorenstein. , eight volumes of the second generation proof have been published (Gorenstein, Lyons & Solomon 1994, 1996, 1998, 1999, 2002, 2005, 2018a, 2018b). In 2012 Solomon estimated that the project would need another 5 volumes, but said that progress on them was slow. It is estimated that the new proof will eventually fill approximately 5,000 pages. (This length stems in part from the second generation proof being written in a more relaxed style.) Aschbacher and Smith wrote their two volumes devoted to the quasithin case in such a way that those volumes can be part of the second generation proof. Gorenstein and his collaborators have given several reasons why a simpler proof is possible. Gorenstein has discussed some of the reasons why there might not be a short proof of the classification similar to the classification of compact Lie groups. This section lists some results that have been proved using the classification of finite simple groups.
https://en.wikipedia.org/wiki?curid=7445
Chalcolithic The Chalcolithic (), a name derived from the "khalkós", "copper" and from "líthos", "stone" or Copper Age, also known as the Eneolithic or Aeneolithic (from Latin "aeneus" "of copper") is an archaeological period which researchers usually regard as part of the broader Neolithic (although scholars originally defined it as a transition between the Neolithic and the Bronze Age). In the context of Eastern Europe, archaeologists often prefer the term "Eneolithic" to "Chalcolithic" or other alternatives. In the Chalcolithic period, copper predominated in metalworking technology. Hence it was the period before it was discovered that by adding tin to copper one could create bronze, a metal alloy harder and stronger than either component. The archaeological site of Belovode, on Rudnik mountain in Serbia, has the worldwide oldest securely-dated evidence of copper smelting at high temperature, from 5000 BC (7000 BP ). The transition from Copper Age to Bronze Age in Europe occurs between the late 5th and the late 3rd millennia BC. In the Ancient Near East the Copper Age covered about the same period, beginning in the late 5th millennium BC and lasting for about a millennium before it gave rise to the Early Bronze Age. The multiple names result from multiple recognitions of the period. Originally, the term Bronze Age meant that either copper or bronze was being used as the chief hard substance for the manufacture of tools and weapons. Ancient writers, who provided the essential cultural references for educated people during the 19th century, used the same names for both. In 1881, John Evans recognized that use of copper often preceded the use of bronze, and distinguished between a "transitional Copper Age" and the "Bronze Age proper". He did not include the transitional period in the three-age system of Early, Middle and Late Bronze Age, but placed it outside the tripartite system, at its beginning. He did not, however, present it as a fourth age but chose to retain the traditional tripartite system. In 1884, Gaetano Chierici, perhaps following the lead of Evans, renamed it in Italian as the "eneo-litica", or "bronze–stone" transition. The phrase was never intended to mean that the period was the only one in which both bronze and stone were used. The Copper Age features the use of copper, excluding bronze; moreover, stone continued to be used throughout both the Bronze Age and the Iron Age. The part "-litica" simply names the Stone Age as the point from which the transition began and is not another "-lithic" age. Subsequently, British scholars used either Evans's "Copper Age" or the term "Eneolithic" (or Æneolithic), a translation of Chierici's "eneo-litica". After several years, a number of complaints appeared in the literature that "Eneolithic" seemed to the untrained eye to be produced from "e-neolithic", "outside the Neolithic", clearly not a definitive characterization of the Copper Age. Around 1900, many writers began to substitute "Chalcolithic" for Eneolithic, to avoid the false segmentation. It was then that the misunderstanding began among those who did not know Italian. The Chalcolithic was seen as a new "-lithic" age, a part of the Stone Age in which copper was used, which may appear paradoxical. Today, "Copper Age", "Eneolithic" and "Chalcolithic" are used synonymously to mean Evans's original definition of Copper Age. The literature of European archaeology in general avoids the use of "Chalcolithic" (the term "Copper Age" is preferred), whereas Middle Eastern archaeologists regularly use it. "Chalcolithic" is not generally used by British prehistorians, who disagree as to whether it applies in the British context. The emergence of metallurgy may have occurred first in the Fertile Crescent. The earliest use of lead is documented here from the late Neolithic settlement of Yarim Tepe in Iraq, "The earliest lead (Pb) finds in the ancient Near East are a 6th millennium BC bangle from Yarim Tepe in northern Iraq and a slightly later conical lead piece from Halaf period Arpachiyah, near Mosul. As native lead is extremely rare, such artifacts raise the possibility that lead smelting may have begun even before copper smelting." Copper smelting is also documented at this site at about the same time period (soon after 6000 BC), although the use of lead seems to precede copper smelting. Early metallurgy is also documented at the nearby site of Tell Maghzaliyah, which seems to be dated even earlier, and completely lacks pottery. The Timna Valley contains evidence of copper mining in 7000–5000 BC. The process of transition from Neolithic to Chalcolithic in the Middle East is characterized in archaeological stone tool assemblages by a decline in high quality raw material procurement and use. This dramatic shift is seen throughout the region, including the Tehran Plain, Iran. Here, analysis of six archaeological sites determined a marked downward trend in not only material quality, but also in aesthetic variation in the lithic artefacts. Fazeli et al. use these results as evidence of the loss of craft specialisation caused by increased use of copper tools. The Tehran Plain findings illustrate the effects of the introduction of copper working technologies on the in-place systems of lithic craft specialists and raw materials. Networks of exchange and specialized processing and production that had evolved during the Neolithic seem to have collapsed by the Middle Chalcolithic ( 4500–3500 BC) and been replaced by the use of local materials by a primarily household-based production of stone tools. An archaeological site in Serbia contains the oldest securely dated evidence of coppermaking from 7,500 years ago. The find in June 2010 extends the known record of copper smelting by about 800 years, and suggests that copper smelting may have been invented in separate parts of Asia and Europe at that time rather than spreading from a single source. In Serbia, a copper axe was found at Prokuplje, which indicates use of metal in Europe by 7,500 years ago (5500 BC), many years earlier than previously believed. Knowledge of the use of copper was far more widespread than the metal itself. The European Battle Axe culture used stone axes modeled on copper axes, even with moulding carved in the stone. Ötzi the Iceman, who was found in the Ötztal Alps in 1991 and whose remains were dated to about 3300 BC, was found with a Mondsee copper axe. Examples of Chalcolithic cultures in Europe include Vila Nova de São Pedro and Los Millares on the Iberian Peninsula. Pottery of the Beaker people has been found at both sites, dating to several centuries after copper-working began there. The Beaker culture appears to have spread copper and bronze technologies in Europe, along with Indo-European languages. In Britain, copper was used between the 25th and 22nd centuries BC, but some archaeologists do not recognise a British Chalcolithic because production and use was on a small scale. According to Parpola (2005), ceramic similarities between the Indus Civilization, southern Turkmenistan, and northern Iran during 4300–3300 BC of the Chalcolithic period suggest considerable mobility and trade. The term "Chalcolithic" has also been used in the context of the South Asian Stone Age. In Bhirrana, the earliest Indus civilization site, copper bangles and arrowheads were found. The inhabitants of Mehrgarh in present-day Pakistan fashioned tools with local copper ore between 7000–3300 BC. At the Nausharo site dated to 4500 years ago, a pottery workshop in province of Balochistan, Pakistan, were unearthed 12 blades or blade fragments. These blades are long and and relatively thin. Archaeological experiments show that these blades were made with a copper indenter and functioned as a potter's tool to trim and shape unfired pottery. Petrographic analysis indicates local pottery manufacturing, but also reveals the existence of a few exotic black-slipped pottery items from the Indus Valley. There was an independent invention of copper and bronze smelting first by Andean civilizations in South America extended later by sea commerce to the Mesoamerican civilization in West Mexico (see Metallurgy in pre-Columbian America and Metallurgy in pre-Columbian Mesoamerica). The term "Chalcolithic" is also applied to American civilizations that already used copper and copper alloys thousands of years before the European migration. Besides cultures in the Andes and Mesoamerica, the Old Copper Complex, centered in the Upper Great Lakes region—present-day Michigan and Wisconsin in the United States—mined and fabricated copper as tools, weapons, and personal ornaments. The evidence of smelting or alloying that has been found is subject to some dispute and a common assumption by archaeologists is that objects were cold-worked into shape. Artifacts from some of these sites have been dated to 4000–1000 BC, making them some of the oldest Chalcolithic sites in the world. Furthermore, some archaeologists find artifactual and structural evidence of casting by Hopewellian and Mississippian peoples to be demonstrated in the archaeological record. In the 5th millennium BC copper artifacts start to appear in East Asia, such as in the Jiangzhai and Hongshan cultures, but those metal artifacts were not widely used. While copper objects make occasionally appear early on, these finds do not represent a regular practice of copper metallurgy which, in East Asia, begins with the entry of Afanasievo groups into western Mongolia towards the end of 4th millennium and beginning of the third millennium BC. In the region of the Aïr Mountains in Niger, we have the development of independent copper smelting between 3000 and 2500 BC. The process was not in a developed state, indicating smelting was not foreign. It became mature about 1500 BC.
https://en.wikipedia.org/wiki?curid=7446
Circumcision and law Laws restricting, regulating, or banning circumcision, some dating back to ancient times, have been enacted in many countries and communities. In modern states, circumcision is generally presumed to be legal, but laws pertaining to assault or child custody have been applied in cases involving circumcision. There are currently no states that unequivocally ban infant male circumcision for non-therapeutic reasons. In the case of non-therapeutic circumcision of children, proponents of laws in favor of the procedure often point to the rights of the parents or practitioners, namely the right of freedom of religion. Those against the procedure point to the boy's right of freedom "from" religion. In several court cases, judges have pointed to the irreversible nature of the act, the grievous harm to the boy's body, and the right to self-determination, and bodily integrity. There are ancient religious requirements for circumcision. The Hebrew Bible commands Jews to circumcise their male children on the eighth day of life, and to circumcise their male slaves (). Laws banning circumcision are also ancient. The ancient Greeks prized the foreskin and disapproved of the Jewish custom of circumcision. 1 Maccabees, 1:60–61 states that King Antiochus IV of Syria, the occupying power of Judea in 170 BCE, outlawed circumcision on penalty of death. one of the grievances leading to the Maccabean Revolt. According to the "Historia Augusta", the Roman emperor Hadrian issued a decree banning circumcision in the empire, and some modern scholars argue that this was a main cause of the Jewish Bar Kokhba revolt of 132 CE. The Roman historian Cassius Dio, however, made no mention of such a law, and blamed the Jewish uprising instead on Hadrian's decision to rebuild Jerusalem as Aelia Capitolina, a city dedicated to Jupiter. Antoninus Pius permitted Jews to circumcise their own sons. However, he forbade the circumcision of non-Jews that were either foreign-slaves or non-Jewish members of the household, contrary to He also made it illegal for a man to convert to Judaism. Antoninus Pius exempted the Egyptian priesthood from the otherwise universal ban on circumcision. Before glasnost, according to an article in The Jewish Press, Jewish ritual circumcision was forbidden in the Soviet Union. However, David E. Fishman, professor of Jewish History at the Jewish Theological Seminary of America, states that, whereas the "heder" and "yeshiva", the organs of Jewish education, "were banned by virtue of the law separating church and school, and subjected to tough police and administrative actions," circumcision was not proscribed by law or suppressed by executive measures. Jehoshua A. Gilboa writes that while circumcision was not officially or explicitly banned, pressure was exerted to make it difficult. "Mohels" in particular were concerned that they could be punished for any health issue that might develop, even if it arose some time after the circumcision. In 1967 all religion in Communist Albania was banned, along with the practice of circumcision. The practice was driven underground and many boys were secretly circumcised. Whereas child custody regulations have been applied to cases involving circumcision, there seems to be no state which currently unequivocally bans infant male circumcision for non-therapeutic reasons, albeit the legality of such circumcision is disputed in some legislations. The present table provides a non-exhaustive overview comparing legal restrictions and requirements on non-therapeutic infant circumcision in several countries. Some countries require one or both parents to consent to the operation; some of these (Finland, United Kingdom) have experienced legal battles between parents when one of them had their son's circumcision carried out or planned without the other's consent. Some countries require the procedure to be performed by or supervised by a qualified physician (or a qualified nurse in Sweden), and with (local) anaesthesia applied to the boy or man. The Royal Australasian College of Physicians (RACP) finds that routine infant circumcision is not warranted in Australia and New Zealand and that, since circumcision involves physical injury, physicians ought to raise and consider with parents and considered the option of leaving circumcision until later, when the boy is old enough to make a decision for himself: In 1993, a non-binding research paper of the Queensland Law Reform Commission ("Circumcision of Male Infants") concluded that "On a strict interpretation of the assault provisions of the Queensland Criminal Code, routine circumcision of a male infant could be regarded as a criminal act", and that doctors who perform circumcision on male infants may be liable to civil claims by that child at a later date. No prosecutions have occurred in Queensland, and circumcisions continue to be performed. In 1999, a Perth man won A$360,000 in damages after a doctor admitted he botched a circumcision operation at birth which left the man with a badly deformed penis. In 2002, Queensland police charged a father with grievous bodily harm for having his two sons, then aged nine and five, circumcised without the knowledge and against the wishes of the mother. The mother and father were in a family court dispute. The charges were dropped when the police prosecutor revealed that he did not have all family court paperwork in court and the magistrate refused to grant an adjournment. Cosmetic circumcision for newborn males is currently banned in all Australian public hospitals, South Australia being the last state to adopt the ban in 2007; the procedure was not forbidden from being performed in private hospitals. In the same year, the Tasmanian President of the Australian Medical Association, Haydn Walters, stated that they would support a call to ban circumcision for non-medical, non-religious reasons. In 2009, the Tasmanian Law Reform Institute released its Issues Paper investigating the law relating to male circumcision in Tasmania, it "highlights the uncertainty in relation to whether doctors can legally perform circumcision on infant males". The Tasmania Law Reform Institute released its recommendations for reform of Tasmanian law relative to male circumcision on 21 August 2012. The report makes fourteen recommendations for reform of Tasmanian law relative to male circumcision. The Belgian Advisory Committee on Bioethics finds that circumcision is a radical operation, and that physical integrity of the child takes precedence over parents' belief systems: In 2012, "Le Soir" reported a 21% increase in the amount of circumcisions in Belgium from 2006 and 2011. In the previous 25 years, one in three Belgian-born boys had allegedly been circumcised. A questionnaire to hospitals in Wallonia and Brussels showed that about 80 to 90% of the procedures had religious or cultural motives. The Ministry of Health stressed the importance of safe circumstances, physicians warned that 'no surgical procedure is without risk' and that circumcision was 'not a necessary procedure'. In 2017, it was estimated that about 15% of Belgian men were circumcised. The incidence has been gradually rising: in 2002, about 17,800 boys or men underwent circumcision, which increased to almost 26,200 in 2016. The expenses of undergoing circumcision are covered by the National Institute for Disease and Disability Insurance (RIZIV/INAMI), costing about 2.7 million euros in 2016. After inquiries were submitted to the Belgian Bioethics Advisory Committee in early 2014, an ethics commission was set up to review the morality of covering the costs of medically unnecessary surgery through taxpayer money, especially considering that many taxpayers regard the practice as immoral. By July 2017, the commission reportedly reached consensus on discontinuing the financial coverage of non-medical circumcision, but was still debating whether to advise the government to institute a total ban of the practice. The commission's final (non-binding) recommendation, presented on 19 September 2017, was to cease public funding for non-medical circumcision, and to not circumcise anyone underage until they can consent or reject the procedure after being properly informed. This was in line with the 1990 Convention on the Rights of the Child, and mirrors the 2013 non-binding Parliamentary Assembly of the Council of Europe's resolution against underage non-therapeutic circumcision. However, Health Minister Maggie De Block rejected the commission's advice, arguing the RIZIV 'cannot know whether there is a medical motive or not' when parents request a circumcision, and when they are denied a professional procedure, chances are parents will have a non-expert perform it, leading to worse results for the children. The Health Minister's response was received with mixed reactions. The Canadian Paediatric Society doesn't recommend routine circumcision, finding that medical necessity has not been clearly established, and as such, that it should be deferred until the individual concerned is able to make his own choices: According to the College of Physicians and Surgeons of British Columbia: The Danish Medical Association finds that circumcision ought to be an informed, personal choice, in line with individual's right to self-determination, and thus that parents have no right to make this decision for their children, leaving up to the individual when he's old enough to make it: A January 2018 survey by polling company Megafon for the television network TV2 showed that 83% of Danish citizens favoured a ban on circumcising boys. A citizens' initiative, put forward by Intact Denmark, demanding a minimum age of 18 for circumcision to protect “children’s fundamental rights” reached 50,000 signatures on 1 June 2018, forcing the Danish Parliament to consider it. It is unsure whether the proposal will gain a majority. In September 2018, the Parliament's justice administration (Lovsekretariat) completed its review of the petition, finding it not to be in breach of the Danish Constitution, and therefore valid for parliamentary consideration. At the first parliamentary reading in November 23, 2018, spokesmen for DF, V, S and RV reported that their parties would vote against the proposal. EL and SF were supportive of a ban, but took issue with details of the proposal, particularly the legalization of circumcision of women of age (i.e., 18 years old). LA, KF and ALT did not make a collective decision, leaving their members to decide for themselves. National security was cited as an argument against. For example, Minister of Health Ellen Trane Nørby, listed "foreign-political implications as well as consequences of security, among other things in relation to the stationed soldiers of Denmark" as some of the cabinet's reasons for opposing the proposal. When asked whether he felt that national concerns outweighed child rights, Socialdemokraterne spokesman Flemming Møller Mortensen said Venstre spokesman Jane Heitmann similarly reported concerns in her party that such a ban "could be abused to misrepresent Denmark and have consequences for the international reputation of Denmark and for Danes living abroad. We do not wish to take that risk". In November 2018, the Venstre party and the Danish People's Party criticised the current wording of the proposal, raising concerns that it could legalise female genital mutilation, while the Ministry of Health denied Intact Denmark's claim that current legislation does not comply with the UN Convention on the Rights of the Child. A study commissioned by the European Parliament Committee on Civil Liberties, Justice and Home Affairs published in February 2013 stated that "Male circumcision for non-therapeutic reasons appears to be practiced with relative regularity and frequency throughout Europe," and said it was "the only scenario, among the topics discussed in the present chapter, in which the outcome of the balancing between the right to physical integrity and religious freedom is in favour of the latter." The study recommended that "the best interests of children should be paramount, while acknowledging the relevance of this practice for Muslims and Jews. Member States should ensure that circumcision of underage children is performed according to the medical profession’s art and under conditions that do not put the health of minors at risk. The introduction of regulations by the Member States in order to set the conditions and the appropriate medical training for those called to perform it is warranted." On 1 October 2013, the Parliamentary Assembly of the Council of Europe adopted a non-binding resolution in which they state they are "particularly worried about a category of violation of the physical integrity of children," and included in this category "circumcision of young boys for religious reasons." On 7 October, Israel's president Shimon Peres wrote a personal missive to the Secretary General of the Council of Europe, Thorbjørn Jagland, to stop the ban, arguing: "The Jewish communities across Europe would be greatly afflicted to see their cultural and religious freedom impeded upon by the Council of Europe, an institution devoted to the protection of these very rights." Two days later, Jagland clarified that the resolution was non-binding and that “Nothing in the body of our legally binding standards would lead us to put on equal footing the issue of female genital mutilation and the circumcision of young boys for religious reasons.” As of February 2018, no European country has a ban on male circumcision, but Iceland was planning to become the first to outlaw the practice for non-medical reasons. The Finnish Ombudsman for Equality finds that circumcising young boys without a medical reason is legally highly questionable, The Finnish Supreme Court found that non-therapeutic circumcision of boys is assault, and the Finnish Ombudsman for Children proposed that Finland should ban non-therapeutic circumcision of young boys: In August 2006, a Finnish court ruled that the circumcision of a four-year-old boy arranged by his mother, who is Muslim, to be an illegal assault. The boy's father, who had not been consulted, reported the incident to the police. A local prosecutor stated that the prohibition of circumcision is not gender-specific in Finnish law. A lawyer for the Ministry of Social Affairs and Health stated that there is neither legislation nor prohibition on male circumcision, and that "the operations have been performed on the basis of common law." The case was appealed and in October 2008 the Finnish Supreme Court ruled that the circumcision, "carried out for religious and social reasons and in a medical manner, did not have the earmarks of a criminal offence. It pointed out in its ruling that the circumcision of Muslim boys is an established tradition and an integral part of the identity of Muslim men". In 2008, the Finnish government was reported to be considering a new law to legalise circumcision if the practitioner is a doctor and if the child consents. In December 2011, Helsinki District Court said that the Supreme Court's decision does not mean that circumcision is legal for any non-medical reasons. The court referred to the Convention on Human rights and Biomedicine of the Council of Europe, which was ratified in Finland in 2010. In February 2010, a Jewish couple were fined for causing bodily harm to their then infant son who was circumcised in 2008 by a mohel brought in from the UK. Normal procedure for persons of Jewish faith in Finland is to have a locally certified mohel who works in Finnish healthcare perform the operation. In the 2008 case, the infant was not anesthetized and developed complications that required immediate hospital care. The parents were ordered to pay 1500 euros in damages to their child. The German Association of Pediatricians (BVKJ) finds no medical reason for non-therapeutic circumcision and that the AAP (2012) recommendation scientifically unsustainable, and that boys should have the same constitutional legal right to physical integrity as girls: In October 2006, a Turkish national who performed ritual circumcisions on seven boys was convicted of causing dangerous bodily harm by the state court in Düsseldorf. In September 2007, a Frankfurt am Main appeals court found that the circumcision of an 11-year-old boy without his approval was an unlawful personal injury. The boy, whose parents were divorced, was visiting his Muslim father during a vacation when his father forced him to be ritually circumcised. The boy had planned to sue his father for . In May 2012, the Cologne regional appellate court ruled that religious circumcision of male children amounts to bodily injury, and is a criminal offense in the area under its jurisdiction. The decision based on the article "Criminal Relevance of Circumcising Boys. A Contribution to the Limitation of Consent in Cases of Care for the Person of the Child" published by Holm Putzke, a German law professor at the University of Passau. The court arrived at its judgment by application of the human rights provisions of the Basic Law, a section of the Civil Code, and some sections of the Criminal Code to non-therapeutic circumcision of male children. Some observers said it could set a legal precedent that criminalizes the practice. Jewish and Muslim groups were outraged by the ruling, viewing it as trampling on freedom of religion. The German ambassador to Israel, Andreas Michaelis, told Israeli lawmakers that Germany was working to resolve the issue and that it doesn't apply at a national level, but instead only to the local jurisdiction of the court in Cologne. The Council of the Coordination of Muslims in Germany condemned the ruling, stating that it is "a serious attack on religious freedom." Ali Kizilkaya, a spokesman of the council, stated that, "The ruling does not take everything into account, religious practice concerning circumcision of young Muslims and Jews has been carried out over the millennia on a global level." The Roman Catholic archbishop of Aachen, Heinrich Mussinghoff, said that the ruling was "very surprising", and the contradiction between "basic rights on freedom of religion and the well-being of the child brought up by the judges is not convincing in this very case." Hans Ulrich Anke, the head of the Protestant Church in Germany, said the ruling should be appealed since it didn't "sufficiently" consider the religious significance of the rite. A spokesman, Steffen Seibert, for German Chancellor Angela Merkel stated that Jewish and Muslim communities will be free to practice circumcision responsibly, and the government would find a way around the local ban in Cologne. The spokesman stated "For everyone in the government it is absolutely clear that we want to have Jewish and Muslim religious life in Germany. Circumcision carried out in a responsible manner must be possible in this country without punishment.". In July, a group of rabbis, imams, and others said that they view the ruling against circumcision "an affront on our basic religious and human rights." The joint statement was signed by leaders of groups including Germany's Turkish-Islamic Union for Religious Affairs, the Islamic Center Brussels, the Rabbinical Centre of Europe, the European Jewish Parliament and the European Jewish Association, who met with members of European Parliament from Germany, Finland, Belgium, Italy, and Poland. European rabbis, who urged Jews to continue circumcision, planned further talks with Muslim and Christian leaders to determine how they can oppose the ban together. The Jewish Hospital of Berlin suspended the practice of male circumcision. On 19 July 2012, a joint resolution of the CDU/CSU, SPD and FDP factions in the Bundestag requesting the executive branch to draft a law permitting circumcision of boys to be performed without unnecessary pain in accordance with best medical practice carried with a broad majority. "The New York Times" reported that the German Medical Association "condemned the ruling for potentially putting children at risk by taking the procedure out of the hands of doctors, but it also warned surgeons not to perform circumcisions for religious reasons until legal clarity was established." The ruling was supported by Deutsche Kinderhilfe, a German child rights organization, which asked for a two-year moratorium to discuss the issue and pointed out that religious circumcision may contravene the Convention on the Rights of the Child (Article 24.3: "States Parties shall take all effective and appropriate measures with a view to abolishing traditional practices prejudicial to the health of children."). The German Academy for Pediatric and Adolescent Medicine (Deutsche Akademie für Kinder- und Jugendmedizin e.V., DAKJ), the German Association for Pediatric Surgery (Deutsche Gesellschaft für Kinderchirurgie, DGKCH) and the Professional Association of Pediatric and Adolescent Physicians (Berufsverband der Kinder- und Jugendärzte) took a firm stand against non-medical routine infant circumcision. In July, in Berlin, a criminal complaint was lodged against Rabbi Yitshak Ehrenberg for "causing bodily harm" by performing religious circumcision, and for vocal support of the continuation of the practice. In September, the prosecutors dismissed the complaint, concluding that "there is no proof to establish that the rabbi's conduct met the 'condition of a criminal' violation." In September, Reuters reported "Berlin's senate said doctors could legally circumcise infant boys for religious reasons in its region, given certain conditions." On 12 December 2012, following a series of hearings and consultations, the Bundestag adopted the proposed law explicitly permitting non-therapeutic circumcision to be performed under certain conditions; it is now §1631(d) in the German Civil Code. The vote tally was 434 ayes, 100 noes, and 46 abstentions. Following approval by the Bundesrat and signing by the Bundespräsident, the new law became effective on 28 December 2012 a day after its publication in the Federal Gazette. In May 2005, Iceland amended its General Penal Code to illegalize female genital mutilation In February 2018, the Progressive Party proposed a bill that would change the words "girl child" to "child" and "her sexual organs" to "[their] sexual organs", thereby making Iceland the first European country to ban male circumcision for non-medical reasons. The bill discussed in the Alþing, the Icelandic parliament, claimed the practice harmed the physical integrity of young boys, was often performed without anaesthesia and in an unhygienic manner by religious leaders instead of medical experts. These facts were deemed incompatible with the United Nations Convention on the Rights of the Child (1990). Critics argued the bill infringed on religious freedom or constituted antisemitism or anti-Muslim bigotry, making it hard for Muslims and Jews to live there. Silja Dögg Gunnarsdóttir, who proposed the ban, retorted that Iceland had already prohibited female circumcision in 2005, and “If we have laws banning circumcision for girls, then we should do so for boys.” On 29 April, the bill was sent back to Parliament for revisions. On 25 March 2018, members of Jews Against Circumcision spoke in the Alþing expressing their support for the proposed ban, dismissing claims that it was motivated by antisemitism, anti-Muslim bigotry, xenophobia or anti-immigration sentiment. Despite rumours to the contrary spread by the "Times of Israel" and others in late April, the bill was not scrapped, but is still a “work in progress.” In October 2005 a Nigerian man was cleared of a charge of reckless endangerment over the death of a baby from hemorrhage and shock after he had circumcised the child. The judge directed the jury not to "bring what he called their white western values to bear when they were deciding this case" and effectively imposed a not guilty verdict on the jury. After deliberating for an hour and a half they found the defendant not guilty. In Israel, Jewish circumcision is entirely legal. Though illegal, female circumcision is still practiced among the Negev Bedouin, and tribal secrecy among the Bedouin makes it difficult for authorities to enforce the ban. In 2013, a Rabbinical court in Israel ordered a mother, Elinor Daniel, to circumcise her son or pay a fine of 500 Israeli Shekel for every day that the child is not circumcised. She appealed against the Rabbinical court ruling and the High Court ruled in her favour stating, among other considerations, the basic right of freedom from religion. The Royal Dutch Medical Association (KNMG) finds non-therapeutic circumcision of male minors to be in conflict with children's right to autonomy and physical integrity, and that there are good reasons for its legal prohibition, as exists for female genital mutilation: In May 2008 a father who had his two sons, aged 3 and 6 circumcised against the will of their mother was found not guilty of abuse as the circumcision was performed by a physician and due to the court's restraint in setting a legal precedent; instead he was given a 6-week suspended jail sentence for taking the boys away from their mother against her will. The Norwegian Ombudsman for Children opposes circumcising children, and it is right to wait until children are old enough to decide for themselves: In June 2012, the centre-right Centre Party proposed a ban on circumcision on males under eighteen, after an Oslo infant died in May following a circumcision. In September 2013, the Children's ombudsmen in all Nordic countries issued a statement by which they called for a ban on circumcision of minors for non-medical reasons, stating that such circumcisions violate the rights of children after the Convention on the Rights of the Child to co-determination and protection from harmful traditions. A bill on ritual circumcision of boys was passed (against two votes) in the Norwegian Parliament in June 2014, with the new law going into effect on 1 January 2015. This law protects the right of Jews to brit mila and obligates the Norwegian Health Care regions to offer the Muslim minority a safe and affordable procedure. Local anaesthesia needs to be applied and a licensed physician needs to be present at the circumcision, which hospitals started to perform in March 2015. In May 2017, the right-wing Progress Party proposed to ban circumcision for males under sixteen. The Children's Act 2005 makes the circumcision of male children under 16 unlawful except for religious or medical reasons. In the Eastern Cape province the Application of Health Standards in Traditional Circumcision Act, 2001, regulates traditional circumcision, which causes the death or mutilation of many youths by traditional surgeons each year. Among other provisions, the minimum age for circumcision is age 18. In 2004, a 22-year-old Rastafarian convert was forcibly circumcised by a group of Xhosa tribal elders and relatives. When he first fled, two police returned him to those who had circumcised him. In another case, a medically circumcised Xhosa man was forcibly recircumcised by his father and community leaders. He laid a charge of unfair discrimination on the grounds of his religious beliefs, seeking an apology from his father and the Congress of Traditional Leaders of South Africa. According to South African newspapers, the subsequent trial became "a landmark case around forced circumcision." In October 2009, the Eastern Cape High Court at Bhisho (sitting as an Equality Court) clarified that circumcision is unlawful unless done with the full consent of the initiate. The Slovenian Human Rights Ombudsman finds that circumcision for non-medical reasons is a violation of children's rights, that ritual circumcision for religious reasons is unacceptable in Slovenia for both legal and ethical reasons, and should not be performed by doctors: The Swedish Medical Association finds no known medical benefits to circumcision of children, and thus strong reasons to wait until the boy is old and mature enough to give informed consent, aiming at ceasing all non-medically justified circumcision without prior consent: In 2001, the Parliament of Sweden enacted a law allowing only persons certified by the National Board of Health to circumcise infants. It requires a medical doctor or an anesthesia nurse to accompany the circumciser and for anaesthetic to be applied beforehand. After the first two months of life circumcisions can only be performed by a physician. The stated purpose of the law was to increase the safety of the procedure. Swedish Jews and Muslims objected to the law, and in 2001, the World Jewish Congress called it "the first legal restriction on Jewish religious practice in Europe since the Nazi era." The requirement for an anaesthetic to be administered by a medical professional is a major issue, and the low degree of availability of certified professionals willing to conduct circumcision has also been subject to criticism. According to a survey, two out of three paediatric surgeons said they refuse to perform non-therapeutic circumcision, and less than half of all county councils offer it in their hospitals. However, in 2006, the U.S. State Department stated, in a report on Sweden, that most Jewish mohels had been certified under the law and 3000 Muslim and 40–50 Jewish boys were circumcised each year. An estimated 2000 of these are performed by persons who are neither physicians nor have officially recognised certification. The Swedish National Board of Health and Welfare reviewed the law in 2005 and recommended that it be maintained, but found that the law had failed with regard to the intended consequence of increasing the safety of circumcisions. A later report by the Board criticised the low level of availability of legal circumcisions, partly due to reluctance among health professionals. To remedy this, the report suggested a new law obliging all county councils to offer non-therapeutic circumcision in their hospitals, but this was later abandoned in favour of a non-binding recommendation. In 2013, the children's ombudsmen of all Nordic countries — Finland, Iceland, Denmark, Sweden and Norway — released a joint declaration in 2013 proposing a ban on non-medical circumcision of male minors. In October 2018, the right-wing populist Sweden Democrats party submitted a draft motion to parliament calling for a ban. At the annual conference of the Centre Party in September 2019, 314 to 166 commissioners voted in favor of prohibiting boys' circumcision. Several Jewish and Islamic organisations voiced their opposition to a potential ban. The Left Party has also expressed support for a prohibition on circumcising boys before the age of 18; other parties have so far not backed a potential ban, though the Green Party found the practice 'problematic'. Male circumcision has traditionally been presumed to be legal under British law, however some authors have argued that there is no solid foundation for this view in English law. While legal, the British Medical Association finds it ethically unacceptable to circumcise a child or young person, either with or without competence, who refuses the procedure, irrespective of the parents’ wishes, and that parental preference alone does not constitute sufficient grounds for performing NTMC on a child unable to express his own view: The passage of the Human Rights Act 1998 has led to some speculation that the lawfulness of the circumcision of male children is unclear. One 1999 case, "Re "J" (child's religious upbringing and circumcision)" said that circumcision in Britain required the consent of all those with parental responsibility (however this comment was not part of the reason for the judgement and therefore is not legally binding), or the permission of the court, acting for the best interests of the child, and issued an order prohibiting the circumcision of a male child of a non-practicing Muslim father and non-practicing Christian mother with custody. The reasoning included evidence that circumcision carried some medical risk; that the operation would be likely to weaken the relationship of the child with his mother, who strongly objected to circumcision without medical necessity; that the child may be subject to ridicule by his peers as the odd one out and that the operation might irreversibly reduce sexual pleasure, by permanently removing some sensory nerves, even though cosmetic foreskin restoration might be possible. The court did not rule out circumcision against the consent of one parent. It cited a hypothetical case of a Jewish mother and an agnostic father with a number of sons, all of whom, by agreement, had been circumcised as infants in accordance with Jewish laws; the parents then have another son who is born after they have separated; the mother wishes him to be circumcised like his brothers; the father for no good reason, refuses his agreement. In such a case, a decision in favor of circumcision was said to be likely. In 2001 the General Medical Council had found a doctor who had botched circumcision operations guilty of abusing his professional position and that he had acted "inappropriately and irresponsibly", and struck him off the register. A doctor who had referred patients to him, and who had pressured a mother into agreeing to the surgery, was also condemned. He was put on an 18-month period of review and retraining, and was allowed to resume unrestricted practice as a doctor in March 2003, after a committee found that he had complied with conditions it placed on him. According to the "Northern Echo", he "told the committee he has now changed his approach to circumcision referrals, accepting that most cases can be treated without the need for surgery.". Fox and Thomson (2005) argue that consent cannot be given for non-therapeutic circumcision. They say there is "no compelling legal authority for the common view that circumcision is lawful." In 2005 a Muslim man had his son circumcised against the wishes of the child's mother who was the custodial parent. In 2009 it was reported that a 20-year-old man whose father had him ritually circumcised as a baby is preparing to sue the doctor who circumcised him. This is believed to be the first time a person who was circumcised as an infant has made a claim in the UK. The case is expected to be heard in 2010. In a 2015 case regarding female circumcision, a judge concluded that non-therapeutic circumcision of male children is a "significant harm". In 2016, the Family Court in Exeter ruled that a Muslim father could not have his two sons (aged 6 and 4) circumcised after their mother disagreed. Mrs Justice Roberts declared that the boys should first grow old enough "to the point where each of the boys themselves will make their individual choices once they have the maturity and insight to appreciate the consequences and longer-term effects of the decisions which they reach." In June 2017, Nottinghamshire Police arrested three people on suspicion of "conspiracy to commit grievous bodily harm". The alleged victim was purportedly circumcised while in its Muslim father's care at his grandparents' in July 2013 without the consent of his mother (a non-religious white British woman who conceived the child after a casual affair with the man, whom she had separated from after the incident). The mother first contacted social services and eventually the police in November 2014. The police initially dismissed the complaint, but after the mother got help from the anti-circumcision group Men Do Complain and leading human rights lawyer Saimo Chahal QC, they reopened the case, and ended up arresting three suspects involved. In November 2017, the Crown Prosecution Service explained to the mother in a letter they were not going to prosecute the doctor, who claimed he was unaware of the mother's non-consent. However, Chahal appealed this decision, which she said "lacks any semblance of a considered and reasoned decision and is flawed and irrational", and threatened to bring the case to court. The by then 29-year-old mother finally sued the doctor in April 2018. Niall McCrae, mental health expert from King's College London, argued that this case could mean 'the end of ritual male circumcision in the UK', drawing comparisons with earlier rulings against female genital mutilation. Circumcision of adults who grant personal informed consent for the surgical operation is legal. In the United States, non-therapeutic circumcision of male children has long been assumed to be lawful in every jurisdiction provided that one parent grants surrogate informed consent. Adler (2013) has recently challenged the validity of this assumption. As with every country, doctors who circumcise children must take care that all applicable rules regarding informed consent and safety are satisfied. While anti-circumcision groups have occasionally proposed legislation banning non-therapeutic child circumcision, it has not been supported in any legislature. After a failed attempt to adopt a local ordinance banning circumcision on a San Francisco ballot, the state of California enacted in October 2011 a law protecting circumcision from local attempts to ban the practice. In 2012, New York City required those performing "metzitzah b'peh", the oral suction of the open circumcision wound required by Hasidim, to obey stringent consent requirements, including documentation. Agudath Israel of America and other Jewish groups have planned to sue the city in response. Disputes between parents Occasionally the courts are asked to make a ruling when parents cannot agree on whether or not to circumcise a child. In January 2001 a dispute between divorcing parents in New Jersey was resolved when the mother, who sought to have the boy circumcised withdrew her request. The boy had experienced two instances of foreskin inflammation and she wanted to have him circumcised. The father, who had experienced a traumatic circumcision as a child objected and they turned to the courts for a decision. The Medical Society of New Jersey and the Urological Society of New Jersey both opposed any court ordered medical treatment. As the parties came to an agreement, no precedent was set. In June 2001 a Nevada court settled a dispute over circumcision between two parents but put a strict gag order on the terms of the settlement. In July 2001 a dispute between parents in Kansas over circumcision was resolved when the mother's request to have the infant circumcised was withdrawn. In this case the father opposed circumcision while the mother asserted that not circumcising the child was against her religious beliefs. (The woman's pastor had stated that circumcision was "important" but was not necessary for salvation.) On 24 July 2001 the parents reached agreement that the infant would not be circumcised. On 14 July 2004 a mother appealed to the Missouri Supreme Court to prevent the circumcision of her son after a county court and the Court of Appeals had denied her a writ of prohibition. However, in early August 2004, before the Supreme Court had given its ruling, the father, who had custody of the boy, had him circumcised. In October 2006 a judge in Chicago granted an injunction blocking the circumcision of a 9-year-old boy. In granting the injunction the judge stated that "the boy could decide for himself whether to be circumcised when he turns 18." In November 2007, the Oregon Supreme Court heard arguments from a divorced Oregon couple over the circumcision of their son. The father wanted his son, who turned 13 on 2 March 2008, to be circumcised in accordance with the father's religious views; the child's mother opposes the procedure. The parents dispute whether the boy is in favor of the procedure. A group opposed to circumcision filed briefs in support of the mother's position, while some Jewish groups filed a brief in support of the father. On 25 January 2008, the Court returned the case to the trial court with instructions to determine whether the child agrees or objects to the proposed circumcision. The father appealed to the US Supreme Court to allow him to have his son circumcised but his appeal was rejected. The case then returned to the trial court. When the trial court interviewed the couple's son, now 14 years old, the boy stated that he did not want to be circumcised. This also provided the necessary circumstances to allow the boy to change residence to live with his mother. The boy was not circumcised. Other disputes In September 2004 the North Dakota Supreme Court rejected a mother's attempt to prosecute her doctor for circumcising her child without fully informing her of the consequences of the procedure. The judge and jury found that the plaintiffs were adequately informed of possible complications, and the jury further found that it is not incumbent on the doctors to describe every "insignificant" risk. In March 2009 a Fulton County, Ga., State Court jury awarded $2.3 million in damages to a 4-year-old boy and his mother for a botched circumcision in which too much tissue was removed causing permanent disfigurement. In August 2010 an eight-day-old boy was circumcised in a Florida hospital against the stated wishes of the parents. The hospital admitted that the boy was circumcised by mistake; the mother has sued the hospital and the doctor involved in the case.
https://en.wikipedia.org/wiki?curid=7447
Context menu A context menu (also called contextual, shortcut, and pop up or pop-up menu) is a menu in a graphical user interface (GUI) that appears upon user interaction, such as a right-click mouse operation. A context menu offers a limited set of choices that are available in the current state, or context, of the operating system or application to which the menu belongs. Usually the available choices are actions related to the selected object. From a technical point of view, such a context menu is a graphical control element. Context menus first appeared in the Smalltalk environment on the Xerox Alto computer, where they were called "pop-up menus"; they were invented by Dan Ingalls in the mid-1970s. Microsoft Office v3.0 introduced the context menu for copy and paste functionality in 1990. Borland demonstrated extensive use of the context menu in 1991 at the Second Paradox Conference in Phoenix Arizona. Lotus 1-2-3/G for OS/2 v1.0 added additional formatting options in 1991. Borland Quattro Pro for Windows v1.0 introduced the Properties context menu option in 1992. Context menus are opened via various forms of user interaction that target a region of the GUI that supports context menus. The specific form of user interaction and the means by which a region is targeted vary: Windows mouse click behavior is such that the context menu doesn't open while the mouse button is pressed, but only opens the menu when the button is released, so the user has to click again (this time with the first mouse button) to select a context menu item. This behavior differs from that of macOS and most free software GUIs. Context menus are sometimes hierarchically organized, allowing navigation through different levels of the menu structure. The implementations differ: Microsoft Word was one of the first applications to only show sub-entries of some menu entries after clicking an arrow icon on the context menu, otherwise executing an action associated with the parent entry. This makes it possible to quickly repeat an action with the parameters of the previous execution, and to better separate options from actions. The following window managers provide context menu functionality: Context menus have received some criticism from usability analysts when improperly used, as some applications make certain features "only" available in context menus, which may confuse even experienced users (especially when the context menus can only be activated in a limited area of the application's client window). Context menus usually open in a fixed position under the pointer, but when the pointer is near a screen edge the menu will be displaced - thus reducing consistency and impeding use of muscle memory. If the context menu is being triggered by keyboard, such as by using Shift + F10, the context menu appears near the focused widget instead of the position of the pointer, to save recognition efforts. Microsoft's guidelines call for always using the term "context menu", and explicitly deprecate "shortcut menu".
https://en.wikipedia.org/wiki?curid=7450
Jews as the chosen people In Judaism, "chosenness" is the belief that the Jews, via descent from the ancient Israelites, are the chosen people, i.e. chosen to be in a covenant with God. The idea of the Israelites being chosen by God is found most directly in the Book of Deuteronomy as the verb "baḥar" (), and is alluded to elsewhere in the Hebrew Bible using other terms such as "holy people". Much is written about these topics in rabbinic literature. The three largest Jewish denominations—Orthodox Judaism, Conservative Judaism and Reform Judaism—maintain the belief that the Jews have been chosen by God for a purpose. Sometimes this choice is seen as charging the Jewish people with a specific mission—to be a light unto the nations, and to exemplify the covenant with God as described in the Torah. This view, however, did not preclude a belief that God has a relationship with other peoples—rather, Judaism held that God had entered into a covenant with all humankind, and that Jews and non-Jews alike have a relationship with God. Biblical references as well as rabbinic literature support this view: Moses refers to the "God of the spirits of all flesh" (), and the Tanakh (Hebrew Bible) also identifies prophets outside the community of Israel. Based on these statements, some rabbis theorized that, in the words of Nethanel ibn Fayyumi, a Yemenite Jewish theologian of the 12th century, "God permitted to every people something he forbade to others...[and] God sends a prophet to every people according to their own language."(Levine, 1907/1966) The Mishnah states that "Humanity was produced from one man, Adam, to show God's greatness. When a man mints a coin in a press, each coin is identical. But when the King of Kings, the Holy One, blessed be He, creates people in the form of Adam not one is similar to any other." (Mishnah Sanhedrin 4:5) The Mishnah continues, and states that anyone who kills or saves a single human, not Jewish, life, has done the same (save or kill) to an entire world. The Tosefta, an important supplement to the Mishnah, also states: "Righteous people of all nations have a share in the world to come" (Sanhedrin 105a). According to the Israel Democracy Institute, approximately two thirds of Israeli Jews believe that Jews are the "chosen people". According to the Bible, Israel's character as the chosen people is unconditional as it says in , The Torah also says, God promises that he will never exchange his people with any other: Other Torah verses about chosenness, The obligation imposed upon the Israelites was emphasized by the prophet Amos (): Sometimes this choice is seen as charging the Jewish people with a specific mission—to be a light unto the nations, and to exemplify the covenant with God as described in the Torah. This view, however, did not preclude a belief that God has a relationship with other peoples—rather, Judaism held that God had entered into a covenant with all humankind, and that Jews and non-Jews alike have a relationship with God. Biblical references as well as rabbinic literature support this view: Moses refers to the "God of the spirits of all flesh" (), and the Tanakh (Hebrew Bible) also identifies prophets outside the community of Israel. Based on these statements, some rabbis theorized that, in the words of Natan'el al-Fayyumi, a Yemenite Jewish theologian of the 12th century, "God permitted to every people something he forbade to others...[and] God sends a prophet to every people according to their own language."(Levine, 1907/1966) The Mishnah states that "Humanity was produced from one man, Adam, to show God's greatness. When a man mints a coin in a press, each coin is identical. But when the King of Kings, the Holy One, blessed be He, creates people in the form of Adam not one is similar to any other." (Mishnah Sanhedrin 4:5) The Mishnah continues, and states that anyone who kills or saves a single human, not Jewish, life, has done the same (save or kill) to an entire world. The Tosefta, a collection of important post-Talmudic discourses, also states: "Righteous people of all nations have a share in the world to come" (Sanhedrin 105a). Most Jewish texts do not state that "God chose the Jews" by itself. Rather, this is usually linked with a mission or purpose, such as proclaiming God's message among all the nations, even though Jews cannot become "unchosen" if they shirk their mission. This implies a special duty, which evolves from the belief that Jews have been pledged by the covenant which God concluded with the biblical patriarch Abraham, their ancestor, and again with the entire Jewish nation at Mount Sinai. In this view, Jews are charged with living a holy life as God's priest-people. In the Jewish prayerbook (the Siddur), chosenness is referred to in a number of ways. The blessing for reading the Torah reads, "Praised are You, Lord our God, King of the Universe, Who has chosen us out of all the nations and bestowed upon us His Torah." In the "Kiddush", a prayer of sanctification, in which the Sabbath is inaugurated over a cup of wine, the text reads, "For you have chosen us and sanctified us out of all the nations, and have given us the Sabbath as an inheritance in love and favour. Praised are you, Lord, who hallows the Sabbath." In the "Kiddush" recited on festivals it reads, "Blessed are You ... who have chosen us from among all nations, raised us above all tongues, and made us holy through His commandments." The Aleinu prayer refers to the concept of Jews as a chosen people: "It is our duty to praise the Master of all, to exalt the Creator of the Universe, who has not made us like the nations of the world and has not placed us like the families of the earth; who has not designed our destiny to be like theirs, nor our lot like that of all their multitude. We bend the knee and bow and acknowledge before the Supreme King of Kings, the Holy One, blessed be he, that it is he who stretched forth the heavens and founded the earth. His seat of glory is in the heavens above; his abode of majesty is in the lofty heights. According to the Rabbis, "Israel is of all nations the most willful or headstrong one, and the Torah was to give it the right scope and power of resistance, or else the world could not have withstood its fierceness." "The Lord offered the Law to all nations; but all refused to accept it except Israel." How do we understand "A Gentile who consecrates his life to the study and observance of the Law ranks as high as the high priest", says R. Meïr, by deduction from Lev. xviii. 5; II Sam. vii. 19; Isa. xxvi. 2; Ps. xxxiii. 1, cxviii. 20, cxxv. 4, where all stress is laid not on Israel, but on man or the righteous one. Maimonides states: "It is now abundantly clear that the pledges Hashem made to Avraham and his descendants would be fulfilled exclusively first in Yitzchak and then in Yaakov, Yitzchak son. This is confirmed by a passage that states, "He is ever mindful of His covenant ... that He made with Avraham, swore to Yitzchak, and confirmed in a decree for Yaakov, for Yisrael, as an eternal covenant." (Tehillim 105: 8,9)." The Gemara states this regarding a non-Jew who studies Torah [his 7 mitzvot] and regarding this, see Shita Mekubetzes, Bava Kama 38a who says that this is an exaggeration. In any case, this statement was not extolling the non-Jew. The Rishonim explain that it is extolling the Torah. Tosfos explains that it uses the example of a "kohen gadol" (high priest), because this statement is based on the verse, ""y'kara hi mipnimim"" (it is more precious than pearls). This is explained elsewhere in the Gemara to mean that the Torah is more precious "pnimim" (translated here as "inside" instead of as "pearls"; thus that the Torah is introspectively absorbed into the person), which refers to "lifnai v'lifnim" (translated as "the most inner of places"), that is the Holy of Holies where the "kahon gadol" went. In any case, in Midrash Rabba (Bamidbar 13:15) this statement is made with an important addition: a non-Jew who converts and studies Torah etc. The Nation of Israel is likened to the olive. Just as this fruit yields its precious oil only after being much pressed and squeezed, so Israel's destiny is one of great oppression and hardship, in order that it may thereby give forth its illuminating wisdom. Poverty is the quality most befitting Israel as the chosen people (Ḥag. 9b). Only on account of its good works is Israel among the nations "as the lily among thorns", or "as wheat among the chaff." Rabbi Lord Immanuel Jakobovits, former Chief Rabbi of the United Synagogue of Great Britain (Modern Orthodox Judaism), described chosenness in this way: "Yes, I do believe that the chosen people concept as affirmed by Judaism in its holy writ, its prayers, and its millennial tradition. In fact, I believe that every people—and indeed, in a more limited way, every individual—is "chosen" or destined for some distinct purpose in advancing the designs of Providence. Only, some fulfill their mission and others do not. Maybe the Greeks were chosen for their unique contributions to art and philosophy, the Romans for their pioneering services in law and government, the British for bringing parliamentary rule into the world, and the Americans for piloting democracy in a pluralistic society. The Jews were chosen by God to be 'peculiar unto Me' as the pioneers of religion and morality; that was and is their national purpose." Modern Orthodox theologian Michael Wyschogrod wrote: "[T]he initial election of Abraham himself was not earned. ... We are simply told that God commanded Abraham to leave his place of birth and go to a land that God would show him. He is also promised that his descendants will become a numerous people. But nowhere does the Bible tell us why Abraham rather than someone else was chosen. The implication is that God chooses whom He wishes and that He owes no accounting to anyone for His choices."
https://en.wikipedia.org/wiki?curid=7451
Chaparral Chaparral () is a shrubland or heathland plant community found primarily in the US state of California and in the northern portion of the Baja California Peninsula, Mexico. It is shaped by a Mediterranean climate (mild, wet winters and hot dry summers) and wildfire, featuring summer-drought-tolerant plants with hard sclerophyllous evergreen leaves, as contrasted with the associated soft-leaved, drought-deciduous, scrub community of coastal sage scrub, found below the chaparral biome. Chaparral covers 5% of the state of California, and of the associated Mediterranean shrubland, an additional 3.5%. The name comes from the Portuguese and Spanish word (from the Basque word ), for evergreen oak shrubland. In its natural state, chaparral is characterized by infrequent fires, with intervals ranging between 30-50+ years. The fires are high intensity and stand-replacing. Recent human activity has increased fire frequency. The shoots of the Chaparral plants do not have thick bark, therefore, they are not very resistant; however, the vegetative recovery is resilient (with long fire-free periods) because its regeneration is rapid from dormant seed banks and root-crowns. There are two kinds of Chaparral plant species: Because sprouters can respond with more rapid growth than seeders, they tend to dominate the early years after a fire. If fires become frequent (<10 years), seeders cannot reach reproductive size before the next fire. System shifts to sprouter-dominance. If fire become even more frequent (< 5 years), sprouters exhaust their below- ground reserves and there will be a conversion to a non-chaparral (grassland) biome. Mature chaparral (stands that have survived for greater intervals between fires) is characterized by nearly impenetrable, dense thickets (except the more open chaparral of the desert). These plants are highly flammable during the late summer and autumn months when conditions are characteristically hot and dry. They grow as woody shrubs with thick, leathery, and often small leaves, contain green leaves all year (are evergreen), and are typically drought resistant (with some exceptions). After the first rains following a fire, the landscape is dominated by small flowering herbaceous plants, known as fire followers, which die back with the summer dry period. Similar plant communities are found in the four other Mediterranean climate regions around the world, including the Mediterranean Basin (where it is known as maquis), central Chile (where it is called matorral), the South African Cape Region (known there as fynbos), and in Western and Southern Australia (as kwongan). According to the California Academy of Sciences, Mediterranean shrubland contains more than 20 percent of the world's plant diversity. The word "chaparral" is a loan word from Spanish , meaning both "small" and "dwarf" evergreen oak (typically holm oak and kermes oak species), which itself comes from a Basque word, , that has the same meaning. Conservation International and other conservation organizations consider chaparral to be a biodiversity hotspot – a biological community with a large number of different species – that is under threat by human activity. The California chaparral and woodlands ecoregion, of the Mediterranean forests, woodlands, and scrub biome, has three sub-ecoregions with ecosystem—plant community subdivisions: For the numerous individual plant and animal species found within the California chaparral and woodlands ecoregion, see: Some of the indicator plants of the California chaparral and woodlands ecoregion include: Another phytogeography system uses two California chaparral and woodlands subdivisions: the cismontane chaparral and the transmontane (desert) chaparral. Cismontane chaparral ("this side of the mountain") refers to the chaparral ecosystem in the Mediterranean forests, woodlands, and scrub biome in California, growing on the western (and coastal) sides of large mountain range systems, such as the western slopes of the Sierra Nevada in the San Joaquin Valley foothills, western slopes of the Peninsular Ranges and California Coast Ranges, and south-southwest slopes of the Transverse Ranges in the Central Coast and Southern California regions. In Central and Southern California chaparral forms a dominant habitat. Members of the chaparral biota native to California, all of which tend to regrow quickly after fires, include: The complex ecology of chaparral habitats supports a very large number of animal species. The following is a short list of birds which are an integral part of the cismontane chaparral ecosystems. Transmontane chaparral or desert chaparral —"transmontane" ("the other side of the mountain") "chaparral"—refers to the desert shrubland habitat and chaparral plant community growing in the rainshadow of these ranges. Transmontane chaparral features xeric desert climate, not Mediterranean climate habitats, and is also referred to as desert chaparral. Desert chaparral is a regional ecosystem subset of the deserts and xeric shrublands biome, with some plant species from the California chaparral and woodlands ecoregion. Unlike cismontane chaparral, which forms dense, impenetrable stands of plants, desert chaparral is open, with only about 50 percent of the ground covered. Individual shrubs can reach up to in height. Transmontane chaparral or desert chaparral is found on the eastern slopes of major mountain range systems on the western sides of the deserts of California. The mountain systems include the southeastern Transverse Ranges (the San Bernardino and San Gabriel Mountains) in the Mojave Desert north and northeast of the Los Angeles basin and Inland Empire; and the northern Peninsular Ranges (San Jacinto, Santa Rosa, and Laguna Mountains), which separate the Colorado Desert (western Sonoran Desert) from lower coastal Southern California. It is distinguished from the cismontane chaparral found on the coastal side of the mountains, which experiences higher winter rainfall. Naturally, desert chaparral experiences less winter rainfall than cismontane chaparral. Plants in this community are characterized by small, hard (sclerophyllic) evergreen (non-deciduous) leaves. Desert chaparral grows above California's desert cactus scrub plant community and below the pinyon-juniper woodland. It is further distinguished from the deciduous sub-alpine scrub above the pinyon-juniper woodlands on the same side of the Peninsular ranges. Transmontane (desert) chaparral typically grows on the lower ( elevation) northern slopes of the southern Transverse Ranges (running east to west in San Bernardino and Los Angeles counties) and on the lower () eastern slopes of the Peninsular Ranges (running south to north from lower Baja California to Riverside and Orange counties and the Transverse Ranges). It can also be found in higher-elevation sky islands in the interior of the deserts, such as in the upper New York Mountains within the Mojave National Preserve in the Mojave Desert. The California transmontane (desert) chaparral is found in the rain shadow deserts of the following: There is overlap of animals with those of the adjacent desert and pinyon-juniper communities. Chaparral is a coastal biome with hot, dry summers and mild, rainy winters. The Chaparral area receives about of precipitation a year. This makes the chaparral most vulnerable to fire in the late summer and fall. The chaparral ecosystem as a whole is adapted to be able to recover from infrequent wildfires (historically these communities had a fire every 30-50 years); indeed, chaparral regions are known culturally and historically for their impressive fires. (This does create a conflict with human development adjacent to and expanding into chaparral systems.) Additionally, Native Americans burned chaparral to promote grasslands for textiles and food. Before a major fire, typical chaparral plant communities are dominated by manzanita, chamise (also called greasewood; "Adenostoma fasciculatum") and "Ceanothus" species, toyon (which can sometimes be interspersed with scrub oaks), and other drought-resistant shrubs with hard (sclerophyllous) leaves; these plants resprout (see resprouter) from underground burls after a fire. The shoots of these plants are, however, not resistant to chaparral crown-fire regimes as the bark is simply not thick enough. Plants that are long-lived in the seed bank or serotenous with induced germination after fire include chamise", Ceanothus," and fiddleneck"." Some chaparral plant communities may grow so dense and tall that it becomes difficult for large animals and humans to penetrate, but may be teeming with smaller fauna in the understory. Many chaparral plant species require some fire cue (heat, smoke, or charred wood, and chemical changes in the soil following fires) for germination. Others, such as annual and herbaceous species like "Phacelia" require fires to allow sunlight to reach them, and are known as fire followers. During the time shortly after a fire, chaparral communities may contain soft-leaved herbaceuous annual plants that dominate the community for the first few years – until the burl resprouts and seedlings of chaparral perennials create an overstory, blocking the sunlight from other plants in the community. When the overstory regrows, seeds of annuals and smaller plants may lie dormant until the next fire creates the conditions required for germination. Mid-sized plants such as "Ceonothus" fix nitrogen, while others cannot, which, together with the need for exposure to the sun, creates a symbiotic relationship of the entire community with infrequent fires. Because of the hot, dry conditions that exist in the California summer and fall, chaparral is one of the most fire-prone plant communities in North America. Some fires are caused by lightning, but these are usually during periods of high humidity and low winds and are easily controlled. Nearly all of the very large wildfires are caused by human activity during periods of very hot, dry easterly Santa Ana winds. These man-made fires are commonly caused by power line failures, vehicle fires and collisions, sparks from machinery, arson, or campfires. In natural Chaparral communities without human interference, the fires are ignition-prone as there are plenty of ground fuels and the temperatures are fire-permitting during the dry season. Though adapted to infrequent fires, an increased frequency of fire can modify the community. A moderate frequency of fire (less than ten years) will result in the loss of seeder plants such as This moderate frequency disallows seeder plants to reach their reproductive size before the next fire and the community shifts to a sprouter-dominance. High frequency fires (less than five years) can cause the additional loss of sprouters by exhausting their reserves below-ground. Changes in fire frequency can also cause a phenomenon known as disclimax, which is when natural succession of plants is constantly interrupted or arrested at an intermediate stage. Today, frequent accidental ignitions can convert chaparral from a native shrubland to non-native annual grassland and drastically reduce species diversity, especially under drought brought about by climate change. There are two assumptions relating to California chaparral fire regimes that appear to have caused considerable debate, and sometimes confusion and controversy, within the fields of wildfire ecology and land management. The perspective that older chaparral is unhealthy or unproductive may have originated during the 1940s when studies were conducted measuring the amount of forage available to deer populations in chaparral stands. However, according to recent studies, California chaparral is extraordinarily resilient to very long periods without fire and continues to maintain productive growth throughout pre-fire conditions. Seeds of many chaparral plants actually require 30 years or more worth of accumulated leaf litter before they will successfully germinate (e.g., scrub oak, "Quercus berberidifolia"; toyon, "Heteromeles arbutifolia"; and holly-leafed cherry, "Prunus ilicifolia"). When intervals between fires drop below 10 to 15 years, many chaparral species are eliminated and the system is typically replaced by non-native, invasive, weedy grassland. The idea that older chaparral is responsible for causing large fires was originally proposed in the 1980s by comparing wildfires in Baja California and southern California . It was suggested that fire suppression activities in southern California allowed more fuel to accumulate, which in turn led to larger fires (in Baja, fires often burn without active suppression efforts ). This is similar to the argument that fire suppression in western United States has allowed ponderosa pine forests to become "overstocked". In the past, surface fires burned through these forests at intervals of anywhere between 4 and 36 years, clearing out the understory and creating a more ecologically balanced system. However, chaparral has a crown-fire regime, meaning that fires consume the entire system whenever they burn, with a historical frequency of 30 to 50 years. In one study, a detailed analysis of historical fire data concluded that fire suppression activities have been ineffective at excluding fire from southern California chaparral, unlike in ponderosa pine forests. In addition, the number of fires is increasing in step with population growth. Chaparral stand age does not have a significant correlation to its tendency to burn. Low humidity, low fuel moisture, and high winds appear to be the primary factors in determining when and where a chaparral fire occurs and how large it becomes. Fires can be beneficial to plant communities by clearing away canopies of litter, inducing serotenous germination, and sanitizing the soils from pathogens.
https://en.wikipedia.org/wiki?curid=7455
Clipper A clipper was a type of mid-19th-century merchant sailing vessel, designed for speed. Clippers were generally narrow for their length, small by later 19th century standards, could carry limited bulk freight, and had a large total sail area. "Clipper" does not refer to a specific sailplan; clippers, by sailplan, may be schooners, brigs, brigantines, etc., or indeed "ships" as restrictively defined in the Age of Sail. Clipper ships were mostly constructed in British and American shipyards, though France, Brazil, the Netherlands and other nations also produced some. Clippers sailed all over the world, primarily on the trade routes between the United Kingdom and China, in transatlantic trade, and on the New York-to-San Francisco route around Cape Horn during the California Gold Rush. Dutch clippers were built beginning in the 1850s for the tea trade and passenger service to Java. The boom years of the clipper ship era began in 1843 in response to a growing demand for faster delivery of tea from China. This continued under the stimulating influence of the discovery of gold in California and Australia in 1848 and 1851, and ended with the opening of the Suez Canal in 1869. While composite iron-framed wooden clippers continued to be built into the 1870s, sailing ships of the next generation had iron hulls. The last full-rigged composite passenger clipper, "Torrens", was launched in 1875, while iron-hulled clippers continued to be built for the Australian wool trade into the 1890s. The term "clipper" most likely derives from the verb "clip", which in former times meant, among other things, to run or fly swiftly. Dryden, the English poet, used the word "clip" to describe the swift flight of a falcon in the 17th century when he said "And, with her eagerness the quarry missed, Straight flies at check, and clips it down the wind." The ships appeared to clip along the ocean water. The term "clip" became synonymous with "speed" and was also applied to fast horses and sailing ships. "To clip it," and "going at a good clip," remained familiar expressions in the early 20th century. The first application of the term "clipper", in a nautical sense, is uncertain. The Baltimore Clipper was in use over the last quarter of the 18th century through to the first half of the 19th century, but under a different name for much of that time. At first, these vessels were referred to as "Virginia-built" or "pilot-boat model"—with the name "Baltimore-built" appearing during the War of 1812. It was not until the final days of the slave trade ("c." 1835 -1850)—just as the type was dying out—that the name "Baltimore Clipper" became common. The retrospective application of the word "clipper" to this type has been a source of confusion. The Oxford English Dictionary's earliest quote (referring to the Baltimore Clipper) is from 1824. British newspaper usage of the term can be found as early as 1832 and in shipping adverts from 1835. There is discussion of clippers being faster than other vessels by captains called before court in "A report of the trial of Pedro Gibert et al before the United States Circuit Court" of 1834;"Edward H. Faucon": I have been to sea twelve years. I am master of a vessel, and...I should think there would be thirty per cent difference in favor of the clipper... "Samuel Austin Turner": I...know the Mexican. Should think, in a royal breeze, she would run six knots, while a clipper would sail one third faster. In a fresh, fair wind, the difference would be smaller—perhaps none at all. Don’t think the brig would ever have the advantage of the clipper. There is no single definition of the characteristics of a clipper ship, but mariner and author Alan Villiers describes them as follows:To sailors, three things made a ship a clipper. She must be sharp-lined, built for speed. She must be tall-sparred and carry the utmost spread of canvas. And she must "use" that sail, day and night, fair weather and foul. Optimized for speed, clippers were too fine-lined to carry much cargo. They typically carried extra sails, such as skysails and moonrakers on the masts, and studding sails on booms extending out from the hull or yards, which required extra sailors to handle them. In high winds where other ships would shorten sail, clippers drove on, heeling so much that their lee rails were in the water. The first ships to which the term "clipper" seems to have been applied were the Baltimore clippers, developed in Chesapeake Bay before the American Revolution, and reaching their zenith between 1795 and 1815. They were small, rarely exceeding 200 tons OM. Their hulls were sharp ended and displayed a lot of deadrise. They were rigged as schooners, brigs or brigantines. In the War of 1812 some were lightly armed, sailing under Letters of Marque and Reprisal, when the type—exemplified by "Chasseur", launched at Fells Point, Baltimore in 1814—became known for her incredible speed; the deep draft enabled the Baltimore clipper to sail close to the wind. Clippers, running the British blockade of Baltimore, came to be recognized for speed rather than cargo space. Speed was also required for the Chinese opium trade between India and China. Small, sharp-bowed British vessels, called opium clippers, were the result. Meanwhile, Baltimore clippers continued to be built, in particular specifically for the opium trade between India and China, a trade that had become unprofitable for American shipowners by 1849. "Ann McKim", built in Baltimore in 1833 by the Kennard & Williamson shipyard, is considered to be the original clipper ship. She measured 494 tons OM, and was built on the enlarged lines of a Baltimore clipper, with sharply raked stem, counter stern and square rig. Although "Ann McKim" was the first large clipper ship ever constructed, it cannot be said that she founded the clipper ship era, or even that she directly influenced shipbuilders, since no other ship was built like her; but she may have suggested the clipper design in vessels of ship rig. She did, however, influence the building of "Rainbow" in 1845, the first extreme clipper ship. In Aberdeen, Scotland, the shipbuilders Alexander Hall and Sons developed the "Aberdeen" clipper bow in the late 1830s: the first was "Scottish Maid" launched in 1839. "Scottish Maid", 150 tons OM, was the first British clipper ship. ""Scottish Maid" was intended for the Aberdeen-London trade, where speed was crucial to compete with steamships. The Hall brothers tested various hulls in a water tank and found the clipper design most effective. The design was influenced by tonnage regulations. Tonnage measured a ship's cargo capacity and was used to calculate tax and harbour dues. The new 1836 regulations measured depth and breadth with length measured at half midship depth. Extra length above this level was tax-free and became a feature of clippers. "Scottish Maid" proved swift and reliable and the design was widely copied." The earliest British clipper ships were built for trade within the British Isles ("Scottish Maid" was built for the Aberdeen to London trade). Then followed the vast clipper trade of tea, opium, spices and other goods from the Far East to Europe, and the ships became known as "tea clippers". From 1839, larger American clipper ships started to be built beginning with "Akbar", 650 tons OM, in 1839, and including the 1844-built Houqua, 581 tons OM. These larger vessels were built predominantly for use in the China tea trade and known as "tea clippers". Smaller clipper vessels also continued to be built predominantly for the China opium trade and known as "opium clippers" such as the 1842-built "Ariel", 100 tons OM. Then in 1845 "Rainbow", 757 tons OM, the first extreme clipper was launched in New York. These American clippers were larger vessels designed to sacrifice cargo capacity for speed. They had a bow lengthened above the water, a drawing out and sharpening of the forward body, and the greatest breadth further aft. Extreme clippers were built in the period 1845 to 1855. In 1851, shipbuilders in Medford, Massachusetts built the "Antelope", often called the "Antelope of Boston" to distinguish her from other ships of the same name. This vessel is sometimes called one of the first medium clipper ships. A ship-design journalist noted that "the design of her model was to combine large stowage capacity with good sailing qualities." The "Antelope" was relatively flat-floored and had only an 8-inch dead rise at half floor. The medium clipper, though still very fast, could carry more cargo. After 1854 extreme clippers were replaced in American shipbuilding yards by medium clippers. The "Flying Cloud" was a clipper ship that claimed the fastest passage between New York and San Francisco, 89 days 8 hours. (The other contender for this "blue ribbon" title was the medium clipper "Andrew Jackson"—there is an unresolvable argument over timing these voyages "from pilot to pilot"). "Flying Cloud" was the most famous of the clippers built by Donald McKay. She was known for her extremely close race with the "Hornet" in 1853; for having a woman navigator, Eleanor Creesy, wife of Josiah Perkins Creesy, who skippered the "Flying Cloud" on two record-setting voyages from New York to San Francisco; and for sailing in the Australia and timber trades. Clipper ships largely ceased being built in American shipyards in 1859 when, unlike the earlier boom years, only four clipper ships were built; a few were built in the 1860s. The last American clipper ship was "the "Pilgrim"" launched in 1873 from the shipyards of Medford, Massachusetts, built by Joshua T. Foster. Among shipowners of the day, “Medford-built” came to mean the best. British clipper ships continued to be built after 1859. Earlier British clipper ships had become known as extreme clippers, and were considered to be "as sharp as the American"-built ships. From 1859 a new design was developed for British clipper ships that was nothing like the American clippers; these ships continued to be called extreme clippers. The new design had a sleek graceful appearance, less sheer, less freeboard, lower bulwarks, and smaller breadth. They were built for the China tea trade, starting with "Falcon" in 1859, and continuing until 1870. It is estimated that 25 to 30 of these ships were built, with no more than four or five in a year. The earlier ships were made from wood, though some were made from iron, just as some British clippers had been made from iron prior to 1859. In 1863 the first tea clippers of composite construction were brought out, combining the best of both worlds. Composite clippers had the strength of iron spars with wooden hulls, and copper sheathing could be added to prevent the fouling that occurred on iron hulls. After 1869, with the opening of the Suez Canal that greatly advantaged steam vessels (see Decline below), the tea trade collapsed for clippers. From the late 1860s until the early 1870s the clipper trade increasingly focused on the Britain to Australia and New Zealand route, carrying goods and immigrants, services that had begun earlier with the Australian Gold Rush of the 1850s. British-built clipper ships and many American-built British-owned ships were used. Even in the 1880s, sailing ships were still the main carriers of cargoes between Britain, and Australia and New Zealand. This trade eventually became unprofitable, and the ageing clipper fleet became unseaworthy. Among the most notable clippers were the China clippers, also called tea clippers or opium clippers, designed to ply the trade routes between Europe and the East Indies. The last example of these still in reasonable condition is "Cutty Sark", preserved in dry dock at Greenwich, United Kingdom. Damaged by fire on 21 May 2007 while undergoing conservation, the ship was permanently elevated three metres above the dry dock floor in 2010 as part of a plan for long-term preservation. Before the early 18th century, the East India Company paid for its tea mainly in silver. When the Chinese Emperor chose to embargo European manufactured commodities and demand payment for all Chinese goods in silver, the price rose, restricting trade. The East India Company began to produce something desired by the Chinese as much as tea was by the British: opium. This had a significant influence on both India and China. Opium was also imported into Britain and was not prohibited because it was thought to be medically beneficial. Laudanum, made from opium, was used as a pain killer, to induce sleep and to suppress anxiety. The famous literary opium addicts Thomas De Quincey, Samuel Taylor Coleridge and Wilkie Collins also took it for its pleasurable effects. The Limehouse area in London was notorious for its opium dens, many of which catered for Chinese sailors as well as English addicts. Clippers were built for seasonal trades such as tea, where an early cargo was more valuable, or for passenger routes. One passenger ship survives, the "City of Adelaide" designed by William Pile of Sunderland. The fast ships were ideally suited to low-volume, high-profit goods, such as tea, opium, spices, people, and mail. The return could be spectacular. The "Challenger" returned from Shanghai with "the most valuable cargo of tea and silk ever to be laden in one bottom". Competition among the clippers was public and fierce, with their times recorded in the newspapers. The ships had short expected lifetimes, and rarely lasted more than twenty years before they were broken up for salvage. Given their speed and maneuverability, many clippers were fitted with cannon or carronades and used for piracy, privateering, smuggling, or interdiction service. The last China clippers were acknowledged as the fastest sail vessels. When fully rigged and riding a tradewind, they had peak average speeds of over . The Great Tea Race of 1866 showcased their speed. China clippers are the fastest commercial sailing vessels ever made; their speeds have often been exceeded by modern yachts, but never by a commercial sail vessel. Only the fastest windjammers could attain similar speeds. The 24h record of the "Champion of the Seas" wasn't broken until 1984 (by a multihull), or 2001 (by another monohull). The American clippers sailing from the East Coast to the California goldfields were working in a booming market. Freight rates were high everywhere in the first years of the 1850s. This started to fade in late 1853. The ports of California and Australia reported that they were overstocked with goods that had been shipped earlier in the year. This gave an accelerating fall in freight rates that was halted, however, by the start of the Crimean War in March 1854, as many ships were now being chartered by the French and British governments. The end of the Crimean War in April 1856 released all this capacity back on the world shipping markets - the result being a severe slump. The next year saw the Panic of 1857, which had effects on both sides of the Atlantic. The USA was just starting to recover from this in 1861 when the American Civil War started, causing significant disruption to trade in both Union and Confederate states. As the economic situation deteriorated in 1853, American shipowners either did not order new vessels, or specified an ordinary clipper or a medium clipper instead of an extreme clipper. No extreme clipper was launched in an American shipyard after the end of 1854 and only a few medium clippers after 1860. By contrast, British trade recovered well at the end of the 1850s. Tea clippers had continued to be launched during the depressed years, apparently little affected by the economic downturn. The long-distance route to China was not realistically challenged by steamships in the early part of the 1860s. No true steamer (as opposed to an auxiliary steamship) had the fuel efficiency to carry sufficient cargo to make a profitable voyage. The auxiliary steamships struggled to make any profit. The situation changed in 1866 when the Alfred Holt-designed and owned "SS Agamemnon" made her first voyage to China. Holt had persuaded the Board of Trade to allow higher steam pressures in British merchant vessels. Running at 60 psi instead of the previously permitted 25 psi, and using an efficient compound engine, "Agamemnon" had the fuel efficiency to steam at 10 knots to China and back, with coaling stops at Mauritius on the outward and return legs - crucially carrying sufficient cargo to make a profit. In 1869 the Suez Canal opened, giving steamships a route about 3,000 NM shorter than that taken by sailing ships round the Cape of Good Hope. Despite initial conservatism by tea merchants, by 1871 tea clippers found strong competition from steamers in the tea ports of China. A typical passage time back to London for a steamer was 58 days, while the very fastest clippers could occasionally make the trip in less than 100 days; the average was 123 days in the 1867–68 tea season. The freight rate for a steamer in 1871 was roughly double that paid to a sailing vessel. Some clipper owners were severely caught out by this: several extreme clippers had been launched in 1869, including "Cutty Sark", "Norman Court" and "the Caliph". Of the many clipper ships built during the mid-19th century, only two are known to survive. The only intact survivor is "Cutty Sark", which was preserved as a museum ship in 1954 at Greenwich for public display. The other known survivor is "City of Adelaide", unlike "Cutty Sark" she was reduced to a hulk over the years. She eventually sank at her moorings in 1991, but was raised the following year and remained on dry land for years. Adelaide (a.k.a. S.V. "Carrick") is the older of the two survivors, and was transported to Australia for conservation. Departures of clipper ships, mostly from New York and Boston to San Francisco, were advertised by clipper ship sailing cards. These cards, slightly larger than today’s postcards, were produced by letterpress and wood engraving on coated card stock. Most clipper cards were printed in the 1850s and 1860s, and represented the first pronounced use of color in American advertising art. Relatively few (perhaps 3,500) cards survive today. With their rarity and importance as artifacts of nautical, Western, and printing history, clipper cards are highly prized by both private collectors and institutions.
https://en.wikipedia.org/wiki?curid=7461
Clive Anderson Clive Stuart Anderson (born 10 December 1952) is an English television and radio presenter, comedy writer and former barrister. Winner of a British Comedy Award in 1991, Anderson began experimenting with comedy and writing comedic scripts during his 15-year legal career, before starring in "Whose Line Is It Anyway?" on BBC Radio 4, then later Channel 4. He has also hosted many radio programmes, and made guest appearances on "Have I Got News for You", "Mock the Week" and "QI". Anderson was educated at Stanburn Primary School and Harrow County School for Boys where his group of friends included Geoffrey Perkins and Michael Portillo. His Scottish father was manager of the Bradford & Bingley's Wembley branch. Anderson attended Selwyn College, Cambridge, where, from 1974 to 1975, he was President of Footlights. He was called to the bar at the Middle Temple in 1976 and became a practising barrister, specialising in criminal law. While still practising law, he continued performing, including taking a show to the Edinburgh Fringe in 1981 with Griff Rhys Jones. Anderson was involved in the fledgling alternative comedy scene in the early 1980s and was the first act to come on stage at The Comedy Store when it opened in 1979. He made his name as host of the improvised television comedy show "Whose Line Is It Anyway?", which ran for 10 series. Anderson hosted his own chat-show, "Clive Anderson Talks Back", on Channel 4, which ran for 10 series. Anderson moved to the BBC in 1996. The show's name was changed to "Clive Anderson All Talk" and it was aired on BBC One. In one incident in 1997, Anderson interviewed the Bee Gees, and throughout the interview he repeatedly joked about their songs from the "Saturday Night Fever" era, also referring to their original intended name as 'Les Tosseurs' but his comment, 'you'll always be tossers to me' ultimately prompted the band to walk out of the interview. Anderson once had a glass of water poured over his head by a perturbed Richard Branson, to which Anderson remarked "I'm used to that; I've flown Virgin". He also said to Jeffrey Archer, "Is there no beginning to your talents?" Archer retorted that "The old ones are always the best," for Anderson to reply "Yes, I've read your books." The last series of "Clive Anderson All Talk" aired in 2001. He has made ten appearances on "Have I Got News for You". He has also frequently appeared on "QI". In 2007, he featured as a regular panellist on the ITV comedy show "News Knight". One heated exchange on "Have I Got News for You" occurred when he joked to fellow guest Piers Morgan that the "Daily Mirror" was now, thanks to Morgan (then its editor), almost as good as "The Sun". When asked by Morgan, "What do you know about editing newspapers?", he swiftly replied, "About as much as you do." In 2005, he presented the short-lived quiz "Back in the Day" for Channel 4. On 25 February 2008, he started to present "Brainbox Challenge", a new game show, for BBC Two. Later that year, he presented a reality TV talent show-themed television series produced by the BBC entitled "Maestro", starring eight celebrities. In 2009, Anderson was the television host of the BBC's "Last Night of the Proms". Anderson presents legal show "Unreliable Evidence" on Radio 4. He also covered the Sunday morning 11 am-1 pm show on BBC Radio 2 through the end of January 2008. It was announced in April 2008 that Anderson, who had previously filled in for host Ned Sherrin from 2006 until Sherrin's death in 2007, would be taking over as permanent host of "Loose Ends". He also hosted six series of "Clive Anderson's Chat Room" on BBC Radio 2 from 2004–2009. Anderson has appeared on BBC Radio 4's "The Unbelievable Truth" hosted by David Mitchell. Anderson also presents the radio show "The Guessing Game" on BBC Radio Scotland. Anderson has also appeared on BBC Radio 5 Live's "Fighting Talk". Anderson is a comedy sketch writer who has written for Frankie Howerd, "Not the Nine O'Clock News", and Griff Rhys Jones and Mel Smith. One of his early comedy writing projects was "Black Cinderella Two Goes East" with Rory McGrath for BBC Radio 4 in 1978. As well as writing comedy, Anderson is also a frequent contributor to newspapers, and was a regular columnist for "The Sunday Correspondent". Anderson lives in Highbury, North London, with his wife, Jane, with whom he has three children. His daughter Flora is a comedian. He supports Arsenal, Rangers and Albion Rovers and is President of the Woodland Trust and Vice Patron of the Solicitors' Benevolent Association. He also has a holiday home in Dalmally, Argyll. The show "Whose Line is it Anyway?" won a BAFTA award in 1990. Later, Anderson won both the "Top Entertainment Presenter" and "Top Radio Comedy Personality" at the British Comedy Awards in 1991.
https://en.wikipedia.org/wiki?curid=7462
Cold fusion Cold fusion is a hypothesized type of nuclear reaction that would occur at, or near, room temperature. It would contrast starkly with the "hot" fusion that is known to take place naturally within stars and artificially in hydrogen bombs and prototype fusion reactors under immense pressure and at temperatures of millions of degrees, and be distinguished from muon-catalyzed fusion. There is currently no accepted theoretical model that would allow cold fusion to occur. In 1989, two electrochemists, Martin Fleischmann and Stanley Pons, reported that their apparatus had produced anomalous heat ("excess heat") of a magnitude they asserted would defy explanation except in terms of nuclear processes. They further reported measuring small amounts of nuclear reaction byproducts, including neutrons and tritium. The small tabletop experiment involved electrolysis of heavy water on the surface of a palladium (Pd) electrode. The reported results received wide media attention and raised hopes of a cheap and abundant source of energy. Many scientists tried to replicate the experiment with the few details available. Hopes faded due to the large number of negative replications, the withdrawal of many reported positive replications, the discovery of flaws and sources of experimental error in the original experiment, and finally the discovery that Fleischmann and Pons had not actually detected nuclear reaction byproducts. By late 1989, most scientists considered cold fusion claims dead, and cold fusion subsequently gained a reputation as pathological science. In 1989 the United States Department of Energy (DOE) concluded that the reported results of excess heat did not present convincing evidence of a useful source of energy and decided against allocating funding specifically for cold fusion. A second DOE review in 2004, which looked at new research, reached similar conclusions and did not result in DOE funding of cold fusion. A small community of researchers continues to investigate cold fusion, now often preferring the designation low-energy nuclear reactions (LENR) or condensed matter nuclear science (CMNS). Since articles about cold fusion are rarely published in peer-reviewed mainstream scientific journals any longer, they do not attract the level of scrutiny expected for mainstream scientific publications. Nuclear fusion is normally understood to occur at temperatures in the tens of millions of degrees. This is called "thermonuclear fusion". Since the 1920s, there has been speculation that nuclear fusion might be possible at much lower temperatures by catalytically fusing hydrogen absorbed in a metal catalyst. In 1989, a claim by Stanley Pons and Martin Fleischmann (then one of the world's leading electrochemists) that such cold fusion had been observed caused a brief media sensation before the majority of scientists criticized their claim as incorrect after many found they could not replicate the excess heat. Since the initial announcement, cold fusion research has continued by a small community of researchers who believe that such reactions happen and hope to gain wider recognition for their experimental evidence. The ability of palladium to absorb hydrogen was recognized as early as the nineteenth century by Thomas Graham. In the late 1920s, two Austrian born scientists, Friedrich Paneth and Kurt Peters, originally reported the transformation of hydrogen into helium by nuclear catalysis when hydrogen was absorbed by finely divided palladium at room temperature. However, the authors later retracted that report, saying that the helium they measured was due to background from the air. In 1927 Swedish scientist John Tandberg reported that he had fused hydrogen into helium in an electrolytic cell with palladium electrodes. On the basis of his work, he applied for a Swedish patent for "a method to produce helium and useful reaction energy". Due to Paneth and Peters's retraction and his inability to explain the physical process, his patent application was denied. After deuterium was discovered in 1932, Tandberg continued his experiments with heavy water. The final experiments made by Tandberg with heavy water were similar to the original experiment by Fleischmann and Pons. Fleischmann and Pons were not aware of Tandberg's work. The term "cold fusion" was used as early as 1956 in a "New York Times" article about Luis Alvarez's work on muon-catalyzed fusion. Paul Palmer and then Steven Jones of Brigham Young University used the term "cold fusion" in 1986 in an investigation of "geo-fusion", the possible existence of fusion involving hydrogen isotopes in a planetary core. In his original paper on this subject with Clinton Van Siclen, submitted in 1985, Jones had coined the term "piezonuclear fusion". The most famous cold fusion claims were made by Stanley Pons and Martin Fleischmann in 1989. After a brief period of interest by the wider scientific community, their reports were called into question by nuclear physicists. Pons and Fleischmann never retracted their claims, but moved their research program to France after the controversy erupted. Martin Fleischmann of the University of Southampton and Stanley Pons of the University of Utah hypothesized that the high compression ratio and mobility of deuterium that could be achieved within palladium metal using electrolysis might result in nuclear fusion. To investigate, they conducted electrolysis experiments using a palladium cathode and heavy water within a calorimeter, an insulated vessel designed to measure process heat. Current was applied continuously for many weeks, with the heavy water being renewed at intervals. Some deuterium was thought to be accumulating within the cathode, but most was allowed to bubble out of the cell, joining oxygen produced at the anode. For most of the time, the power input to the cell was equal to the calculated power leaving the cell within measurement accuracy, and the cell temperature was stable at around 30 °C. But then, at some point (in some of the experiments), the temperature rose suddenly to about 50 °C without changes in the input power. These high temperature phases would last for two days or more and would repeat several times in any given experiment once they had occurred. The calculated power leaving the cell was significantly higher than the input power during these high temperature phases. Eventually the high temperature phases would no longer occur within a particular cell. In 1988 Fleischmann and Pons applied to the United States Department of Energy for funding towards a larger series of experiments. Up to this point they had been funding their experiments using a small device built with $100,000 out-of-pocket. The grant proposal was turned over for peer review, and one of the reviewers was Steven Jones of Brigham Young University. Jones had worked for some time on muon-catalyzed fusion, a known method of inducing nuclear fusion without high temperatures, and had written an article on the topic entitled "Cold nuclear fusion" that had been published in "Scientific American" in July 1987. Fleischmann and Pons and co-workers met with Jones and co-workers on occasion in Utah to share research and techniques. During this time, Fleischmann and Pons described their experiments as generating considerable "excess energy", in the sense that it could not be explained by chemical reactions alone. They felt that such a discovery could bear significant commercial value and would be entitled to patent protection. Jones, however, was measuring neutron flux, which was not of commercial interest. To avoid future problems, the teams appeared to agree to publish their results simultaneously, though their accounts of their 6 March meeting differ. In mid-March 1989, both research teams were ready to publish their findings, and Fleischmann and Jones had agreed to meet at an airport on 24 March to send their papers to "Nature" via FedEx. Fleischmann and Pons, however, pressured by the University of Utah, which wanted to establish priority on the discovery, broke their apparent agreement, disclosing their work at a press conference on 23 March (they claimed in the press release that it would be published in "Nature" but instead submitted their paper to the "Journal of Electroanalytical Chemistry"). Jones, upset, faxed in his paper to "Nature" after the press conference. Fleischmann and Pons' announcement drew wide media attention. But the 1986 discovery of high-temperature superconductivity had made the scientific community more open to revelations of unexpected scientific results that could have huge economic repercussions and that could be replicated reliably even if they had not been predicted by established theories. Many scientists were also reminded of the Mössbauer effect, a process involving nuclear transitions in a solid. Its discovery 30 years earlier had also been unexpected, though it was quickly replicated and explained within the existing physics framework. The announcement of a new purported clean source of energy came at a crucial time: adults still remembered the 1973 oil crisis and the problems caused by oil dependence, anthropogenic global warming was starting to become notorious, the anti-nuclear movement was labeling nuclear power plants as dangerous and getting them closed, people had in mind the consequences of strip mining, acid rain, the greenhouse effect and the Exxon Valdez oil spill, which happened the day after the announcement. In the press conference, Chase N. Peterson, Fleischmann and Pons, backed by the solidity of their scientific credentials, repeatedly assured the journalists that cold fusion would solve environmental problems, and would provide a limitless inexhaustible source of clean energy, using only seawater as fuel. They said the results had been confirmed dozens of times and they had no doubts about them. In the accompanying press release Fleischmann was quoted saying: "What we have done is to open the door of a new research area, our indications are that the discovery will be relatively easy to make into a usable technology for generating heat and power, but continued work is needed, first, to further understand the science and secondly, to determine its value to energy economics." Although the experimental protocol had not been published, physicists in several countries attempted, and failed, to replicate the excess heat phenomenon. The first paper submitted to "Nature" reproducing excess heat, although it passed peer review, was rejected because most similar experiments were negative and there were no theories that could explain a positive result; this paper was later accepted for publication by the journal "Fusion Technology". Nathan Lewis, professor of chemistry at the California Institute of Technology, led one of the most ambitious validation efforts, trying many variations on the experiment without success, while CERN physicist Douglas R. O. Morrison said that "essentially all" attempts in Western Europe had failed. Even those reporting success had difficulty reproducing Fleischmann and Pons' results. On 10 April 1989, a group at Texas A&M University published results of excess heat and later that day a group at the Georgia Institute of Technology announced neutron production—the strongest replication announced up to that point due to the detection of neutrons and the reputation of the lab. On 12 April Pons was acclaimed at an ACS meeting. But Georgia Tech retracted their announcement on 13 April, explaining that their neutron detectors gave false positives when exposed to heat. Another attempt at independent replication, headed by Robert Huggins at Stanford University, which also reported early success with a light water control, became the only scientific support for cold fusion in 26 April US Congress hearings. But when he finally presented his results he reported an excess heat of only one degree Celsius, a result that could be explained by chemical differences between heavy and light water in the presence of lithium. He had not tried to measure any radiation and his research was derided by scientists who saw it later. For the next six weeks, competing claims, counterclaims, and suggested explanations kept what was referred to as "cold fusion" or "fusion confusion" in the news. In April 1989, Fleischmann and Pons published a "preliminary note" in the "Journal of Electroanalytical Chemistry". This paper notably showed a gamma peak without its corresponding Compton edge, which indicated they had made a mistake in claiming evidence of fusion byproducts. Fleischmann and Pons replied to this critique, but the only thing left clear was that no gamma ray had been registered and that Fleischmann refused to recognize any mistakes in the data. A much longer paper published a year later went into details of calorimetry but did not include any nuclear measurements. Nevertheless, Fleischmann and Pons and a number of other researchers who found positive results remained convinced of their findings. The University of Utah asked Congress to provide $25 million to pursue the research, and Pons was scheduled to meet with representatives of President Bush in early May. On 30 April 1989 cold fusion was declared dead by the "New York Times". The "Times" called it a circus the same day, and the "Boston Herald" attacked cold fusion the following day. On 1 May 1989 the American Physical Society held a session on cold fusion in Baltimore, including many reports of experiments that failed to produce evidence of cold fusion. At the end of the session, eight of the nine leading speakers stated that they considered the initial Fleischmann and Pons claim dead, with the ninth, Johann Rafelski, abstaining. Steven E. Koonin of Caltech called the Utah report a result of ""the incompetence and delusion of Pons and Fleischmann,"" which was met with a standing ovation. Douglas R. O. Morrison, a physicist representing CERN, was the first to call the episode an example of pathological science. On 4 May, due to all this new criticism, the meetings with various representatives from Washington were cancelled. From 8 May only the A&M tritium results kept cold fusion afloat. In July and November 1989, "Nature" published papers critical of cold fusion claims. Negative results were also published in several other scientific journals including "Science", "Physical Review Letters", and "Physical Review C" (nuclear physics). In August 1989, in spite of this trend, the state of Utah invested $4.5 million to create the National Cold Fusion Institute. The United States Department of Energy organized a special panel to review cold fusion theory and research. The panel issued its report in November 1989, concluding that results as of that date did not present convincing evidence that useful sources of energy would result from the phenomena attributed to cold fusion. The panel noted the large number of failures to replicate excess heat and the greater inconsistency of reports of nuclear reaction byproducts expected by established conjecture. Nuclear fusion of the type postulated would be inconsistent with current understanding and, if verified, would require established conjecture, perhaps even theory itself, to be extended in an unexpected way. The panel was against special funding for cold fusion research, but supported modest funding of "focused experiments within the general funding system". Cold fusion supporters continued to argue that the evidence for excess heat was strong, and in September 1990 the National Cold Fusion Institute listed 92 groups of researchers from 10 different countries that had reported corroborating evidence of excess heat, but they refused to provide any evidence of their own arguing that it could endanger their patents. However, no further DOE nor NSF funding resulted from the panel's recommendation. By this point, however, academic consensus had moved decidedly toward labeling cold fusion as a kind of "pathological science". In March 1990 Michael H. Salamon, a physicist from the University of Utah, and nine co-authors reported negative results. University faculty were then "stunned" when a lawyer representing Pons and Fleischmann demanded the Salamon paper be retracted under threat of a lawsuit. The lawyer later apologized; Fleischmann defended the threat as a legitimate reaction to alleged bias displayed by cold-fusion critics. In early May 1990 one of the two A&M researchers, Kevin Wolf, acknowledged the possibility of spiking, but said that the most likely explanation was tritium contamination in the palladium electrodes or simply contamination due to sloppy work. In June 1990 an article in "Science" by science writer Gary Taubes destroyed the public credibility of the A&M tritium results when it accused its group leader John Bockris and one of his graduate students of spiking the cells with tritium. In October 1990 Wolf finally said that the results were explained by tritium contamination in the rods. An A&M cold fusion review panel found that the tritium evidence was not convincing and that, while they couldn't rule out spiking, contamination and measurements problems were more likely explanations, and Bockris never got support from his faculty to resume his research. On 30 June 1991 the National Cold Fusion Institute closed after it ran out of funds; it found no excess heat, and its reports of tritium production were met with indifference. On 1 January 1991 Pons left the University of Utah and went to Europe. In 1992, Pons and Fleischman resumed research with Toyota Motor Corporation's IMRA lab in France. Fleischmann left for England in 1995, and the contract with Pons was not renewed in 1998 after spending $40 million with no tangible results. The IMRA laboratory stopped cold fusion research in 1998 after spending £12 million. Pons has made no public declarations since, and only Fleischmann continued giving talks and publishing papers. Mostly in the 1990s, several books were published that were critical of cold fusion research methods and the conduct of cold fusion researchers. Over the years, several books have appeared that defended them. Around 1998, the University of Utah had already dropped its research after spending over $1 million, and in the summer of 1997, Japan cut off research and closed its own lab after spending $20 million. A 1991 review by a cold fusion proponent had calculated "about 600 scientists" were still conducting research. After 1991, cold fusion research only continued in relative obscurity, conducted by groups that had increasing difficulty securing public funding and keeping programs open. These small but committed groups of cold fusion researchers have continued to conduct experiments using Fleischmann and Pons electrolysis setups in spite of the rejection by the mainstream community. "The Boston Globe" estimated in 2004 that there were only 100 to 200 researchers working in the field, most suffering damage to their reputation and career. Since the main controversy over Pons and Fleischmann had ended, cold fusion research has been funded by private and small governmental scientific investment funds in the United States, Italy, Japan, and India. For example, it was reported in "Nature", in May, 2019, that Google had spent approximately $10 million on cold fusion research. A group of scientists at well-known research labs (e.g, MIT, Lawrence Berkeley National Lab, and others) worked for several years to establish experimental protocols and measurement techniques in an effort to re-evaluate cold fusion to a high standard of scientific rigor. Their reported conclusion: no cold fusion. Cold fusion research continues today in a few specific venues, but the wider scientific community has generally marginalized the research being done and researchers have had difficulty publishing in mainstream journals. The remaining researchers often term their field Low Energy Nuclear Reactions (LENR), Chemically Assisted Nuclear Reactions (CANR), Lattice Assisted Nuclear Reactions (LANR), Condensed Matter Nuclear Science (CMNS) or Lattice Enabled Nuclear Reactions; one of the reasons being to avoid the negative connotations associated with "cold fusion". The new names avoid making bold implications, like implying that fusion is actually occurring. The researchers who continue acknowledge that the flaws in the original announcement are the main cause of the subject's marginalization, and they complain of a chronic lack of funding and no possibilities of getting their work published in the highest impact journals. University researchers are often unwilling to investigate cold fusion because they would be ridiculed by their colleagues and their professional careers would be at risk. In 1994, David Goodstein, a professor of physics at Caltech, advocated for increased attention from mainstream researchers and described cold fusion as: United States Navy researchers at the Space and Naval Warfare Systems Center (SPAWAR) in San Diego have been studying cold fusion since 1989. In 2002 they released a two-volume report, "Thermal and nuclear aspects of the Pd/D2O system," with a plea for funding. This and other published papers prompted a 2004 Department of Energy (DOE) review. In August 2003, the U.S. Secretary of Energy, Spencer Abraham, ordered the DOE to organize a second review of the field. This was thanks to an April 2003 letter sent by MIT's Peter L. Hagelstein, and the publication of many new papers, including the Italian ENEA and other researchers in the 2003 International Cold Fusion Conference, and a two-volume book by U.S. SPAWAR in 2002. Cold fusion researchers were asked to present a review document of all the evidence since the 1989 review. The report was released in 2004. The reviewers were "split approximately evenly" on whether the experiments had produced energy in the form of heat, but "most reviewers, even those who accepted the evidence for excess power production, 'stated that the effects are not repeatable, the magnitude of the effect has not increased in over a decade of work, and that many of the reported experiments were not well documented.'" In summary, reviewers found that cold fusion evidence was still not convincing 15 years later, and they didn't recommend a federal research program. They only recommended that agencies consider funding individual well-thought studies in specific areas where research "could be helpful in resolving some of the controversies in the field". They summarized its conclusions thus: Cold fusion researchers placed a "rosier spin" on the report, noting that they were finally being treated like normal scientists, and that the report had increased interest in the field and caused "a huge upswing in interest in funding cold fusion research." However, in a 2009 BBC article on an American Chemical Society's meeting on cold fusion, particle physicist Frank Close was quoted stating that the problems that plagued the original cold fusion announcement were still happening: results from studies are still not being independently verified and inexplicable phenomena encountered are being labelled as "cold fusion" even if they are not, in order to attract the attention of journalists. In February 2012, millionaire Sidney Kimmel, convinced that cold fusion was worth investing in by a 19 April 2009 interview with physicist Robert Duncan on the US news show "60 Minutes", made a grant of $5.5 million to the University of Missouri to establish the Sidney Kimmel Institute for Nuclear Renaissance (SKINR). The grant was intended to support research into the interactions of hydrogen with palladium, nickel or platinum under extreme conditions. In March 2013 Graham K. Hubler, a nuclear physicist who worked for the Naval Research Laboratory for 40 years, was named director. One of the SKINR projects is to replicate a 1991 experiment in which a professor associated with the project, Mark Prelas, says bursts of millions of neutrons a second were recorded, which was stopped because "his research account had been frozen". He claims that the new experiment has already seen "neutron emissions at similar levels to the 1991 observation". In May 2016, the United States House Committee on Armed Services, in its report on the 2017 National Defense Authorization Act, directed the Secretary of Defense to "provide a briefing on the military utility of recent U.S. industrial base LENR advancements to the House Committee on Armed Services by September 22, 2016." Since the Fleischmann and Pons announcement, the Italian national agency for new technologies, energy and sustainable economic development (ENEA) has funded Franco Scaramuzzi's research into whether excess heat can be measured from metals loaded with deuterium gas. Such research is distributed across ENEA departments, CNR laboratories, INFN, universities and industrial laboratories in Italy, where the group continues to try to achieve reliable reproducibility (i.e. getting the phenomenon to happen in every cell, and inside a certain frame of time). In 2006–2007, the ENEA started a research program which claimed to have found excess power of up to 500 percent, and in 2009, ENEA hosted the 15th cold fusion conference. Between 1992 and 1997, Japan's Ministry of International Trade and Industry sponsored a "New Hydrogen Energy (NHE)" program of US$20 million to research cold fusion. Announcing the end of the program in 1997, the director and one-time proponent of cold fusion research Hideo Ikegami stated "We couldn't achieve what was first claimed in terms of cold fusion. (...) We can't find any reason to propose more money for the coming year or for the future." In 1999 the Japan C-F Research Society was established to promote the independent research into cold fusion that continued in Japan. The society holds annual meetings. Perhaps the most famous Japanese cold fusion researcher is Yoshiaki Arata, from Osaka University, who claimed in a demonstration to produce excess heat when deuterium gas was introduced into a cell containing a mixture of palladium and zirconium oxide, a claim supported by fellow Japanese researcher Akira Kitamura of Kobe University and McKubre at SRI. In the 1990s India stopped its research in cold fusion at the Bhabha Atomic Research Centre because of the lack of consensus among mainstream scientists and the US denunciation of the research. Yet, in 2008, the National Institute of Advanced Studies recommended that the Indian government revive this research. Projects were commenced at Chennai's Indian Institute of Technology, the Bhabha Atomic Research Centre and the Indira Gandhi Centre for Atomic Research. However, there is still skepticism among scientists and, for all practical purposes, research has stalled since the 1990s. A special section in the Indian multidisciplinary journal "Current Science" published 33 cold fusion papers in 2015 by major cold fusion researchers including several Indian researchers. A cold fusion experiment usually includes: Electrolysis cells can be either open cell or closed cell. In open cell systems, the electrolysis products, which are gaseous, are allowed to leave the cell. In closed cell experiments, the products are captured, for example by catalytically recombining the products in a separate part of the experimental system. These experiments generally strive for a steady state condition, with the electrolyte being replaced periodically. There are also "heat-after-death" experiments, where the evolution of heat is monitored after the electric current is turned off. The most basic setup of a cold fusion cell consists of two electrodes submerged in a solution containing palladium and heavy water. The electrodes are then connected to a power source to transmit electricity from one electrode to the other through the solution. Even when anomalous heat is reported, it can take weeks for it to begin to appear—this is known as the "loading time," the time required to saturate the palladium electrode with hydrogen (see "Loading ratio" section). The Fleischmann and Pons early findings regarding helium, neutron radiation and tritium were never replicated satisfactorily, and its levels were too low for the claimed heat production and inconsistent with each other. Neutron radiation has been reported in cold fusion experiments at very low levels using different kinds of detectors, but levels were too low, close to background, and found too infrequently to provide useful information about possible nuclear processes. An excess heat observation is based on an energy balance. Various sources of energy input and output are continuously measured. Under normal conditions, the energy input can be matched to the energy output to within experimental error. In experiments such as those run by Fleischmann and Pons, an electrolysis cell operating steadily at one temperature transitions to operating at a higher temperature with no increase in applied current. If the higher temperatures were real, and not an experimental artifact, the energy balance would show an unaccounted term. In the Fleischmann and Pons experiments, the rate of inferred excess heat generation was in the range of 10–20% of total input, though this could not be reliably replicated by most researchers. Researcher Nathan Lewis discovered that the excess heat in Fleischmann and Pons's original paper was not measured, but estimated from measurements that didn't have any excess heat. Unable to produce excess heat or neutrons, and with positive experiments being plagued by errors and giving disparate results, most researchers declared that heat production was not a real effect and ceased working on the experiments. In 1993, after their original report, Fleischmann reported "heat-after-death" experiments—where excess heat was measured after the electric current supplied to the electrolytic cell was turned off. This type of report has also become part of subsequent cold fusion claims. Known instances of nuclear reactions, aside from producing energy, also produce nucleons and particles on readily observable ballistic trajectories. In support of their claim that nuclear reactions took place in their electrolytic cells, Fleischmann and Pons reported a neutron flux of 4,000 neutrons per second, as well as detection of tritium. The classical branching ratio for previously known fusion reactions that produce tritium would predict, with 1 watt of power, the production of 1012 neutrons per second, levels that would have been fatal to the researchers. In 2009, Mosier-Boss et al. reported what they called the first scientific report of highly energetic neutrons, using CR-39 plastic radiation detectors, but the claims cannot be validated without a quantitative analysis of neutrons. Several medium and heavy elements like calcium, titanium, chromium, manganese, iron, cobalt, copper and zinc have been reported as detected by several researchers, like Tadahiko Mizuno or George Miley. The report presented to the United States Department of Energy (DOE) in 2004 indicated that deuterium-loaded foils could be used to detect fusion reaction products and, although the reviewers found the evidence presented to them as inconclusive, they indicated that those experiments did not use state-of-the-art techniques. In response to doubts about the lack of nuclear products, cold fusion researchers have tried to capture and measure nuclear products correlated with excess heat. Considerable attention has been given to measuring 4He production. However, the reported levels are very near to background, so contamination by trace amounts of helium normally present in the air cannot be ruled out. In the report presented to the DOE in 2004, the reviewers' opinion was divided on the evidence for 4He; with the most negative reviews concluding that although the amounts detected were above background levels, they were very close to them and therefore could be caused by contamination from air. One of the main criticisms of cold fusion was that deuteron-deuteron fusion into helium was expected to result in the production of gamma rays—which were not observed and were not observed in subsequent cold fusion experiments. Cold fusion researchers have since claimed to find X-rays, helium, neutrons and nuclear transmutations. Some researchers also claim to have found them using only light water and nickel cathodes. The 2004 DOE panel expressed concerns about the poor quality of the theoretical framework cold fusion proponents presented to account for the lack of gamma rays. Researchers in the field do not agree on a theory for cold fusion. One proposal considers that hydrogen and its isotopes can be absorbed in certain solids, including palladium hydride, at high densities. This creates a high partial pressure, reducing the average separation of hydrogen isotopes. However, the reduction in separation is not enough by a factor of ten to create the fusion rates claimed in the original experiment. It was also proposed that a higher density of hydrogen inside the palladium and a lower potential barrier could raise the possibility of fusion at lower temperatures than expected from a simple application of Coulomb's law. Electron screening of the positive hydrogen nuclei by the negative electrons in the palladium lattice was suggested to the 2004 DOE commission, but the panel found the theoretical explanations not convincing and inconsistent with current physics theories. Criticism of cold fusion claims generally take one of two forms: either pointing out the theoretical implausibility that fusion reactions have occurred in electrolysis setups or criticizing the excess heat measurements as being spurious, erroneous, or due to poor methodology or controls. There are a couple of reasons why known fusion reactions are an unlikely explanation for the excess heat and associated cold fusion claims. Because nuclei are all positively charged, they strongly repel one another. Normally, in the absence of a catalyst such as a muon, very high kinetic energies are required to overcome this charged repulsion. Extrapolating from known fusion rates, the rate for uncatalyzed fusion at room-temperature energy would be 50 orders of magnitude lower than needed to account for the reported excess heat. In muon-catalyzed fusion there are more fusions because the presence of the muon causes deuterium nuclei to be 207 times closer than in ordinary deuterium gas. But deuterium nuclei inside a palladium lattice are further apart than in deuterium gas, and there should be fewer fusion reactions, not more. Paneth and Peters in the 1920s already knew that palladium can absorb up to 900 times its own volume of hydrogen gas, storing it at several thousands of times the atmospheric pressure. This led them to believe that they could increase the nuclear fusion rate by simply loading palladium rods with hydrogen gas. Tandberg then tried the same experiment but used electrolysis to make palladium absorb more deuterium and force the deuterium further together inside the rods, thus anticipating the main elements of Fleischmann and Pons' experiment. They all hoped that pairs of hydrogen nuclei would fuse together to form helium, which at the time was needed in Germany to fill zeppelins, but no evidence of helium or of increased fusion rate was ever found. This was also the belief of geologist Palmer, who convinced Steven Jones that the helium-3 occurring naturally in Earth perhaps came from fusion involving hydrogen isotopes inside catalysts like nickel and palladium. This led their team in 1986 to independently make the same experimental setup as Fleischmann and Pons (a palladium cathode submerged in heavy water, absorbing deuterium via electrolysis). Fleischmann and Pons had much the same belief, but they calculated the pressure to be of 1027 atmospheres, when cold fusion experiments only achieve a loading ratio of one to one, which only has between 10,000 and 20,000 atmospheres. John R. Huizenga says they had misinterpreted the Nernst equation, leading them to believe that there was enough pressure to bring deuterons so close to each other that there would be spontaneous fusions. Conventional deuteron fusion is a two-step process, in which an unstable high-energy intermediary is formed: Experiments have observed only three decay pathways for this excited-state nucleus, with the branching ratio showing the probability that any given intermediate follows a particular pathway. The products formed via these decay pathways are: Only about one in one million of the intermediaries decay along the third pathway, making its products comparatively rare when compared to the other paths. This result is consistent with the predictions of the Bohr model. If one watt (1 W = 1 J/s ; 1 J = 6.242 × 1018 eV = 6.242 × 1012 MeV since 1 eV = 1.602 × 10−19 joule) of nuclear power were produced from ~2.2575 × 1011 deuteron fusion individual reactions each second consistent with known branching ratios, the resulting neutron and tritium (3H) production would be easily measured. Some researchers reported detecting 4He but without the expected neutron or tritium production; such a result would require branching ratios strongly favouring the third pathway, with the actual rates of the first two pathways lower by at least five orders of magnitude than observations from other experiments, directly contradicting both theoretically predicted and observed branching probabilities. Those reports of 4He production did not include detection of gamma rays, which would require the third pathway to have been changed somehow so that gamma rays are no longer emitted. The known rate of the decay process together with the inter-atomic spacing in a metallic crystal makes heat transfer of the 24 MeV excess energy into the host metal lattice prior to the intermediary's decay inexplicable in terms of conventional understandings of momentum and energy transfer, and even then there would be measurable levels of radiation. Also, experiments indicate that the ratios of deuterium fusion remain constant at different energies. In general, pressure and chemical environment only cause small changes to fusion ratios. An early explanation invoked the Oppenheimer–Phillips process at low energies, but its magnitude was too small to explain the altered ratios. Cold fusion setups utilize an input power source (to ostensibly provide activation energy), a platinum group electrode, a deuterium or hydrogen source, a calorimeter, and, at times, detectors to look for byproducts such as helium or neutrons. Critics have variously taken issue with each of these aspects and have asserted that there has not yet been a consistent reproduction of claimed cold fusion results in either energy output or byproducts. Some cold fusion researchers who claim that they can consistently measure an excess heat effect have argued that the apparent lack of reproducibility might be attributable to a lack of quality control in the electrode metal or the amount of hydrogen or deuterium loaded in the system. Critics have further taken issue with what they describe as mistakes or errors of interpretation that cold fusion researchers have made in calorimetry analyses and energy budgets. In 1989, after Fleischmann and Pons had made their claims, many research groups tried to reproduce the Fleischmann-Pons experiment, without success. A few other research groups, however, reported successful reproductions of cold fusion during this time. In July 1989, an Indian group from the Bhabha Atomic Research Centre (P. K. Iyengar and M. Srinivasan) and in October 1989, John Bockris' group from Texas A&M University reported on the creation of tritium. In December 1990, professor Richard Oriani of the University of Minnesota reported excess heat. Groups that did report successes found that some of their cells were producing the effect, while other cells that were built exactly the same and used the same materials were not producing the effect. Researchers that continued to work on the topic have claimed that over the years many successful replications have been made, but still have problems getting reliable replications. Reproducibility is one of the main principles of the scientific method, and its lack led most physicists to believe that the few positive reports could be attributed to experimental error. The DOE 2004 report said among its conclusions and recommendations: Cold fusion researchers (McKubre since 1994, ENEA in 2011) have speculated that a cell that is loaded with a deuterium/palladium ratio lower than 100% (or 1:1) will not produce excess heat. Since most of the negative replications from 1989 to 1990 did not report their ratios, this has been proposed as an explanation for failed replications. This loading ratio is hard to obtain, and some batches of palladium never reach it because the pressure causes cracks in the palladium, allowing the deuterium to escape. Fleischmann and Pons never disclosed the deuterium/palladium ratio achieved in their cells, there are no longer any batches of the palladium used by Fleischmann and Pons (because the supplier now uses a different manufacturing process), and researchers still have problems finding batches of palladium that achieve heat production reliably. Some research groups initially reported that they had replicated the Fleischmann and Pons results but later retracted their reports and offered an alternative explanation for their original positive results. A group at Georgia Tech found problems with their neutron detector, and Texas A&M discovered bad wiring in their thermometers. These retractions, combined with negative results from some famous laboratories, led most scientists to conclude, as early as 1989, that no positive result should be attributed to cold fusion. The calculation of excess heat in electrochemical cells involves certain assumptions. Errors in these assumptions have been offered as non-nuclear explanations for excess heat. One assumption made by Fleischmann and Pons is that the efficiency of electrolysis is nearly 100%, meaning nearly all the electricity applied to the cell resulted in electrolysis of water, with negligible resistive heating and substantially all the electrolysis product leaving the cell unchanged. This assumption gives the amount of energy expended converting liquid D2O into gaseous D2 and O2. The efficiency of electrolysis is less than one if hydrogen and oxygen recombine to a significant extent within the calorimeter. Several researchers have described potential mechanisms by which this process could occur and thereby account for excess heat in electrolysis experiments. Another assumption is that heat loss from the calorimeter maintains the same relationship with measured temperature as found when calibrating the calorimeter. This assumption ceases to be accurate if the temperature distribution within the cell becomes significantly altered from the condition under which calibration measurements were made. This can happen, for example, if fluid circulation within the cell becomes significantly altered. Recombination of hydrogen and oxygen within the calorimeter would also alter the heat distribution and invalidate the calibration. The ISI identified cold fusion as the scientific topic with the largest number of published papers in 1989, of all scientific disciplines. The Nobel Laureate Julian Schwinger declared himself a supporter of cold fusion in the fall of 1989, after much of the response to the initial reports had turned negative. He tried to publish his theoretical paper "Cold Fusion: A Hypothesis" in "Physical Review Letters", but the peer reviewers rejected it so harshly that he felt deeply insulted, and he resigned from the American Physical Society (publisher of "PRL") in protest. The number of papers sharply declined after 1990 because of two simultaneous phenomena: first, scientists abandoned the field; second, journal editors declined to review new papers. Consequently, cold fusion fell off the ISI charts. Researchers who got negative results turned their backs on the field; those who continued to publish were simply ignored. A 1993 paper in "Physics Letters A" was the last paper published by Fleischmann, and "one of the last reports [by Fleischmann] to be formally challenged on technical grounds by a cold fusion skeptic." The "Journal of Fusion Technology" (FT) established a permanent feature in 1990 for cold fusion papers, publishing over a dozen papers per year and giving a mainstream outlet for cold fusion researchers. When editor-in-chief George H. Miley retired in 2001, the journal stopped accepting new cold fusion papers. This has been cited as an example of the importance of sympathetic influential individuals to the publication of cold fusion papers in certain journals. The decline of publications in cold fusion has been described as a "failed information epidemic". The sudden surge of supporters until roughly 50% of scientists support the theory, followed by a decline until there is only a very small number of supporters, has been described as a characteristic of pathological science. The lack of a shared set of unifying concepts and techniques has prevented the creation of a dense network of collaboration in the field; researchers perform efforts in their own and in disparate directions, making the transition to "normal" science more difficult. Cold fusion reports continued to be published in a small cluster of specialized journals like "Journal of Electroanalytical Chemistry" and "Il Nuovo Cimento". Some papers also appeared in "Journal of Physical Chemistry", "Physics Letters A", "International Journal of Hydrogen Energy", and a number of Japanese and Russian journals of physics, chemistry, and engineering. Since 2005, "Naturwissenschaften" has published cold fusion papers; in 2009, the journal named a cold fusion researcher to its editorial board. In 2015 the Indian multidisciplinary journal "Current Science" published a special section devoted entirely to cold fusion related papers. In the 1990s, the groups that continued to research cold fusion and their supporters established (non-peer-reviewed) periodicals such as "Fusion Facts", "Cold Fusion Magazine", "Infinite Energy Magazine" and "New Energy Times" to cover developments in cold fusion and other fringe claims in energy production that were ignored in other venues. The internet has also become a major means of communication and self-publication for CF researchers. Cold fusion researchers were for many years unable to get papers accepted at scientific meetings, prompting the creation of their own conferences. The first International Conference on Cold Fusion (ICCF) was held in 1990, and has met every 12 to 18 months since. Attendees at some of the early conferences were described as offering no criticism to papers and presentations for fear of giving ammunition to external critics, thus allowing the proliferation of crackpots and hampering the conduct of serious science. Critics and skeptics stopped attending these conferences, with the notable exception of Douglas Morrison, who died in 2001. With the founding in 2004 of the International Society for Condensed Matter Nuclear Science (ISCMNS), the conference was renamed the International Conference on Condensed Matter Nuclear Science – for reasons that are detailed in the subsequent research section above – but reverted to the old name in 2008. Cold fusion research is often referenced by proponents as "low-energy nuclear reactions", or LENR, but according to sociologist Bart Simon the "cold fusion" label continues to serve a social function in creating a collective identity for the field. Since 2006, the American Physical Society (APS) has included cold fusion sessions at their semiannual meetings, clarifying that this does not imply a softening of skepticism. Since 2007, the American Chemical Society (ACS) meetings also include "invited symposium(s)" on cold fusion. An ACS program chair said that without a proper forum the matter would never be discussed and, "with the world facing an energy crisis, it is worth exploring all possibilities." On 22–25 March 2009, the American Chemical Society meeting included a four-day symposium in conjunction with the 20th anniversary of the announcement of cold fusion. Researchers working at the U.S. Navy's Space and Naval Warfare Systems Center (SPAWAR) reported detection of energetic neutrons using a heavy water electrolysis setup and a CR-39 detector, a result previously published in "Naturwissenschaften". The authors claim that these neutrons are indicative of nuclear reactions; without quantitative analysis of the number, energy, and timing of the neutrons and exclusion of other potential sources, this interpretation is unlikely to find acceptance by the wider scientific community. Although details have not surfaced, it appears that the University of Utah forced the 23 March 1989 Fleischmann and Pons announcement to establish priority over the discovery and its patents before the joint publication with Jones. The Massachusetts Institute of Technology (MIT) announced on 12 April 1989 that it had applied for its own patents based on theoretical work of one of its researchers, Peter L. Hagelstein, who had been sending papers to journals from 5 to 12 April. On 2 December 1993 the University of Utah licensed all its cold fusion patents to ENECO, a new company created to profit from cold fusion discoveries, and in March 1998 it said that it would no longer defend its patents. The U.S. Patent and Trademark Office (USPTO) now rejects patents claiming cold fusion. Esther Kepplinger, the deputy commissioner of patents in 2004, said that this was done using the same argument as with perpetual motion machines: that they do not work. Patent applications are required to show that the invention is "useful", and this utility is dependent on the invention's ability to function. In general USPTO rejections on the sole grounds of the invention's being "inoperative" are rare, since such rejections need to demonstrate "proof of total incapacity", and cases where those rejections are upheld in a Federal Court are even rarer: nevertheless, in 2000, a rejection of a cold fusion patent was appealed in a Federal Court and it was upheld, in part on the grounds that the inventor was unable to establish the utility of the invention. A U.S. patent might still be granted when given a different name to disassociate it from cold fusion, though this strategy has had little success in the US: the same claims that need to be patented can identify it with cold fusion, and most of these patents cannot avoid mentioning Fleischmann and Pons' research due to legal constraints, thus alerting the patent reviewer that it is a cold-fusion-related patent. David Voss said in 1999 that some patents that closely resemble cold fusion processes, and that use materials used in cold fusion, have been granted by the USPTO. The inventor of three such patents had his applications initially rejected when they were reviewed by experts in nuclear science; but then he rewrote the patents to focus more on the electrochemical parts so they would be reviewed instead by experts in electrochemistry, who approved them. When asked about the resemblance to cold fusion, the patent holder said that it used nuclear processes involving "new nuclear physics" unrelated to cold fusion. Melvin Miles was granted in 2004 a patent for a cold fusion device, and in 2007 he described his efforts to remove all instances of "cold fusion" from the patent description to avoid having it rejected outright. At least one patent related to cold fusion has been granted by the European Patent Office. A patent only legally prevents others from using or benefiting from one's invention. However, the general public perceives a patent as a stamp of approval, and a holder of three cold fusion patents said the patents were very valuable and had helped in getting investments. A 1990 Michael Winner film "Bullseye!", starring Michael Caine and Roger Moore, referenced the Fleischmann and Pons experiment. The film – a comedy – concerned conmen trying to steal scientists' purported findings. However, the film had a poor reception, described as "appallingly unfunny". In "Undead Science", sociologist Bart Simon gives some examples of cold fusion in popular culture, saying that some scientists use cold fusion as a synonym for outrageous claims made with no supporting proof, and courses of ethics in science give it as an example of pathological science. It has appeared as a joke in "Murphy Brown" and "The Simpsons". It was adopted as a software product name Adobe ColdFusion and a brand of protein bars (Cold Fusion Foods). It has also appeared in advertising as a synonym for impossible science, for example a 1995 advertisement for Pepsi Max. The plot of "The Saint", a 1997 action-adventure film, parallels the story of Fleischmann and Pons, although with a different ending. The film might have affected the public perception of cold fusion, pushing it further into the science fiction realm. "Final Exam", the 16th episode of season 4 of "The Outer Limits", depicts a student named Todtman who has invented a cold fusion weapon, and attempts to use it as a tool for revenge on people who have wronged him over the years. Despite the secret being lost with his death at the end of the episode, it is implied that another student elsewhere is on a similar track, and may well repeat Todtman's efforts. In the "DC's Legends of Tomorrow" episode "No Country for Old Dads," Ray Palmer theorizes that cold fusion could repair the shattered Fire Totem, if it wasn't only theoretical. Damien Dahrk reveals that he assassinated a scientist in 1962 East Berlin that developed a formula for cold fusion. Ray and Dahrk's daughter Nora time travel from 2018 to 1962 in an attempt to rescue the scientist from the younger version of Dahrk and/or retrieve the formula.
https://en.wikipedia.org/wiki?curid=7463
Coal tar Coal tar is a thick dark liquid which is a by-product of the production of coke and coal gas from coal. It has both medical and industrial uses. It may be applied to the affected area to treat psoriasis and seborrheic dermatitis (dandruff). It may be used in combination with ultraviolet light therapy. Industrially it is a railway tie preservative and used in the surfacing of roads. Side effects include skin irritation, sun sensitivity, allergic reactions, and skin discoloration. It is unclear if use during pregnancy is safe for the baby and use during breastfeeding is not typically recommended. The exact mechanism of action is unknown. It is a complex mixture of phenols, polycyclic aromatic hydrocarbons (PAHs), and heterocyclic compounds. It demonstrates antifungal, anti-inflammatory, anti-itch, and antiparasitic properties. Coal tar was discovered around 1665 and used for medical purposes as early as the 1800s. It is on the World Health Organization's List of Essential Medicines, the safest and most effective medicines needed in a health system. Coal tar is available as a generic medication and over the counter. In the United Kingdom 125 ml of 5% shampoo costs the NHS about £1.89. In the United States a month's treatment costs less than $25 USD. Coal tar was one of the key starting materials for the early pharmaceutical industry. Coal tar is used in medicated shampoo, soap and ointment. It demonstrates antifungal, anti-inflammatory, anti-itch, and antiparasitic properties. It may be applied topically as a treatment for dandruff and psoriasis, and to kill and repel head lice. It may be used in combination with ultraviolet light therapy. Pine tar has historically also been used for this purpose. Though it is frequently cited online as having been banned as a medical product by the FDA due to a "lack of evidence having been submitted for proof of effectiveness", pine tar is included in the Code of Federal Regulations, subchapter D: Drugs for Human Use, as an OTC treatment for "Dandruff/seborrheic dermatitis/psoriasis". Coal tar may be used in two forms: crude coal tar () or a coal tar solution () also known as liquor carbonis detergens (LCD). Named brands include Denorex, Balnetar, Psoriasin, Tegrin, T/Gel, and Neutar. When used in the extemporaneous preparation of topical medications, it is supplied in the form of coal tar topical solution USP, which consists of a 20% w/v solution of coal tar in alcohol, with an additional 5% w/v of polysorbate 80 USP; this must then be diluted in an ointment base such as petrolatum. Coal tar was a component of the first sealed roads. In its original development by Edgar Purnell Hooley, tarmac was tar covered with granite chips. Later the filler used was industrial slag. Today, petroleum derived binders and sealers are more commonly used. These sealers are used to extend the life and reduce maintenance cost associated with asphalt pavements, primarily in asphalt road paving, car parks and walkways. Coal tar is incorporated into some parking-lot sealcoat products used to protect the structural integrity of the underlying pavement. Sealcoat products that are coal-tar based typically contain 20 to 35 percent coal-tar pitch. Research shows it is used in United States states from Alaska to Florida, but several areas have banned its use in sealcoat products, including the District of Columbia; the city of Austin, Texas; Dane County, Wisconsin; the state of Washington; and several municipalities in Minnesota and others. Being flammable, coal tar is sometimes used for heating or to fire boilers. Like most heavy oils, it must be heated before it will flow easily. A large part of the binder used in the graphite industry for making "green blocks" is coke oven volatiles (COV), a considerable portion of which is coal tar. During the baking process of the green blocks as a part of commercial graphite production, most of the coal tar binders are vaporised and are generally burned in an incinerator to prevent release into the atmosphere, as COV and coal tar can be injurious to health. Coal tar is also used to manufacture paints, synthetic dyes (notably tartrazine/Yellow #5), and photographic materials. In the coal gas era, there were many companies in Britain whose business was to distill coal tar to separate the higher-value fractions, such as naphtha, creosote and pitch. A great many industrial chemicals were first isolated from coal tar during this time. These companies included: Side effects of coal tar products include skin irritation, sun sensitivity, allergic reactions, and skin discoloration. It is unclear if use during pregnancy is safe for the baby and use during breastfeeding is not typically recommended. According to the National Psoriasis Foundation, coal tar is a valuable, safe and inexpensive treatment option for millions of people with psoriasis and other scalp or skin conditions. According to the FDA, coal tar concentrations between 0.5% and 5% are considered safe and effective for psoriasis. Evidence is inconclusive whether the coal tar in the concentrations seen in non-prescription treatments causes cancer, because there is insufficient data to make a judgment. While coal tar consistently causes cancer in animal studies, short-term treatments of humans have shown no significant increase in rates of cancer. It's possible that the skin can repair itself after short-term exposure to PAHs, but not after long-term exposure. Coal tar was one of the first chemical substances proven to cause cancer from occupational exposure, during research in 1775 on the cause of chimney sweeps' carcinoma. Modern studies have shown that working with coal tar pitch, such as during the paving of roads or when working on roofs, increases the risk of cancer. Coal tar contains many polycyclic aromatic hydrocarbons, and it is believed that their metabolites bind to DNA, damaging it. Long-term skin exposure to these compounds can produce "tar warts", which can progress to squamous cell carcinoma. The International Agency for Research on Cancer lists coal tars as Group 1 carcinogens, meaning they directly cause cancer. Both the U.S. Department of Health and Human Services and the state of California list coal tars as known human carcinogens. In response to public health concerns regarding the carcinogenicity of PAHs some municipalities, such as the city of Milwaukee, have banned the use of common coal tar-based road and driveway sealants citing concerns of elevated PAH content in groundwater. Coal tar causes increased sensitivity to sunlight, so skin treated with topical coal tar preparations should be protected from sunlight. The residue from the distillation of high-temperature coal tar, primarily a complex mixture of three or more membered condensed ring aromatic hydrocarbons, was listed on 28 October 2008 as a substance of very high concern by the European Chemicals Agency. The exact mechanism of action is unknown. Coal tar is a complex mixture of phenols, polycyclic aromatic hydrocarbons (PAHs), and heterocyclic compounds. It is a keratolytic agent, which reduces the growth rate of skin cells and softens the skin's keratin. Coal tar is produced through thermal destruction (pyrolysis) of coal. Its composition varies with the process and type of coal used – lignite, bituminous or anthracite. Coal tar contains approximately 10,000 chemicals, of which only about 50% have been identified. Components include polycyclic aromatic hydrocarbons (4-rings: chrysene, fluoranthene, pyrene, triphenylene, naphthacene, benzanthracene, 5-rings: picene, benzo[a]pyrene, benzo[e]pyrene, benzofluoranthenes, perylene, 6-rings: dibenzopyrenes, dibenzofluoranthenes, benzoperylenes, 7-rings: coronene), as well as methylated and polymethylated derivatives, mono- and polyhydroxylated derivatives, and heterocyclic compounds. Others include benzene, toluene, xylenes, cumenes, coumarone, indene, benzofuran, naphthalene and methyl-naphthalenes, acenaphthene, fluorene, phenol, cresols, pyridine, picolines, phenanthracene, carbazole, quinolines, fluoranthene. Many of these constituents are known carcinogens. Various phenolic coal tar derivatives have analgesic (pain-killer) properties. These included acetanilide, phenacetin, and paracetamol (acetaminophen). Paracetamol is the only coal-tar derived analgesic still in use today, but industrial phenol is now usually synthesized from crude oil rather than coal tar. Coal tar derivatives are contra-indicated for people with the inherited red cell blood disorder glucose-6-phosphate dehydrogenase deficiency (G6PD deficiency), as they can cause oxidative stress leading to red blood cell breakdown. Coal tar is on the World Health Organization's List of Essential Medicines, the most effective and safe medicines needed in a health system. Coal tar is generally available as a generic medication and over the counter. In the United Kingdom 125 ml of 5% shampoo costs the NHS about £1.89. In the United States a month of treatment costs less than . Exposure to coal tar pitch volatiles can occur in the workplace by breathing, skin contact, or eye contact. The Occupational Safety and Health Administration (OSHA) has set the permissible exposure limit) to 0.2 mg/m3 benzene-soluble fraction over an 8-hour workday. The National Institute for Occupational Safety and Health (NIOSH) has set a recommended exposure limit (REL) of 0.1 mg/m3 cyclohexane-extractable fraction over an 8-hour workday. At levels of 80 mg/m3, coal tar pitch volatiles are immediately dangerous to life and health. When used as a medication in the United States, coal tar preparations are considered over-the-counter drug pharmaceuticals and are subject to regulation by the Food and Drug Administration (FDA).
https://en.wikipedia.org/wiki?curid=7466
Catherine of Siena Catherine of Siena (25 March 1347 – 29 April 1380), a lay member of the Dominican Order, was a mystic, activist, and author who had a great influence on Italian literature and the Catholic Church. Canonized in 1461, she is also a Doctor of the Church. She was born and raised in Siena, and at an early age wanted to devote herself to God, against the will of her parents. She joined the Dominican tertiaries. She made herself known very quickly by being marked by mystical phenomena such as invisible stigmata and a mystical marriage. Her influence with Pope Gregory XI played a role in his decision to leave Avignon for Rome. She was then sent by him to negotiate peace with Florence. After Gregory XI's death and peace concluded, she returned to Siena. She dictated to secretaries her set of spiritual treatises "The Dialogue of Divine Providence". The Great Schism of the West led Catherine of Siena to go to Rome with the pope. She sent numerous letters to princes and cardinals to promote obedience to Pope Urban VI and defend what she calls the "vessel of the Church." She died on 29 April 1380, exhausted by her penances. Urban VI celebrated her funeral and burial in the Basilica of Santa Maria sopra Minerva in Rome. The devotion around Catherine of Siena developed rapidly after her death. She was canonized in 1461, declared patron saint of Rome in 1866, and of Italy (together with Francis of Assisi) in 1939. She was the first woman (along with Teresa of Ávila) to be declared a "doctor of the Church," on 4 October 1970 by Pope Paul VI. She was also proclaimed patron saint of Europe in 1999 by Pope John Paul II. Catherine of Siena is one of the outstanding figures of medieval Catholicism, by the strong influence she has had in the history of the papacy and her extensive authorship. She was behind the return of the Pope from Avignon to Rome, and then carried out many missions entrusted by the pope, something quite rare for a woman in the Middle Ages. Her "Dialogue", hundreds of letters, and dozens of prayers, also give her a prominent place in the history of Italian literature. Caterina di Giacomo di Benincasa was born on 25 March 1347 (shortly before the Black Death ravaged Europe) in Siena, Republic of Siena (today Italy), to Lapa Piagenti, the daughter of a local poet, and Giacomo di Benincasa, a cloth dyer who ran his enterprise with the help of his sons. The house where Catherine grew up still exists. Lapa was about forty years old when she gave premature birth to twin daughters Catherine and Giovanna. She had already borne 22 children, but half of them had died. Giovanna was handed over to a wet-nurse and died soon after. Catherine was nursed by her mother and developed into a healthy child. She was two years old when Lapa had her 25th child, another daughter named Giovanna. As a child Catherine was so merry that the family gave her the pet name of "Euphrosyne", which is Greek for "joy" and the name of an Euphrosyne of Alexandria. Catherine is said by her confessor and biographer Raymond of Capua O.P.'s "Life" to have had her first vision of Christ when she was five or six years old: she and a brother were on the way home from visiting a married sister when she is said to have experienced a vision of Christ seated in glory with the Apostles Peter, Paul, and John. Raymond continues that at age seven, Catherine vowed to give her whole life to God. When Catherine was sixteen, her older sister Bonaventura died in childbirth; already anguished by this, Catherine soon learned that her parents wanted her to marry Bonaventura's widower. She was absolutely opposed and started a strict fast. She had learned this from Bonaventura, whose husband had been far from considerate but his wife had changed his attitude by refusing to eat until he showed better manners. Besides fasting, Catherine further disappointed her mother by cutting off her long hair as a protest against being overly encouraged to improve her appearance to attract a husband. Catherine would later advise Raymond of Capua to do during times of trouble what she did now as a teenager: "Build a cell inside your mind, from which you can never flee." In this inner cell she made her father into a representation of Christ, her mother into the Blessed Virgin Mary, and her brothers into the apostles. Serving them humbly became an opportunity for spiritual growth. Catherine resisted the accepted course of marriage and motherhood on the one hand, or a nun's veil on the other. She chose to live an active and prayerful life outside a convent's walls following the model of the Dominicans. Eventually her father gave up and permitted her to live as she pleased. A vision of Saint Dominic gave strength to Catherine, but her wish to join his Order was no comfort to Lapa, who took her daughter with her to the baths in Bagno Vignoni to improve her health. Catherine fell seriously ill with a violent rash, fever and pain, which conveniently made her mother accept her wish to join the "Mantellate", the local association of Dominican tertiaries. Lapa went to the Sisters of the Order and persuaded them to take in her daughter. Within days, Catherine seemed entirely restored, rose from bed and donned the black and white habit of the Third Order of Saint Dominic. Catherine received the habit of a Dominican tertiary from the friars of the order after vigorous protests from the tertiaries themselves, who up to that point had been only widows. As a tertiary, she lived outside the convent, at home with her family like before. The Mantellate taught Catherine how to read, and she lived in almost total silence and solitude in the family home. Her custom of giving away clothing and food without asking anyone's permission cost her family significantly, but she requested nothing for herself. By staying in their midst, she could live out her rejection of them more strongly. She did not want their food, referring to the table laid for her in Heaven with her real family. According to Raymond of Capua, at the age of twenty-one (c. 1368), Catherine experienced what she described in her letters as a "Mystical Marriage" with Jesus, later a popular subject in art as the "Mystic marriage of Saint Catherine". Caroline Walker Bynum explains one surprising and controversial aspect of this marriage that occurs both in artistic representations of the event and in some early accounts of her life: "Underlining the extent to which the marriage was a fusion with Christ's physicality [...] Catherine received, not the ring of gold and jewels that her biographer reports in his bowdlerized version, but the ring of Christ's foreskin." Catherine herself mentions the foreskin-as-wedding ring motif in one of her letters (#221), equating the wedding ring of a virgin with a foreskin; she typically claimed that her own wedding ring to Christ was simply invisible. She wrote in a letter (to encourage a nun who seems to have been undergoing a prolonged period of spiritual trial and torment): "Bathe in the blood of Christ crucified. See that you don't look for or want anything but the crucified, as a true bride ransomed by the blood of Christ crucified-for that is my wish. You see very well that you are a bride and that he has espoused you-you and everyone else-and not with a ring of silver but with a ring of his own flesh. Look at the tender little child who on the eighth day, when he was circumcised, gave up just so much flesh as to make a tiny circlet of a ring!" Raymond of Capua also records that she was told by Christ to leave her withdrawn life and enter the public life of the world. Catherine rejoined her family and began helping the ill and the poor, where she took care of them in hospitals or homes. Her early pious activities in Siena attracted a group of followers, women and men, who gathered around her. As social and political tensions mounted in Siena, Catherine found herself drawn to intervene in wider politics. She made her first journey to Florence in 1374, probably to be interviewed by the Dominican authorities at the General Chapter held in Florence in May 1374, though this is disputed (if she was interviewed, then the absence of later evidence suggests she was deemed sufficiently orthodox). It seems that at this time she acquired Raymond of Capua as her confessor and spiritual director. After this visit, she began travelling with her followers throughout northern and central Italy advocating reform of the clergy and advising people that repentance and renewal could be done through "the total love for God." In Pisa, in 1375, she used what influence she had to sway that city and Lucca away from alliance with the anti-papal league whose force was gaining momentum and strength. She also lent her enthusiasm towards promoting the launch of a new crusade. It was in Pisa in 1375 that, according to Raymond of Capua's biography, she received the stigmata (visible, at Catherine's request, only to herself). Physical travel was not the only way in which Catherine made her views known. From 1375 onwards, she began dictating letters to scribes. These letters were intended to reach men and women of her circle, increasingly widening her audience to include figures in authority as she begged for peace between the republics and principalities of Italy and for the return of the Papacy from Avignon to Rome. She carried on a long correspondence with Pope Gregory XI, asking him to reform the clergy and the administration of the Papal States. Towards the end of 1375, she returned to Siena, to assist a young political prisoner, Niccolò di Tuldo, at his execution. In June 1376 Catherine went to Avignon as ambassador of the Republic of Florence to make peace with the Papal States (on 31 March 1376 Gregory XI had placed Florence under interdict). She was unsuccessful and was disowned by the Florentine leaders, who sent ambassadors to negotiate on their own terms as soon as Catherine's work had paved the way for them. Catherine sent an appropriately scorching letter back to Florence in response. While in Avignon, Catherine also tried to convince Pope Gregory XI, the last Avignon Pope, to return to Rome. Gregory did indeed return his administration to Rome in January 1377; to what extent this was due to Catherine's influence is a topic of much modern debate. Catherine returned to Siena and spent the early months of 1377 founding a women's monastery of strict observance outside the city in the old fortress of Belcaro. She spent the rest of 1377 at Rocca d'Orcia, about twenty miles from Siena, on a local mission of peace-making and preaching. During this period, in autumn 1377, she had the experience which led to the writing of her "Dialogue" and learned to write, although she still seems to have chiefly relied upon her secretaries for her correspondence. Late in 1377 or early in 1378 Catherine again travelled to Florence, at the order of Gregory XI, to seek peace between Florence and Rome. Following Gregory's death in March 1378 riots, the revolts of the Ciompi, broke out in Florence on 18 June, and in the ensuing violence she was nearly assassinated. Eventually, in July 1378, peace was agreed between Florence and Rome; Catherine returned quietly to Florence. In late November 1378, with the outbreak of the Western Schism, the new Pope, Urban VI, summoned her to Rome. She stayed at Pope Urban VI's court and tried to convince nobles and cardinals of his legitimacy, both meeting with individuals at court and writing letters to persuade others. For many years she had accustomed herself to a rigorous abstinence. She received the Holy Eucharist almost daily. This extreme fasting appeared unhealthy in the eyes of the clergy and her own sisterhood. Her confessor, Raymond, ordered her to eat properly. But Catherine claimed that she was unable to, describing her inability to eat as an "infermità" (illness). From the beginning of 1380, Catherine could neither eat nor swallow water. On February 26 she lost the use of her legs. Catherine died in Rome, on 29 April 1380, at the age of thirty-three, having eight days earlier suffered a massive stroke which paralyzed her from the waist down. Her last words were, "Father, into Your Hands I commend my soul and my spirit." There is some internal evidence of Catherine's personality, teaching and work in her nearly four hundred letters, her "Dialogue", and her prayers. Much detail about her life has also, however, been drawn from the various sources written shortly after her death in order to promote her cult and canonisation. Though much of this material is heavily hagiographic, it has been an important source for historians seeking to reconstruct Catherine's life. Various sources are particularly important, especially the works of Raymond of Capua, who was Catherine's spiritual director and close friend from 1374 until her death, and himself became Master General of the Order in 1380. Raymond began writing what is known as the "Legenda Major", his "Life" of Catherine, in 1384, and completed it in 1395. Another important work written after Catherine's death was "Libellus de Supplemento" ("Little Supplement Book"), written between 1412 and 1418 by Tommaso d'Antonio Nacci da Siena (commonly called Thomas of Siena, or Tommaso Caffarini): the work is an expansion of Raymond's "Legenda Major" making heavy use of the notes of Catherine's first confessor, Tommaso della Fonte (notes that do not survive anywhere else). Caffarini later published a more compact account of Catherine's life, entitled the "Legenda Minor". From 1411 onwards, Caffarini also co-ordinated the compiling of the "Processus" of Venice, the set of documents submitted as part of the process of canonisation of Catherine, which provides testimony from nearly all of Catherine's disciples. There is also an anonymous piece entitled "Miracoli della Beata Caterina" ("Miracle of Blessed Catherine"), written by an anonymous Florentine. A few other relevant pieces survive. Three genres of work by Catherine survive: Catherine's theology can be described as mystical, and was employed towards practical ends for her own spiritual life or those of others. She used the language of medieval scholastic philosophy to elaborate her experiential mysticism. Interested mainly with achieving an incorporeal union with God, Catherine practiced extreme fasting and asceticism, eventually to the extent of living solely off the Eucharist every day. For Catherine, this practice was the means to fully realize her love of Christ in her mystical experience, with a large proportion of her ecstatic visions relating to the consumption or rejection of food during her life. She viewed Christ as a "bridge" between the soul and God and transmitted that idea, along with her other teachings, in her book The Dialogue. The Dialogue is highly systematic and explanatory in its presentation of her mystical ideas; however, these ideas themselves are not so much based on reason or logic as they are based in her ecstatic mystical experience. In one of her letters she sent to her confessor, Raymund of Capua, she recorded this revelation from her conversation with Christ, in which he said: "Do you know what you are to Me, and what I am to you, my daughter? I am He who is, you are she who is not". This mystical concept of God as the wellspring of being is seen in the works and ideas of Aquinas and can be seen as a simplistic rendering of apotheosis and a more rudimentary form of the doctrine of divine simplicity. She describes God in her work, 'the Dialogues', as a "sea, in which we are the fish", the point being that the relationship between God and man should not be seen as man contending against the Divine and vice versa, but as God being the endless being that supports all things. She was buried in the (Roman) cemetery of Santa Maria sopra Minerva which lies near the Pantheon. After miracles were reported to take place at her grave, Raymond moved her inside the Basilica of Santa Maria sopra Minerva, where she lies to this day. Her head however, was parted from her body and inserted in a gilt bust of bronze. This bust was later taken to Siena, and carried through that city in a procession to the Dominican church. Behind the bust walked Lapa, Catherine's mother, who lived until she was 89 years old. By then she had seen the end of the wealth and the happiness of her family, and followed most of her children and several of her grandchildren to the grave. She helped Raymond of Capua write his biography of her daughter, and said, "I think God has laid my soul athwart in my body, so that it can't get out." The incorrupt head and thumb were entombed in the Basilica of San Domenico at Siena, where they remain. Pope Pius II, himself from Siena, canonized Catherine on 29 June 1461. On 4 October 1970, Pope Paul VI named Catherine a Doctor of the Church; this title was almost simultaneously given to Teresa of Ávila (27 September 1970), making them the first women to receive this honour. Initially however, her feast day was not included in the General Roman Calendar. When it was added in 1597, it was put on the day of her death, April 29; however, because this conflicted with the feast of Saint Peter of Verona which also fell on the 29th of April, Catherine's feast day was moved in 1628 to the new date of April 30. In the 1969 revision of the calendar, it was decided to leave the celebration of the feast of St Peter of Verona to local calendars, because he was not as well known worldwide, and Catherine's feast was restored to April 29. In his decree of 13 April 1866, Pope Pius IX declared Catherine of Siena to be a co-patroness of Rome. On 18 June 1939 Pope Pius XII named her a joint patron saint of Italy along with Saint Francis of Assisi. On 1 October 1999, Pope John Paul II made her one of Europe's patron saints, along with Saint Teresa Benedicta of the Cross and Saint Bridget of Sweden. She is also the patroness of the historically Catholic American woman's fraternity, Theta Phi Alpha. The people of Siena wished to have Catherine's body. A story is told of a miracle whereby they were partially successful: knowing that they could not smuggle her whole body out of Rome, they decided to take only her head which they placed in a bag. When stopped by the Roman guards, they prayed to Catherine to help them, confident that she would rather have her body (or at least part thereof) in Siena. When they opened the bag to show the guards, it appeared no longer to hold her head but to be full of rose petals. Catherine ranks high among the mystics and spiritual writers of the Church. She remains a greatly respected figure for her spiritual writings, and political boldness to "speak truth to power"—it being exceptional for a woman, in her time period, to have had such influence in politics and on world history. The main churches in honor of Catherine of Siena are: English translations of The "Dialogue" include: The Letters are translated into English as: The Prayers are translated into English as: Raymond of Capua's "Life" was translated into English in 1493 and 1609, and in Modern English is translated as:
https://en.wikipedia.org/wiki?curid=7471
Charles Lyell Sir Charles Lyell, 1st Baronet, (14 November 1797 – 22 February 1875) was a Scottish geologist who demonstrated the power of known natural causes in explaining Earth's history. He is best known as the author of "Principles of Geology" (1830–33), which presented to a wide public audience the idea that Earth was shaped by the same natural processes still in operation today, operating at similar intensities. The philosopher William Whewell termed this gradualistic view "uniformitarianism" and contrasted it with catastrophism, which had been championed by Georges Cuvier and was better accepted in Europe. The combination of evidence and eloquence in "Principles" convinced a wide range of readers of the significance of "deep time" for understanding the Earth and environment. Lyell's scientific contributions included a pioneering explanation of climate change, in which shifting boundaries between oceans and continents could be used to explain long-term variations in temperature and rainfall. Lyell also gave influential explanations of earthquakes and developed the theory of gradual "backed up-building" of volcanoes. In stratigraphy his division of the Tertiary period into the Pliocene, Miocene, and Eocene was highly influential. He incorrectly conjectured that icebergs may be the emphasis behind the transport of glacial erratics, and that silty loess deposits might have settled out of flood waters. His creation of a separate period for human history, entitled the 'Recent', is widely cited as providing the foundations for the modern discussion of the Anthropocene. Building on the innovative work of James Hutton and his follower John Playfair, Lyell favoured an indefinitely long age for the Earth, despite evidence suggesting an old but finite age. He was a close friend of Charles Darwin, and contributed significantly to Darwin's thinking on the processes involved in evolution. As Darwin wrote in "On the Origin of Species", "He who can read Sir Charles Lyell's grand work on the Principles of Geology, which the future historian will recognise as having produced a revolution in natural science, yet does not admit how incomprehensibly vast have been the past periods of time, may at once close this volume." Lyell helped to arrange the simultaneous publication in 1858 of papers by Darwin and Alfred Russel Wallace on natural selection, despite his personal religious qualms about the theory. He later published evidence from geology of the time man had existed on Earth. Lyell was born into a wealthy family, on 14 November 1797, at the family's estate house, Kinnordy House, near Kirriemuir in Forfarshire. He was the eldest of ten children. Lyell's father, also named Charles Lyell, was noted as a translator and scholar of Dante. An accomplished botanist, it was he who first exposed his son to the study of nature. Lyell's grandfather, also Charles Lyell, had made the family fortune supplying the Royal Navy at Montrose, enabling him to buy Kinnordy House. The family seat is located in Strathmore, near the Highland Boundary Fault. Round the house, in the strath, is good farmland, but within a short distance to the north-west, on the other side of the fault, are the Grampian Mountains in the Highlands. His family's second country home was in a completely different geological and ecological area: he spent much of his childhood at Bartley Lodge in the New Forest, in Hampshire in southern England. Lyell entered Exeter College, Oxford, in 1816, and attended William Buckland's geological lectures. He graduated with a BA Hons. second class degree in classics, in December 1819, and gained his M.A. 1821. After graduation he took up law as a profession, entering Lincoln's Inn in 1820. He completed a circuit through rural England, where he could observe geological phenomena. In 1821 he attended Robert Jameson's lectures in Edinburgh, and visited Gideon Mantell at Lewes, in Sussex. In 1823 he was elected joint secretary of the Geological Society. As his eyesight began to deteriorate, he turned to geology as a full-time profession. His first paper, "On a recent formation of freshwater limestone in Forfarshire", was presented in 1822. By 1827, he had abandoned law and embarked on a geological career that would result in fame and the general acceptance of uniformitarianism, a working out of the ideas proposed by James Hutton a few decades earlier. In 1832, Lyell married Mary Horner in Bonn, daughter of Leonard Horner (1785–1864), also associated with the Geological Society of London. The new couple spent their honeymoon in Switzerland and Italy on a geological tour of the area. During the 1840s, Lyell travelled to the United States and Canada, and wrote two popular travel-and-geology books: "Travels in North America" (1845) and "A Second Visit to the United States" (1849). After the Great Chicago Fire, Lyell was one of the first to donate books to help found the Chicago Public Library. In 1866, he was elected a foreign member of the Royal Swedish Academy of Sciences. Lyell's wife died in 1873, and two years later (in 1875) Lyell himself died as he was revising the twelfth edition of "Principles". He is buried in Westminster Abbey. Lyell was knighted (Kt) in 1848, and later, in 1864, made a baronet (Bt), which is an hereditary honour. He was awarded the Copley Medal of the Royal Society in 1858 and the Wollaston Medal of the Geological Society in 1866. Mount Lyell, the highest peak in Yosemite National Park, is named after him; the crater Lyell on the Moon and a crater on Mars were named in his honour; Mount Lyell in western Tasmania, Australia, located in a profitable mining area, bears Lyell's name; and the Lyell Range in north-west Western Australia is named after him as well. In Southwest Nelson in the South Island of New Zealand, the Lyell Range, Lyell River and the gold mining town of Lyell (now only a camping site) were all named after Lyell. The jawless fish "Cephalaspis lyelli", from the Old Red Sandstone of southern Scotland, was named by Louis Agassiz in honour of Lyell. Lyell had private means, and earned further income as an author. He came from a prosperous family, worked briefly as a lawyer in the 1820s, and held the post of Professor of Geology at King's College London in the 1830s. From 1830 onward his books provided both income and fame. Each of his three major books was a work continually in progress. All three went through multiple editions during his lifetime, although many of his friends (such as Darwin) thought the first edition of the "Principles" was the best written. Lyell used each edition to incorporate additional material, rearrange existing material, and revisit old conclusions in light of new evidence. Throughout his life, Lyell kept a remarkable series of nearly three hundred manuscript notebooks and diaries. These were essential to the development of his ideas, and provide a unique record of his travels, conversations, correspondence, reading and field observations. "Principles of Geology", Lyell's first book, was also his most famous, most influential, and most important. First published in three volumes in 1830–33, it established Lyell's credentials as an important geological theorist and propounded the doctrine of uniformitarianism. It was a work of synthesis, backed by his own personal observations on his travels. The central argument in "Principles" was that "the present is the key to the past" – a concept of the Scottish Enlightenment which David Hume had stated as "all inferences from experience suppose ... that the future will resemble the past", and James Hutton had described when he wrote in 1788 that "from what has actually been, we have data for concluding with regard to that which is to happen thereafter." Geological remains from the distant past can, and should, be explained by reference to geological processes now in operation and thus directly observable. Lyell's interpretation of geological change as the steady accumulation of minute changes over enormously long spans of time was a powerful influence on the young Charles Darwin. Lyell asked Robert FitzRoy, captain of HMS "Beagle", to search for erratic boulders on the survey voyage of the "Beagle", and just before it set out FitzRoy gave Darwin Volume 1 of the first edition of Lyell's "Principles". When the "Beagle" made its first stop ashore at St Jago in the Cape Verde islands, Darwin found rock formations which seen "through Lyell's eyes" gave him a revolutionary insight into the geological history of the island, an insight he applied throughout his travels. While in South America Darwin received Volume 2 which considered the ideas of Lamarck in some detail. Lyell rejected Lamarck's idea of organic evolution, proposing instead "Centres of Creation" to explain diversity and territory of species. However, as discussed below, many of his letters show he was fairly open to the idea of evolution. In geology Darwin was very much Lyell's disciple, and brought back observations and his own original theorising, including ideas about the formation of atolls, which supported Lyell's uniformitarianism. On the return of the "Beagle" (October 1836) Lyell invited Darwin to dinner and from then on they were close friends. Although Darwin discussed evolutionary ideas with him from 1842, Lyell continued to reject evolution in each of the first nine editions of the "Principles". He encouraged Darwin to publish, and following the 1859 publication of "On the Origin of Species", Lyell finally offered a tepid endorsement of evolution in the tenth edition of "Principles". "Elements of Geology" began as the fourth volume of the third edition of "Principles": Lyell intended the book to act as a suitable field guide for students of geology. The systematic, factual description of geological formations of different ages contained in "Principles" grew so unwieldy, however, that Lyell split it off as the "Elements" in 1838. The book went through six editions, eventually growing to two volumes and ceasing to be the inexpensive, portable handbook that Lyell had originally envisioned. Late in his career, therefore, Lyell produced a condensed version titled "Student's Elements of Geology" that fulfilled the original purpose. "Geological Evidences of the Antiquity of Man" brought together Lyell's views on three key themes from the geology of the Quaternary Period of Earth history: glaciers, evolution, and the age of the human race. First published in 1863, it went through three editions that year, with a fourth and final edition appearing in 1873. The book was widely regarded as a disappointment because of Lyell's equivocal treatment of evolution. Lyell, a highly religious man with a strong belief in the special status of human reason, had great difficulty reconciling his beliefs with natural selection. Lyell's geological interests ranged from volcanoes and geological dynamics through stratigraphy, palaeontology, and glaciology to topics that would now be classified as prehistoric archaeology and paleoanthropology. He is best known, however, for his role in elaborating the doctrine of uniformitarianism. He played a critical role in advancing the study of loess. From 1830 to 1833 his multi-volume "Principles of Geology" was published. The work's subtitle was "An attempt to explain the former changes of the Earth's surface by reference to causes now in operation", and this explains Lyell's impact on science. He drew his explanations from field studies conducted directly before he went to work on the founding geology text. He was, along with the earlier John Playfair, the major advocate of James Hutton's idea of uniformitarianism, that the earth was shaped entirely by slow-moving forces still in operation today, acting over a very long period of time. This was in contrast to catastrophism, an idea of abrupt geological changes, which had been adapted in England to explain landscape features--such as rivers much smaller than their associated valleys--that seemed impossible to explain other than through violent action. Criticizing the reliance of his contemporaries on what he argued were "ad hoc" explanations, Lyell wrote, Never was there a doctrine more calculated to foster indolence, and to blunt the keen edge of curiosity, than this assumption of the discordance between the former and the existing causes of change... The student was taught to despond from the first. Geology, it was affirmed, could never arise to the rank of an exact science... [With catastrophism] we see the ancient spirit of speculation revived, and a desire manifestly shown to cut, rather than patiently untie, the Gordian Knot.-Sir Charles Lyell, "Principles of Geology", 1854 edition, p.196; quoted by Stephen Jay Gould. Lyell saw himself as "the spiritual saviour of geology, freeing the science from the old dispensation of Moses." The two terms, "uniformitarianism" and "catastrophism", were both coined by William Whewell; in 1866 R. Grove suggested the simpler term "continuity" for Lyell's view, but the old terms persisted. In various revised editions (12 in all, through 1872), "Principles of Geology" was the most influential geological work in the middle of the 19th century, and did much to put geology on a modern footing. Lyell noted the "economic advantages" that geological surveys could provide, citing their felicity in mineral-rich countries and provinces. Modern surveys, like the British Geological Survey (founded in 1835), and the US Geological Survey (founded in 1879), map and exhibit the natural resources within the country. So, in endorsing surveys, as well as advancing the study of geology, Lyell helped to forward the business of modern extractive industries, such as the coal and oil industry. Before the work of Lyell, phenomena such as earthquakes were understood by the destruction that they brought. One of the contributions that Lyell made in "Principles" was to explain the cause of earthquakes. Lyell, in contrast focused on recent earthquakes (150 yrs), evidenced by surface irregularities such as faults, fissures, stratigraphic displacements and depressions. Lyell's work on volcanoes focused largely on Vesuvius and Etna, both of which he had earlier studied. His conclusions supported gradual building of volcanoes, so-called "backed up-building", as opposed to the upheaval argument supported by other geologists. Lyell was a key figure in establishing the classification of more recent geological deposits, long known as the Tertiary period. From May 1828, until February 1829, he travelled with Roderick Impey Murchison (1792–1871) to the south of France (Auvergne volcanic district) and to Italy. In these areas he concluded that the recent strata (rock layers) could be categorised according to the number and proportion of marine shells encased within. Based on this the third volume of his "Principles of Geology", published in 1833, proposed dividing the Tertiary period into four parts, which he named the Eocene, Miocene, Pliocene, and Recent. In 1839, Lyell termed the Pleistocene epoch, distinguishing a more recent fossil layer from the Pliocene. The Recent epoch renamed the Holocene by French paleontologist Paul Gervais in 1867 included all deposits from the era subject to human observation. In recent years Lyell's subdivisions have been widely discussed in relation to debates about the Anthropocene. In "Principles of Geology" (first edition, vol. 3, ch. 2, 1833) Lyell proposed that icebergs could be the means of transport for erratics. During periods of global warming, ice breaks off the poles and floats across submerged continents, carrying debris with it, he conjectured. When the iceberg melts, it rains down sediments upon the land. Because this theory could account for the presence of diluvium, the word "drift" became the preferred term for the loose, unsorted material, today called "till". Furthermore, Lyell believed that the accumulation of fine angular particles covering much of the world (today called loess) was a deposit settled from mountain flood water. Today some of Lyell's mechanisms for geological processes have been disproven, though many have stood the test of time. His observational methods and general analytical framework remain in use today as foundational principles in geology. Lyell initially accepted the conventional view of other men of science, that the fossil record indicated a directional geohistory in which species went extinct. Around 1826, when he was on circuit, he read Lamarck's "Zoological Philosophy" and on 2 March 1827 wrote to Mantell, expressing admiration, but cautioning that he read it "rather as I hear an advocate on the wrong side, to know what can be made of the case in good hands".: He struggled with the implications for human dignity, and later in 1827 wrote private notes on Lamarck's ideas. Lyell reconciled transmutation of species with natural theology by suggesting that it would be as much a "remarkable manifestation of creative Power" as creating each species separately. He countered Lamarck's views by rejecting continued cooling of the Earth in favour of "a fluctuating cycle", a long-term steady-state geohistory as proposed by James Hutton. The fragmentary fossil record already showed "a high class of fishes, close to reptiles" in the Carboniferous period which he called "the first Zoological era", and quadrupeds could also have existed then. In November 1827, after William Broderip found a Middle Jurassic fossil of the early mammal "Didelphis", Lyell told his father that "There was everything but man even as far back as the Oolite." Lyell inaccurately portrayed Lamarckism as a response to the fossil record, and said it was falsified by a lack of progress. He said in the second volume of "Principles" that the occurrence of this one fossil of the higher mammalia "in these ancient strata, is as fatal to the theory of successive development, as if several hundreds had been discovered." In the first edition of "Principles", the first volume briefly set out Lyell's concept of a steady state with no real progression of fossils. The sole exception was the advent of humanity, with no great physical distinction from animals, but with absolutely unique intellectual and moral qualities. The second volume dismissed Lamarck's claims of animal forms arising from habits, continuous spontaneous generation of new life, and man having evolved from lower forms. Lyell explicitly rejected Lamarck's concept of transmutation of species, drawing on Cuvier's arguments, and concluded that species had been created with stable attributes. He discussed the geographical distribution of plants and animals, and proposed that every species of plant or animal was descended from a pair or individual, originated in response to differing external conditions. Species would regularly go extinct, in a "struggle for existence" between hybrids, or a "war one with another" due to population pressure. He was vague about how replacement species formed, portraying this as an infrequent occurrence which could rarely be observed. The leading man of science Sir John Herschel wrote from Cape Town on 20 February 1836, thanking Lyell for sending a copy of "Principles" and praising the book as opening a way for bold speculation on "that mystery of mysteries, the replacement of extinct species by others" – by analogy with other intermediate causes, "the origination of fresh species, could it ever come under our cognizance, would be found to be a natural in contradistinction to a miraculous process". Lyell replied: "In regard to the origination of new species, I am very glad to find that you think it probable that it may be carried on through the intervention of intermediate causes. I left this rather to be inferred, not thinking it worth while to offend a certain class of persons by embodying in words what would only be a speculation." Whewell subsequently questioned this topic, and in March 1837 Lyell told him: As a result of his letters and, no doubt, personal conversations, Huxley and Haeckel were convinced that, at the time he wrote "Principles", he believed new species had arisen by natural methods. Sedgwick wrote worried letters to him about this. By the time Darwin returned from the "Beagle" survey expedition in 1836, he had begun to doubt Lyell's ideas about the permanence of species. He continued to be a close personal friend, and Lyell was one of the first scientists to support "On the Origin of Species", though he did not subscribe to all its contents. Lyell was also a friend of Darwin's closest colleagues, Hooker and Huxley, but unlike them he struggled to square his religious beliefs with evolution. This inner struggle has been much commented on. He had particular difficulty in believing in natural selection as the main motive force in evolution. Lyell and Hooker were instrumental in arranging the peaceful co-publication of the theory of natural selection by Darwin and Alfred Russel Wallace in 1858: each had arrived at the theory independently. Lyell's views on gradual change and the power of a long time scale were important because Darwin thought that populations of an organism changed very slowly. Although Lyell rejected evolution at the time of writing the "Principles", after the Darwin–Wallace papers and the "Origin" Lyell wrote in one of his notebooks: Lyell's acceptance of natural selection, Darwin's proposed mechanism for evolution, was equivocal, and came in the tenth edition of "Principles". "The Antiquity of Man" (published in early February 1863, just before Huxley's "Man's place in nature") drew these comments from Darwin to Huxley: Quite strong remarks: no doubt Darwin resented Lyell's repeated suggestion that he owed a lot to Lamarck, whom he (Darwin) had always specifically rejected. Darwin's daughter Henrietta (Etty) wrote to her father: "Is it fair that Lyell always calls your theory a modification of Lamarck's?" In other respects "Antiquity" was a success. It sold well, and it "shattered the tacit agreement that mankind should be the sole preserve of theologians and historians". But when Lyell wrote that it remained a profound mystery how the huge gulf between man and beast could be bridged, Darwin wrote "Oh!" in the margin of his copy. Places named after Lyell:
https://en.wikipedia.org/wiki?curid=7472
Chelsea F.C. Chelsea Football Club are an English professional football club based in Fulham, London. Founded in 1905, they compete in the Premier League, the top division of English football. Chelsea are among England's most successful clubs having won over thirty competitive honours, including six league titles and six European trophies. Their home ground is Stamford Bridge. Chelsea won their first major honour, the League Championship, in 1955. They won the FA Cup for the first time in 1970 and their first European honour, the UEFA Cup Winners' Cup, in 1971. After a period of decline in the late 1970s and 1980s, the club enjoyed a revival in the 1990s and had more success in cup competitions. The past two decades have been the most successful in Chelsea's history: they won five of their six league titles and the UEFA Champions League during this period. Chelsea are one of five clubs to have won all three of UEFA's main club competitions, and the only London club to have won the Champions League. Chelsea's home kit colours are royal blue shirts and shorts with white socks. The club's crest features a ceremonial lion rampant regardant holding a staff. The club have rivalries with neighbouring teams Arsenal and Tottenham Hotspur, and a historic rivalry with Leeds United. Based on attendance figures, the club have the sixth-largest fanbase in England. In terms of club value, Chelsea are the sixth most valuable football club in the world, worth £2.13 billion ($2.576 billion), and are the eighth highest-earning football club in the world, with earnings of over €428 million in the 2017–18 season. Since 2003, Chelsea have been owned by Russian billionaire Roman Abramovich. In 1904, Gus Mears acquired the Stamford Bridge athletics stadium with the aim of turning it into a football ground. An offer to lease it to nearby Fulham was turned down, so Mears opted to found his own club to use the stadium. As there was already a team named Fulham in the borough, the name of the adjacent borough of Chelsea was chosen for the new club; names like "Kensington FC", "Stamford Bridge FC" and "London FC" were also considered. Chelsea were founded on 10 March 1905 at The Rising Sun pub (now The Butcher's Hook), opposite the present-day main entrance to the ground on Fulham Road, and were elected to the Football League shortly afterwards. Chelsea won promotion to the First Division in their second season, and yo-yoed between the First and Second Divisions in their early years. They reached the 1915 FA Cup Final, where they lost to Sheffield United at Old Trafford, and finished third in the First Division in 1920, the club's best league campaign to that point. Chelsea had a reputation for signing star players and attracted large crowds. The club had the highest average attendance in English football in ten separate seasons including 1907–08, 1909–10, 1911–12, 1912–13, 1913–14 and 1919–20. They were FA Cup semi-finalists in 1920 and 1932 and remained in the First Division throughout the 1930s, but success eluded the club in the inter-war years. Former Arsenal and England centre-forward Ted Drake was appointed manager in 1952 and proceeded to modernise the club. He removed the club's Chelsea pensioner crest, improved the youth set-up and training regime, rebuilt the side with shrewd signings from the lower divisions and amateur leagues, and led Chelsea to their first major trophy success – the League championship – in 1954–55. The following season saw UEFA create the European Champions' Cup, but after objections from The Football League and the FA, Chelsea were persuaded to withdraw from the competition before it started. Chelsea failed to build on this success, and spent the remainder of the 1950s in mid-table. Drake was dismissed in 1961 and replaced by player-coach Tommy Docherty. Docherty built a new team around the group of talented young players emerging from the club's youth set-up and Chelsea challenged for honours throughout the 1960s, enduring several near-misses. They were on course for a treble of League, FA Cup and League Cup going into the final stages of the 1964–65 season, winning the League Cup but faltering late on in the other two. In three seasons the side were beaten in three major semi-finals and were FA Cup runners-up. Under Docherty's successor, Dave Sexton, Chelsea won the FA Cup in 1970, beating Leeds United 2–1 in a final replay. The following year, Chelsea took their first European honour, a UEFA Cup Winners' Cup triumph, with another replayed win, this time over Real Madrid in Athens. The late 1970s through to the '80s was a turbulent period for Chelsea. An ambitious redevelopment of Stamford Bridge threatened the financial stability of the club, star players were sold and the team were relegated. Further problems were caused by a notorious hooligan element among the support, which was to plague the club throughout the decade. In 1982, Chelsea were, at the nadir of their fortunes, acquired by Ken Bates for the nominal sum of £1, although by now the Stamford Bridge freehold had been sold to property developers, meaning the club faced losing their home. On the pitch, the team had fared little better, coming close to relegation to the Third Division for the first time, but in 1983 manager John Neal put together an impressive new team for minimal outlay. Chelsea won the Second Division title in 1983–84 and established themselves in the top division with two top-six finishes, before being relegated again in 1988. The club bounced back immediately by winning the Second Division championship in 1988–89. After a long-running legal battle, Bates reunited the stadium freehold with the club in 1992 by doing a deal with the banks of the property developers, who had been bankrupted by a market crash. Chelsea's form in the new Premier League was unconvincing, although they did reach the 1994 FA Cup Final. The appointment of Ruud Gullit as player-manager in 1996 began an upturn in the team's fortunes. He added several top international players to the side and led the club to their first major honour since 1971, the FA Cup. Gullit was replaced by Gianluca Vialli, whose reign saw Chelsea win the League Cup, the UEFA Cup Winners' Cup and the UEFA Super Cup in 1998, and the FA Cup in 2000. They also mounted a strong title challenge in 1998–99, finishing four points behind champions Manchester United, and made their first appearance in the UEFA Champions League. Vialli was sacked in favour of Claudio Ranieri, who guided Chelsea to the 2002 FA Cup Final and Champions League qualification in 2002–03. In July 2003, Bates sold Chelsea to Russian billionaire Roman Abramovich for £140 million. Over £100 million was spent on new players, but Ranieri was unable to deliver any trophies, and was replaced by José Mourinho. Under Mourinho, Chelsea became the fifth English team to win back-to-back league championships since the Second World War (2004–05 and 2005–06), in addition to winning an FA Cup (2007) and two League Cups (2005 and 2007). After a poor start to the 2007–2008 season, Mourinho was replaced by Avram Grant, who led the club to their first UEFA Champions League final, which they lost on penalties to Manchester United. In 2009, under caretaker manager Guus Hiddink, Chelsea won another FA Cup. In 2009–10, his successor Carlo Ancelotti led them to their first Premier League and FA Cup Double, also becoming the first English top-flight club to score 100 league goals in a season since 1963. In 2012, Roberto Di Matteo led Chelsea to their seventh FA Cup, and their first UEFA Champions League title, beating Bayern Munich 4–3 on penalties, the first London club to win the trophy. The following year the club won the UEFA Europa League, making them the first club to hold two major European titles simultaneously and one of five clubs to have won the three main UEFA trophies. Mourinho returned as manager in 2013 and led Chelsea to League Cup success in March 2015, and the Premier League title two months later. Mourinho was sacked after four months of the following season after a poor start. In 2017, under new coach Antonio Conte, Chelsea won their sixth English title and the following season won their eighth FA Cup. In 2018 Conte was sacked after a 5th-place finish and replaced with Maurizio Sarri, under whom Chelsea reached the League Cup final, which they lost on penalties to Manchester City and won the Europa League for a second time, beating Arsenal 4–1 in the final. Sarri then left the club to become manager of Juventus and was then replaced by former Chelsea player Frank Lampard. Chelsea have only had one home ground, Stamford Bridge, where they have played since the team's foundation. It was officially opened on 28 April 1877 and for the first 28 years of its existence, it was used almost exclusively by the London Athletic Club as an arena for athletics meetings and not at all for football. In 1904 the ground was acquired by businessman Gus Mears and his brother Joseph, who had also purchased nearby land (formerly a large market garden) with the aim of staging football matches on the now 12.5 acre (51,000 m2) site. Stamford Bridge was designed for the Mears family by the noted football architect Archibald Leitch, who had also designed Ibrox, Craven Cottage and Hampden Park. Most football clubs were founded first, and then sought grounds in which to play, but Chelsea were founded for Stamford Bridge. Starting with an open bowl-like design and one grandstand with seating, Stamford Bridge had an original capacity of around 100,000, making it the second biggest stadium in England after Crystal Palace. The early 1930s saw the construction of a terrace on the southern part of the ground with a roof that covered around one fifth of the stand. As the roof resembled that of a corrugated iron shed, the stand eventually became known as the "Shed End", although it is unknown who first coined this name. Starting in the 1960s, it became known as the home of Chelsea's most loyal and vocal supporters. In 1939, another small seated stand was added, the North Stand, which remained until its demolition in 1975. In the early 1970s, the club's owners announced a modernisation of Stamford Bridge with plans for a state-of-the-art 50,000 all-seater stadium. Work began in 1972 but the project was beset with problems and ultimately only the East Stand was completed; the cost brought the club close to bankruptcy, culminating in the freehold being sold to property developers. Following a long legal battle, it was not until the mid-1990s that Chelsea's future at the stadium was secured and renovation work resumed. The north, west and southern parts of the ground were converted into all-seater stands and moved closer to the pitch, a process completed by 2001. The East Stand was retained from the 1970s development. In 1996, the north stand was renamed the Matthew Harding stand, after the club director and benefactor who was killed in a helicopter crash earlier that year. When Stamford Bridge was redeveloped in the Bates era many additional features were added to the complex including two Millennium & Copthorne hotels, apartments, bars, restaurants, the Chelsea Megastore, and an interactive visitor attraction called Chelsea World of Sport. The intention was that these facilities would provide extra revenue to support the football side of the business, but they were less successful than hoped and before the Abramovich takeover in 2003 the debt taken on to finance them was a major burden on the club. Soon after the takeover a decision was taken to drop the "Chelsea Village" brand and refocus on Chelsea as a football club. However, the stadium is sometimes still referred to as part of ""Chelsea Village"" or ""The Village"". The Stamford Bridge freehold, the pitch, the turnstiles and Chelsea's naming rights are now owned by Chelsea Pitch Owners, a non-profit organisation in which fans are the shareholders. The CPO was created to ensure the stadium could never again be sold to developers. As a condition for using the Chelsea FC name, the club has to play its first team matches at Stamford Bridge, which means that if the club moves to a new stadium, they may have to change their name. Chelsea's training ground is located in Cobham, Surrey. Chelsea moved to Cobham in 2004. Their previous training ground in Harlington was taken over by QPR in 2005. The new training facilities in Cobham were completed in 2007. Stamford Bridge hosted the FA Cup Final from 1920 to 1922, has held ten FA Cup Semi-finals (most recently in 1978), ten FA Charity Shield matches (the last in 1970), and three England international matches, the last in 1932; it was also the venue for an unofficial "Victory International" in 1946. The 2013 UEFA Women's Champions League Final was played at Stamford Bridge as well. The stadium has also been used for a variety of other sports. In October 1905 it hosted a rugby union match between the All Blacks and Middlesex, and in 1914 hosted a baseball match between the touring New York Giants and the Chicago White Sox. It was the venue for a boxing match between world flyweight champion Jimmy Wilde and Joe Conn in 1918. The running track was used for dirt track racing between 1928 and 1932, greyhound racing from 1933 to 1968, and Midget car racing in 1948. In 1980, Stamford Bridge hosted the first international floodlit cricket match in the UK, between Essex and the West Indies. It was also the home stadium of the London Monarchs American Football team for the 1997 season. The current club ownership have stated that a larger stadium is necessary in order for Chelsea to stay competitive with rival clubs who have significantly larger stadia, such as Arsenal and Manchester United. Owing to its location next to a main road and two railway lines, fans can only enter Stamford Bridge via the Fulham Road exits, which places constraints on expansion due to health and safety regulations. The club have consistently affirmed their desire to keep Chelsea at their current home, but have nonetheless been linked with a move to various nearby sites, including the Earls Court Exhibition Centre, Battersea Power Station and the Chelsea Barracks. In October 2011, a proposal from the club to buy back the freehold to the land on which Stamford Bridge sits was voted down by Chelsea Pitch Owners shareholders. In May 2012, the club made a formal bid to purchase Battersea Power Station, with a view to developing the site into a new stadium, but lost out to a Malaysian consortium. The club subsequently announced plans to redevelop Stamford Bridge into a 60,000-seater stadium, and in January 2017 these plans were approved by Hammersmith and Fulham council. However, on 31 May 2018, the club released a statement saying that the new stadium project had been put on hold indefinitely, citing "the current unfavourable investment climate." Chelsea have had four main crests, which all underwent minor variations. The first, adopted when the club was founded, was the image of a Chelsea pensioner, the army veterans who reside at the nearby Royal Hospital Chelsea. This contributed to the club's original "pensioner" nickname, and remained for the next half-century, though it never appeared on the shirts. When Ted Drake became Chelsea manager in 1952, he began to modernise the club. Believing the Chelsea pensioner crest to be old-fashioned, he insisted that it be replaced. A stop-gap badge which comprised the initials C.F.C. was adopted for a year. In 1953, the club crest was changed to an upright blue lion looking backwards and holding a staff. It was based on elements in the coat of arms of the Metropolitan Borough of Chelsea with the "lion rampant regardant" taken from the arms of then club president Viscount Chelsea and the staff from the Abbots of Westminster, former Lords of the Manor of Chelsea. It also featured three red roses, to represent England, and two footballs. This was the first Chelsea crest to appear on the shirts, in the early 1960s. In 1986, with Ken Bates now owner of the club, Chelsea's crest was changed again as part of another attempt to modernise and because the old rampant lion badge could not be trademarked. The new badge featured a more naturalistic non-heraldic lion, in white and not blue, standing over the C.F.C. initials. This lasted for the next 19 years, with some modifications such as the use of different colours, including red from 1987 to 1995, and yellow from 1995 until 1999, before the white returned. With the new ownership of Roman Abramovich, and the club's centenary approaching, combined with demands from fans for the popular 1950s badge to be restored, it was decided that the crest should be changed again in 2005. The new crest was officially adopted for the start of the 2005–06 season and marked a return to the older design, used from 1953 to 1986, featuring a blue heraldic lion holding a staff. For the centenary season this was accompanied by the words '100 YEARS' and 'CENTENARY 2005–2006' on the top and bottom of the crest respectively. Chelsea have always worn blue shirts, although they originally used the paler eton blue, which was taken from the racing colours of then club president, Earl Cadogan, and was worn with white shorts and dark blue or black socks. The light blue shirts were replaced by a royal blue version in around 1912. In the 1960s Chelsea manager Tommy Docherty changed the kit again, switching to blue shorts (which have remained ever since) and white socks, believing it made the club's colours more modern and distinctive, since no other major side used that combination; this kit was first worn during the 1964–65 season. Since then Chelsea have always worn white socks with their home kit apart from a short spell from 1985 to 1992, when blue socks were reintroduced. Chelsea's away colours are usually all yellow or all white with blue trim. More recently, the club have had a number of black or dark blue away kits which alternate every year. As with most teams, they have also had some more unusual ones. At Docherty's behest, in the 1966 FA Cup semi-final they wore blue and black stripes, based on Inter Milan's kit. In the mid-1970s, the away strip was a red, white and green kit inspired by the Hungarian national side of the 1950s. Other away kits include an all jade strip worn from 1986–89, red and white diamonds from 1990–92, graphite and tangerine from 1994–96, and luminous yellow from 2007–08. The graphite and tangerine strip has appeared in lists of the worst football kits ever. Chelsea are among the most widely supported football clubs in the world. They have the sixth highest average all-time attendance in English football, and regularly attract over 40,000 fans to Stamford Bridge; they were the seventh best-supported Premier League team in the 2013–14 season, with an average gate of 41,572. Chelsea's traditional fanbase comes from all over the Greater London area including working-class parts such as Hammersmith and Battersea, wealthier areas like Chelsea and Kensington, and from the home counties. There are also numerous official supporters clubs in the United Kingdom and all over the world. Between 2007 and 2012, Chelsea were ranked fourth worldwide in annual replica kit sales, with an average of 910,000. As of 2018, Chelsea had 72.2 million followers on social media, the fourth highest among football clubs. At matches, Chelsea fans sing chants such as "Carefree" (to the tune of "Lord of the Dance", whose lyrics were probably written by supporter Mick Greenaway), "Ten Men Went to Mow", "We All Follow the Chelsea" (to the tune of "Land of Hope and Glory"), "Zigga Zagga", and the celebratory "Celery". The latter is often accompanied by fans throwing celery at each other, although the vegetable was banned inside Stamford Bridge after an incident involving Arsenal midfielder Cesc Fàbregas at the 2007 League Cup Final. During the 1970s and 1980s in particular, Chelsea supporters were associated with football hooliganism. The club's "football firm", originally known as the Chelsea Shed Boys, and subsequently as the Chelsea Headhunters, were nationally notorious for football violence, alongside hooligan firms from other clubs such as West Ham United's Inter City Firm and Millwall's Bushwackers, before, during and after matches. The increase of hooligan incidents in the 1980s led chairman Ken Bates to propose erecting an electric fence to deter them from invading the pitch, a proposal that the Greater London Council rejected. Since the 1990s, there has been a marked decline in crowd trouble at matches, as a result of stricter policing, CCTV in grounds and the advent of all-seater stadia. In 2007, the club launched the Back to the Shed campaign to improve the atmosphere at home matches, with notable success. According to Home Office statistics, 126 Chelsea fans were arrested for football-related offences during the 2009–10 season, the third highest in the division, and 27 banning orders were issued, the fifth-highest in the division. Chelsea have long-standing rivalries with North London clubs Arsenal and Tottenham Hotspur. A strong rivalry with Leeds United dates back to several heated and controversial matches in the 1960s and 1970s, particularly the 1970 FA Cup Final. More recently a rivalry with Liverpool has grown following repeated clashes in cup competitions. Chelsea's fellow West London sides Brentford, Fulham and Queens Park Rangers are generally not considered major rivals, as matches have only taken place intermittently due to the clubs often being in separate divisions. A 2004 survey by Planetfootball.com found that Chelsea fans consider their main rivalries to be with (in descending order): Arsenal, Tottenham Hotspur and Manchester United. In the same survey, fans of Arsenal, Fulham, Leeds United, QPR, Tottenham, and West Ham United named Chelsea as one of their three main rivals. In a 2008 poll conducted by the Football Fans Census, Chelsea fans named Liverpool, Arsenal and Manchester United as their most disliked clubs. In the same survey, "Chelsea" was the top answer to the question "Which other English club do you dislike the most?" A 2012 survey, conducted among 1200 supporters of the top four league divisions across the country, found that many clubs' main rivals had changed since 2003 and reported that Chelsea fans consider Tottenham to be their main rivals, above Arsenal and Manchester United. Additionally, fans of Arsenal, Brentford, Fulham, Liverpool, Manchester United, QPR, Tottenham and West Ham identified Chelsea as one of their top three rivals. Chelsea's highest appearance-maker is ex-captain Ron Harris, who played in 795 competitive games for the club between 1961 and 1980. Four other players made more than 500 appearances for the club: Peter Bonetti (729; 1959–79), John Terry (717; 1998–2017), Frank Lampard (648; 2001–2014) and John Hollins (592; 1963–1975 and 1983–1984). With 103 caps (101 while at the club) for England, Lampard is Chelsea's most capped international player. Every starting player in Chelsea's 57 games of the 2013–14 season was a full international – a new club record. Lampard is Chelsea's all-time top goalscorer, having scored 211 goals in 648 games (2001–2014); he passed Bobby Tambling's longstanding record of 202 in May 2013. Eight other players have also scored over 100 goals for Chelsea: George Hilsdon (1906–1912), George Mills (1929–1939), Roy Bentley (1948–1956), Jimmy Greaves (1957–1961), Peter Osgood (1964–1974 and 1978–1979), Kerry Dixon (1983–1992), Didier Drogba (2004–2012 and 2014–2015), and Eden Hazard (2012–2019). Greaves holds the club record for the most goals scored in one season (43 in 1960–61). While a Chelsea player, Greaves also became the youngest ever player to score 100 goals in the English top-flight, at 20 years and 290 days. Chelsea's biggest winning scoreline in a competitive match is 13–0, achieved against Jeunesse Hautcharage in the Cup Winners' Cup in 1971. The club's biggest top-flight win was an 8–0 victory against Wigan Athletic in 2010, which was matched in 2012 against Aston Villa. Chelsea's biggest loss was an 8–1 reverse against Wolverhampton Wanderers in 1953. The club's 21–0 aggregate victory over Jeunesse Hautcharage in the UEFA Cup Winners' Cup in 1971 is also a record in European competition. Officially, Chelsea's highest home attendance is 82,905 for a First Division match against Arsenal on 12 October 1935. However, an estimated crowd of over 100,000 attended a friendly match against Soviet team Dynamo Moscow on 13 November 1945. From 20 March 2004 to 26 October 2008, Chelsea went a record 86 consecutive league matches at home without defeat, beating the previous record of 63 matches unbeaten set by Liverpool between 1978 and 1980. Chelsea hold the English record for the fewest goals conceded during a league season (15), the highest number of clean sheets overall in a Premier League season (25) (both set during the 2004–05 season), and the most consecutive clean sheets from the start of a league season (6, set during the 2005–06 season). Chelsea's streak of eleven consecutive away league wins, set between 5 April 2008 and 6 December 2008, is a record for the English top flight. Chelsea are the only Premier League side to win their opening nine league games of the season, doing so in 2005–06. From 2009 to 2013, Chelsea were unbeaten in a record 29 consecutive FA Cup matches (excluding penalty shoot-outs). On 25 August 1928, Chelsea, along with Arsenal, became the first club to play with shirt numbers, in their match against Swansea Town. They were the first English side to travel by aeroplane to a domestic away match, when they visited Newcastle United on 19 April 1957, and the first First Division side to play a match on a Sunday, when they faced Stoke City on 27 January 1974. On 26 December 1999, Chelsea became the first British side to field an entirely foreign starting line-up (no British or Irish players) in a Premier League match against Southampton. In May 2007, Chelsea were the first team to win the FA Cup at the new Wembley Stadium, having also been the last to win it at the old Wembley. They were the first English club to be ranked No. 1 under UEFA's five-year coefficient system in the 21st century. They were the first Premier League team, and the first team in the English top flight since 1962–63, to score at least 100 goals in a single season, reaching the milestone during the 2009–10 season. Chelsea are the only London club to win the UEFA Champions League. Upon winning the 2012–13 UEFA Europa League, Chelsea became the first English club to win all four UEFA club trophies and the only club to hold the Champions League and the Europa League at the same time. Chelsea have twice broken the record for the highest transfer fee paid by a British club. Their £30.8 million purchase of Andriy Shevchenko from AC Milan in June 2006 was a British record until surpassed by the £32.5 million paid by Manchester City for Robinho in September 2008. The club's £50 million purchase of Fernando Torres from Liverpool in January 2011 held the record until Ángel Di María signed for Manchester United in August 2014 for £59.7 million. The club's £71 million purchase of Kepa Arrizabalaga in August 2018 remains a world record fee paid for a goalkeeper. Chelsea Football Club were founded by Gus Mears in 1905. After his death in 1912, his descendants continued to own the club until 1982, when Ken Bates bought the club from Mears' great-nephew Brian Mears for £1. Bates bought a controlling stake in the club and floated Chelsea on the AIM stock exchange in March 1996. In the mid-1990s Chelsea fan and businessman Matthew Harding became a director and loaned the club £26 million to build the new North Stand and invest in new players. In July 2003, Roman Abramovich purchased just over 50% of Chelsea Village plc's share capital, including Bates' 29.5% stake, for £30 million and over the following weeks bought out most of the remaining 12,000 shareholders at 35 pence per share, completing a £140 million takeover. Other shareholders at the time of the takeover included the Matthew Harding estate (21%), BSkyB (9.9%) and various anonymous offshore trusts. After passing the 90% share threshold, Abramovich took the club back into private hands, delisting it from the AIM on 22 August 2003. At the time of the Abramovich takeover, the club also had debts of around £100 million, which included a ten-year £75 million Eurobond taken out in 1997 by the Bates regime to buy the freehold of Stamford Bridge and finance the redevelopment of the stadium. The 9% interest on the loan cost the club around £7 million a year and according to Bruce Buck, Chelsea were struggling to pay an instalment due in July 2003. Abramovich paid off some of this debt immediately, but the outstanding £36 million on the Eurobond was not fully repaid until 2008. Since then, the club have had no external debt. Abramovich changed the ownership name to Chelsea FC plc, whose ultimate parent company is Fordstam Limited, which is controlled by him. Chelsea are additionally funded by Abramovich via interest free soft loans channelled through his holding company Fordstam Limited. The loans stood at £709 million in December 2009, when they were all converted to equity by Abramovich, leaving the club themselves debt free, although the debt remains with Fordstam. Chelsea did not turn a profit in the first nine years of Abramovich's ownership, and made record losses of £140m in June 2005. In November 2012, Chelsea announced a profit of £1.4 million for the year ending 30 June 2012, the first time the club had made a profit under Abramovich's ownership. This was followed by a loss in 2013 and then their highest ever profit of £18.4 million for the year to June 2014. In 2018 Chelsea announced a record after-tax profit of £62 million. Chelsea have been described as a global brand; a 2012 report by Brand Finance ranked Chelsea fifth among football brands and valued the club's brand value at US$398 million – an increase of 27% from the previous year, also valuing them at US$10 million more than the sixth best brand, London rivals Arsenal – and gave the brand a strength rating of AA (very strong). In 2016, "Forbes" magazine ranked Chelsea the seventh most valuable football club in the world, at £1.15 billion ($1.66 billion). , Chelsea are ranked eighth in the Deloitte Football Money League with an annual commercial revenue of £322.59 million. The club's recent accounting records highlight £26.6m they have lost in compensation to former head coach Antonio Conte for sacking and to pay off his backroom staff and the legal costs that followed. Chelsea's kit has been manufactured by Nike since July 2017. Previously, the kit was manufactured by Adidas, which was originally contracted to supply the club's kit from 2006 to 2018. The partnership was extended in October 2010 in a deal worth £160 million over eight years. This deal was again extended in June 2013 in a deal worth £300 million over another ten years. In May 2016, Adidas announced that by mutual agreement, the kit sponsorship would end six years early on 30 June 2017. Chelsea had to pay £40m in compensation to Adidas. In October 2016, Nike was announced as the new kit sponsor, in a deal worth £900m over 15 years, until 2032. Previously, the kit was manufactured by Umbro (1975–81), Le Coq Sportif (1981–86), The Chelsea Collection (1986–87), Umbro (1987–2006), and Adidas (2006–2017). Chelsea's first shirt sponsor was Gulf Air, agreed during the 1983–84 season. The club were then sponsored by Grange Farms, Bai Lin Tea and Simod before a long-term deal was signed with Commodore International in 1989; Amiga, an offshoot of Commodore, also appeared on the shirts. Chelsea were subsequently sponsored by Coors beer (1994–97), Autoglass (1997–2001), Emirates (2001–05), Samsung Mobile (2005–08), Samsung (2008–15). and Yokohama Tyres (2015-20). Worth £40 million per year, the deal is second in English football to Chevrolet's £50 million-per-year sponsorship of Manchester United. Chelsea's current sponsor is Three, starting from July 2020. Following the introduction of sleeve sponsors in the Premier League, Chelsea had Alliance Tyres as their first sleeve sponsor in the 2017–18 season. For the 2018–19 season, they have Hyundai Motor Company as the new sleeve sponsor. The club also has a variety of other sponsors and official partners, which include Carabao, Delta Air Lines, Beats by Dre, Singha, EA Sports, Rexona, Hublot, Ericsson, William Hill, Levy Restaurants, Wipro, Grand Royal Whisky, Bangkok Bank, Guangzhou R&F, Mobinil, IndusInd Bank, and Ole777. In 1930, Chelsea featured in one of the earliest football films, "The Great Game". One-time Chelsea centre forward, Jack Cock, who by then was playing for Millwall, was the star of the film and several scenes were shot at Stamford Bridge, including on the pitch, the boardroom, and the dressing rooms. It included guest appearances by then-Chelsea players Andrew Wilson, George Mills, and Sam Millington. Owing to the notoriety of the Chelsea Headhunters, a football firm associated with the club, Chelsea have also featured in films about football hooliganism, including 2004's "The Football Factory". Chelsea also appear in the Hindi film "Jhoom Barabar Jhoom". In April 2011, Montenegrin comedy series "Nijesmo mi od juče" made an episode in which Chelsea play against FK Sutjeska Nikšić for qualification of the UEFA Champions League. Up until the 1950s, the club had a long-running association with the music halls; their underachievement often provided material for comedians such as George Robey. It culminated in comedian Norman Long's release of a comic song in 1933, ironically titled "On the Day That Chelsea Went and Won the Cup", the lyrics of which describe a series of bizarre and improbable occurrences on the hypothetical day when Chelsea finally won a trophy. In Alfred Hitchcock's 1935 film "The 39 Steps", Mr Memory claims that Chelsea last won the Cup in 63 BC, "in the presence of the Emperor Nero." Scenes in a 1980 episode of "Minder" were filmed during a real match at Stamford Bridge between Chelsea and Preston North End with Terry McCann (played by Dennis Waterman) standing on the terraces. The song "Blue is the Colour" was released as a single in the build-up to the 1972 League Cup Final, with all members of Chelsea's first team squad singing; it reached number five in the UK Singles Chart. The song has since been adopted as an anthem by a number of other sports teams around the world, including the Vancouver Whitecaps (as "White is the Colour") and the Saskatchewan Roughriders (as "Green is the Colour"). In the build-up to the 1997 FA Cup Final, the song "Blue Day", performed by Suggs and members of the Chelsea squad, reached number 22 in the UK charts. Chelsea also operate a women's football team, Chelsea Football Club Women, formerly known as Chelsea Ladies. They have been affiliated to the men's team since 2004 and are part of the club's Community Development programme. They play their home games at Kingsmeadow, the home ground of the EFL League One club AFC Wimbledon. The club were promoted to the Premier Division for the first time in 2005 as Southern Division champions and won the Surrey County Cup nine times between 2003 and 2013. In 2010 Chelsea Ladies were one of the eight founder members of the FA Women's Super League. In 2015, Chelsea Ladies won the FA Women's Cup for the first time, beating Notts County Ladies at Wembley Stadium, and a month later clinched their first FA WSL title to complete a league and cup double. In 2018 they won a second league and FA Cup double. John Terry, former captain of the Chelsea men's team, is the President of Chelsea Women. Source: Chelsea F.C. The following managers won at least one trophy when in charge of Chelsea:
https://en.wikipedia.org/wiki?curid=7473
Arthur St. Clair Arthur St. Clair ( – August 31, 1818) was a Scottish-American soldier and politician. Born in Thurso, Scotland, he served in the British Army during the French and Indian War before settling in Pennsylvania, where he held local office. During the American Revolutionary War, he rose to the rank of major general in the Continental Army, but lost his command after a controversial retreat from Fort Ticonderoga. After the war, he served as President of the Continental Congress, which during his term passed the Northwest Ordinance. He was then made governor of the Northwest Territory in 1788, and then the portion that would become Ohio in 1800. In 1791, St. Clair commanded the American forces in what was the United States's worst ever defeat by the American Indians. Politically out-of-step with the Jefferson administration, he was replaced as governor in 1802. St. Clair was born in Thurso, Caithness, Scotland. Little is known of his early life. Early biographers estimated his year of birth as 1734, but subsequent historians uncovered a birth date of March 23, 1736, which in the modern calendar system means that he was born in 1737. His parents, unknown to early biographers, were probably William Sinclair, a merchant, and Elizabeth Balfour. He reportedly attended the University of Edinburgh before being apprenticed to the renowned physician William Hunter. In 1757, St. Clair purchased a commission in the British Army, Royal American Regiment, and came to America with Admiral Edward Boscawen's fleet for the French and Indian War. He served under General Jeffery Amherst at the capture of Louisburg, Nova Scotia on July 26, 1758. On April 17, 1759, he received a lieutenant's commission and was assigned under the command of General James Wolfe, under whom he served at the Battle of the Plains of Abraham which resulted in the capture of Quebec City. On April 16, 1762, he resigned his commission, and, in 1764, he settled in Ligonier Valley, Pennsylvania, where he purchased land and erected mills. He was the largest landowner in Western Pennsylvania. In 1770, St. Clair became a justice of the court, of quarter sessions and of common pleas, a member of the proprietary council, a justice, recorder, and clerk of the orphans' court, and prothonotary of Bedford and Westmoreland counties. In 1774, the colony of Virginia took claim of the area around Pittsburgh, Pennsylvania, and some residents of Western Pennsylvania took up arms to eject them. St. Clair issued an order for the arrest of the officer leading the Virginia troops. Lord Dunmore's War eventually settled the boundary dispute. By the mid-1770s, St. Clair considered himself more of an American than a British subject. In January 1776, he accepted a commission in the Continental Army as a colonel of the 3rd Pennsylvania Regiment. He first saw service in the later days of the Quebec invasion, where he saw action in the Battle of Trois-Rivières. He was appointed a brigadier general in August 1776, and was sent by Gen. George Washington to help organize the New Jersey militia. He took part in George Washington's crossing of the Delaware River on the night of December 25–26, 1776, before the Battle of Trenton on the morning of December 26. Many biographers credit St. Clair with the strategy that led to Washington's capture of Princeton, New Jersey on January 3, 1777. St. Clair was promoted to major general in February 1777. In April 1777, St. Clair was sent to defend Fort Ticonderoga. His small garrison could not resist British General John Burgoyne's larger force in the Saratoga campaign. St. Clair was forced to retreat at the Siege of Fort Ticonderoga on July 5, 1777. He withdrew his forces and played no further part in the campaign. In 1778 he was court-martialed for the loss of Ticonderoga. The court exonerated him and he returned to duty, although he was no longer given any battlefield commands. He still saw action, however, as an aide-de-camp to General Washington, who retained a high opinion of him. St. Clair was at Yorktown when Lord Cornwallis surrendered his army. St. Clair was a member of the Pennsylvania Council of Censors in 1783, and was elected a delegate to the Confederation Congress, serving from November 2, 1785, until November 28, 1787. Chaos ruled the day in early 1787 with Shays's Rebellion in full force and the states refusing to settle land disputes or contribute to the now six-year-old federal government. On February 2, 1787, the delegates finally gathered into a quorum and elected St. Clair to a one-year term as President of the Continental Congress. Congress enacted its most important piece of legislation, the Northwest Ordinance, during St. Clair's tenure as president. Time was running out for the Confederation Congress, however; during St. Clair's presidency, the Philadelphia Convention was drafting a new United States Constitution, which would abolish the old Congress. Under the Northwest Ordinance of 1787, which created the Northwest Territory, General St. Clair was appointed governor of what is now Ohio, Indiana, Illinois, Michigan, along with parts of Wisconsin and Minnesota. He named Cincinnati, Ohio, after the Society of the Cincinnati, and it was there that he established his home. As Governor, he formulated Maxwell's Code (named after its printer, William Maxwell), the first written laws of the territory. He also sought to end Native American claims to Ohio land and clear the way for white settlement. In 1789, he succeeded in getting certain Indians to sign the Treaty of Fort Harmar, but many native leaders had not been invited to participate in the negotiations, or had refused to do so. Rather than settling the Indians' claims, the treaty provoked them to further resistance in what is also sometimes known as the "Northwest Indian War" (or "Little Turtle's War"). Mutual hostilities led to a campaign by General Josiah Harmar, whose 1,500 militiamen were defeated by the Indians in October 1790. In March 1791, St. Clair succeeded Harmar as commander of the United States Army and was commissioned as a major general. He personally led a punitive expedition involving two Regular Army regiments and some militia. In October 1791 as an advance post for his campaign, Fort Jefferson (Ohio) was built under the direction of General Arthur St. Clair. Located in present-day Darke County in far western Ohio, the fort was built of wood and intended primarily as a supply depot; accordingly, it was originally named Fort Deposit. One month later, near modern-day Fort Recovery, his force advanced to the location of Indian settlements near the headwaters of the Wabash River, but on November 4 they were routed in battle by a tribal confederation led by Miami Chief Little Turtle and Shawnee chief Blue Jacket. They were aided by British collaborators Alexander McKee and Simon Girty. More than 600 soldiers and scores of women and children were killed in the battle, which has since borne the name "St. Clair's Defeat", also known as the "Battle of the Wabash", the "Columbia Massacre," or the "Battle of a Thousand Slain". It remains the greatest defeat of a US Army by Native Americans in history, with about 623 American soldiers killed in action and about 50 Native Americans killed. The wounded were many, including St. Clair and Capt. Robert Benham. Although an investigation exonerated him, St. Clair resigned his army commission in March 1792 at the request of President Washington, but he continued to serve as Governor of the Northwest Territory. A Federalist, St. Clair hoped to see two states made of the Ohio Territory in order to increase Federalist power in Congress. However, he was opposed by Ohio Democratic-Republicans for what were perceived as his partisanship, high-handedness, and arrogance in office. In 1802, St. Clair remarked the U.S. Congress had no power to interfere in the affairs of those in the Ohio Territory. He also stated the people of the territory "are no more bound by an act of Congress than we would be bound by an edict of the first consul of France." This led President Thomas Jefferson to remove him from office as territorial governor. He thus played no part in the organizing of the state of Ohio in 1803. The first Ohio Constitution provided for a weak governor and a strong legislature, in part as a reaction to St. Clair's method of governance. St. Clair met Phoebe Bayard, a member of one of the most prominent families in Boston, and they were married in 1760. Miss Bayard's mother's maiden name was Bowdoin and she was the sister of James Bowdoin, colonial governor of Massachusetts. In retirement St. Clair lived with his daughter, Louisa St. Clair Robb, and her family on the ridge between Ligonier and Greensburg. Arthur St. Clair died in poverty in Greensburg, Pennsylvania, on August 31, 1818, at the age of 81. His remains are buried under a Masonic monument in St. Clair Park in downtown Greensburg. St Clair had been a petitioner for a Charter for Nova Caesarea Lodge #10 in Cincinnati, Ohio in 1791. This Lodge exists today, as Nova Caesarea Harmony #2. His wife Phoebe died shortly after and is buried beside him. A portion of the Hermitage, St. Clair's home in Oak Grove, Pennsylvania (north of Ligonier), was later moved to Ligonier, Pennsylvania, where it is now preserved, along with St. Clair artifacts and memorabilia at the Fort Ligonier Museum. An American Civil War steamer was named USS "St. Clair". Places named in honor of Arthur St. Clair include: In Pennsylvania: In Ohio: Other States: In Scotland: Notes Books
https://en.wikipedia.org/wiki?curid=2639
Ajaccio Ajaccio (, also ; or , ; , ; ) is a French commune, prefecture of the department of Corse-du-Sud, and head office of the "Collectivité territoriale de Corse" (capital city of Corsica). It is also the largest settlement on the island. Ajaccio is located on the west coast of the island of Corsica, southeast of Marseille. The original city went into decline in the Middle Ages, but began to prosper again after the Genoese built a citadel in 1492 to the south of the earlier settlement. After the Corsican Republic was declared in 1755 the Genoese continued to hold several citadels, including Ajaccio, until the French took control of the island. The inhabitants of the commune are known as "Ajacciens" or "Ajacciennes". The most famous of these is Napoleon Bonaparte who was born in Ajaccio in 1769, and whose ancestral home, the Maison Bonaparte, is now a museum. Other dedications to him in the city include Ajaccio Napoleon Bonaparte Airport. Ajaccio is located on the west coast of the island of Corsica, southeast of Marseille. The commune occupies a sheltered position at the foot of wooded hills on the northern shore of the Gulf of Ajaccio between Gravona and the "pointe de la Parata" and includes the "îles Sanguinaires" (Bloody Islands). The harbour lies to the east of the original citadel below a hill overlooking a peninsula which protects the harbour in the south where the Quai de la Citadelle and the Jettée de la Citadelle are. The modern city not only encloses the entire harbour but takes up the better part of the Gulf of Ajaccio and in suburban form extends for some miles up the valley of the Gravona River. The flow from that river is nearly entirely consumed as the city's water supply. Many beaches and coves border its territory and the terrain is particularly rugged in the west where the highest point is . Although the commune of Ajaccio has a large area (82.03 km2), only a small portion of this is urbanized. Therefore, the urban area of Ajaccio is located in the east of the commune on a narrow coastal strip forming a densely populated arc. The rest of the territory is natural with habitation of little importance and spread thinly. Suburbanization occurs north and east of the main urban area. The original urban core, close to the old marshy plain of "Cannes" was abandoned in favour of the current city which was built near the "Punta della Lechia". It has undergone various improvements, particularly under Napoleon, who originated the two current major structural arteries (the "Cours Napoleon" oriented north–south and the "Cours Grandval" oriented east–west). Ajaccio experienced a demographic boom in the 1960s, which explains why 85% of dwellings are post-1949. This is reflected in the layout of the city which is marked by very large areas of low-rise buildings and concrete towers, especially on the heights ("Jardins de l'Empereur") and in the north of the city - e.g. the waterfront, "Les Cannes", and "Les Salines". A dichotomy appears in the landscape between the old city and the imposing modern buildings. Ajaccio gives the image of a city built on two different levels. The city has a Mediterranean climate which is "Csa" in the Köppen climate classification. The average annual sunshine is 2726 hours. There are important local climatic variations, especially with wind exposure and total precipitation, between the city centre, the airport, and the "îles Sanguinaires". The annual average rainfall is at the "Campo dell'Oro" weather station (as per the chart) and at the "Parata": the third-driest place in metropolitan France. The heat and dryness of summer are somewhat tempered by the proximity of the Mediterranean Sea except when the sirocco is blowing. In autumn and spring, heavy rain-storm episodes may occur. Winters are mild and snow is rare. Ajaccio is the French city which holds the record for the number of thunderstorms in the reference period 1971–2000 with an average of 39 thunderstorm days per year. On 14 September 2009, the city was hit by a tornado with an intensity of F1 on the Fujita scale. There was little damage except torn billboards, flying tiles, overturned cars, and broken windows but no casualties. Weather Data for Ajaccio Several hypotheses have been advanced as to the etymology of the name "Ajaccio" ("Aiacciu" in Corsican, "Addiazzo" on old documents). Among these, the most prestigious suggests that the city was founded by the Greek legendary hero "Ajax" and named after him. Other more realistic explanations are, for example, that the name could be related to the Tuscan "agghiacciu" meaning "sheep pens". Another explanation, supported by Byzantine sources from around the year 600AD called the city "Agiation" which suggests a possible Greek origin for the word, "agathè" could mean "good luck" or "good mooring" (this was also the root of the name of the city of Agde). The city was not mentioned by the Greek geographer Ptolemy of Alexandria in the 2nd century AD despite the presence of a place called "Ourkinion" in the "Cinarca" area. It is likely that the city of Ajaccio had its first development at this time. The 2nd century was a period of prosperity in the Mediterranean basin (the Pax Romana) and there was a need for a proper port at the head of the several valleys that lead to the Gulf able to accommodate large ships. Some important underwater archaeological discoveries recently made of Roman ships tend to confirm this. Further excavations conducted recently led to the discovery of important early Christian remains likely to significantly a reevaluation upwards of the size of Ajaccio city in Late Antiquity and the beginning of the Middle Ages. The city was in any case already significant enough to be the seat of a diocese, mentioned by Pope Gregory the Great in 591. The city was then further north than the location chosen later by the Genoese - in the location of the existing quarters of "Castel Vecchio" and "Sainte-Lucie". The earliest certain written record of a settlement at Ajaccio with a name ancestral to its name was the exhortation in Epistle 77 written in 601AD by Gregory the great to the Defensor Boniface, one of two known rectors of the early Corsican church, to tell him not to leave Aléria and Adjacium without bishops. There is no earlier use of the term and Adjacium is not an attested Latin word, which probably means that it is a Latinization of a word in some other language. The Ravenna Cosmography of about 700 AD cites Agiation, which sometimes is taken as evidence of a prior Greek city, as -ion appears to be a Greek ending. There is, however, no evidence at all of a Greek presence on the west coast and the Ionians at Aléria on the east coast had been expelled by the Etruscans long before Roman domination. Ptolemy, who must come the closest to representing indigenous names, lists the Lochra River just south of a feature he calls the "sandy shore" on the southwest coast. If the shore is the Campo dell'Oro (Place of Gold) the Lochra would seem to be the combined mouth of the Gravona and Prunelli Rivers, neither one of which sounds like Lochra. North of there was a Roman city, Ourchinion. The western coastline was so distorted, however, that it is impossible to say where Adjacium was; certainly, he would have known its name and location if he had had any first-hand knowledge of the island and if in fact it was there. Ptolemy's Ourchinion is further north than Ajaccio and does not have the same name. It could be Sagone. The lack of correspondence between Ptolemaic and historical names known to be ancient has no defense except in the case of the two Roman colonies, Aleria and Mariana. In any case the population of the region must belong to Ptolemy's Tarabeni or Titiani people, neither of which are ever heard about again. The population of the city throughout the centuries maintained an oral tradition that it had originally been Roman. Travellers of the 19th century could point to the Hill of San Giovanni on the northwest shore of the Gulf of Ajaccio, which still had a cathedral said to have been the 6th-century seat of the Bishop of Ajaccio. The Castello Vecchio ("old castle"), a ruined citadel, was believed to be Roman but turned out to have Gothic features. The hill was planted with vines. The farmers kept turning up artifacts and terracotta funerary urns that seemed to be Roman. In the 20th century the hill was covered over with buildings and became a part of downtown Ajaccio. In 2005 construction plans for a lot on the hill offered the opportunity to the Institut national de recherches archéologiques preventatives (Inrap) to excavate. They found the baptistry of a 6th-century cathedral and large amounts of pottery dated to the 6th and 7th centuries AD; in other words, an early Christian town. A cemetery had been placed over the old church. In it was a single Roman grave covered over with roof tiles bearing short indecipherable inscriptions. The finds of the previous century had included Roman coins. This is the only evidence so far of a Roman city continuous with the early Christian one. It has been established that after the 8th century the city, like most other Corsican coastal communities, strongly declined and disappeared almost completely. Nevertheless, a castle and a cathedral were still in place in 1492 which last was not demolished until 1748. Towards the end of the 15th century, the Genoese were eager to assert their dominance in the south of the island and decided to rebuild the city of Ajaccio. Several sites were considered: the "Pointe de la Parata" (not chosen because it was too exposed to the wind), the ancient city (finally considered unsafe because of the proximity of the salt ponds), and finally the "Punta della Lechia" which was finally selected. Work began on the town on 21 April 1492 south of the Christian village by the Bank of Saint George at Genoa, who sent Cristoforo of Gandini, an architect, to build it. He began with a castle on Capo di Bolo, around which he constructed residences for several hundred people. The new city was essentially a colony of Genoa. The Corsicans were restricted from the city for some years. Nevertheless, the town grew rapidly and became the administrative capital of the province of "Au Delà Des Monts" (more or less the current "Corse-du-Sud"). Bastia remained the capital of the entire island. Although at first populated exclusively by the Genoese, the city slowly opened to the Corsicans while the Ajaccians, almost to the French conquest, were legally citizens of the Republic of Genoa and were happy to distinguish themselves from the insular "paesani" who lived mainly in "Borgu", a suburb outside the city walls (the current "rue Fesch" was the main street). Ajaccio was occupied from 1553 to 1559 by the French, but it again fell to the Genoese after the Treaty of Cateau Cambresis in the latter year. Subsequently, the Republic of Genoa was strong enough to keep Corsica until 1755, the year Pasquale Paoli proclaimed the Corsican Republic. Paoli took most of the island for the republic, but he was unable to force Genoese troops out of the citadels of Saint-Florent, Calvi, Ajaccio, Bastia and Algajola. Leaving them there, he went on to build the nation, while the Republic of Genoa was left to ponder prospects and solutions. Their ultimate solution was to sell Corsica to France in 1768 and French troops of the Ancien Régime replaced Genoese ones in the citadels, including Ajaccio's. Corsica was formally annexed to France in 1780. Napoleon Bonaparte (born as Napoleone di Buonaparte) was born at Ajaccio in the same year as the Battle of Ponte Novu, 1769. The Buonaparte family at the time had a modest four-story home in town (now a museum known as Maison Bonaparte) and a rarely used country home in the hills north of the city (now site of the Arboretum des Milelli). The father of the family, attorney Carlo di Buonaparte, was secretary to Pasquale Paoli during the Corsican Republic. After the defeat of Paoli, the Comte de Marbeuf began to meet with some leading Corsicans to outline the shape of the future and enlist their assistance. The Comte was among a delegation from Ajaccio in 1769, offered his loyalty and was appointed assessor. Marbeuf also offered Carlo di Buonaparte an appointment for one of his sons to the Military College of Brienne, but Napoleone did not speak French which was a requirement and he had to be at least ten years of age. There is a dispute concerning Napoleon's age because of this requirement; the emperor is known to have altered the civic records at Ajaccio concerning himself and it is possible that he was born in Corte in 1768 when his father was there on business. In any case Napoleon was sent to a school in Autun to learn basic French, then after a year went to Brienne from 1779–1784. At Brienne Napoleon concentrated on studies. He wrote a boyish history of Corsica. He did not share his father's views but held Pasquale Paoli in high esteem and was at heart a Corsican nationalist. The top students were encouraged to go into the artillery. After graduation and a brief sojourn at the Military School of Paris Napoleon applied for a second-lieutenancy in the artillery regiment of La Fère at Valence and after a time was given the position. Meanwhile, his father died and his mother was cast into poverty in Corsica, still having four children to support. Her only income was Napoleon's meager salary. The regiment was in Auxonne when the revolution broke out in the summer of 1789. Napoleon returned on leave to Ajaccio in October, became a Jacobin and began to work for the revolution. The National Assembly in Paris united Corsica to France and pardoned its exiles. Paoli returned in 1790 after 21 years and kissed the soil on which he stood. He and Napoleon met and toured the battlefield of Paoli's defeat. A national assembly at Orezza created the department of Corsica and Paoli was subsequently elected president. He commanded the national guard raised by Napoleon. After a brief return to his regiment Napoleon was promoted to First Lieutenant and came home again on leave in 1791. The death of a rich uncle relieved the family's poverty. All officers were recalled from leave in 1792, intervention threatened and war with Austria (Marie-Antoinette's homeland) began. Napoleon returned to Paris for review, was exonerated, then promoted to Captain and given leave to escort his sister, a schoolgirl, back to Corsica at state expense. His family was prospering; his estate increased. Napoleon became a Lieutenant-Colonel in the Corsican National Guard. Paoli sent him off on an expedition to Sardinia, ordered by France, under Paolis's nephew but the nephew had secret orders from Paoli to make sure the expedition failed. Paoli was now a conservative, opposing the execution of the king and supporting an alliance with Great Britain. Returning from Sardinia Napoleon with his family and all his supporters were instrumental in getting Paoli denounced at the National Convention in Paris in 1793. Napoleon earned the hatred of the Paolists by pretending to support Paoli and then turning against him (payment, one supposes, for Sardinia). Paoli was convicted in absentia, a warrant was issued for his arrest (which could not be served) and Napoleon was dispatched to Corsica as Inspector General of Artillery to take the citadel of Ajaccio from the royalists who had held it since 1789. The Paolists combining with the royalists defeated the French in two pitched battles and Napoleon and his family went on the run, hiding by day, while the Paolists burned their estate. Napoleon and his mother, Laetitia, were taken out by ship in June 1793, by friends while two of the girls found refuge with other friends. They landed in Toulon with only Napoleon's pay for their support. The Bonapartes moved to Marseille but in August Toulon offered itself to the British and received the protection of a fleet under Admiral Hood. The Siege of Toulon began in September under revolutionary officers mainly untrained in the art of war. Napoleon happened to present socially one evening and during a casual conversation over a misplaced 24-pounder explained the value of artillery. Taken seriously he was allowed to bring up over 100 guns from coastal emplacements but his plan for the taking of Toulon was set aside as one incompetent officer superseded another. By December they decided to try his plan and made him a Colonel. Placing the guns at close range he used them to keep the British fleet away while he battered down the walls of Toulon. As soon as the Committee of Public Safety heard of the victory Napoleon became a Brigadier General, the start of his meteoric rise to power. The Bonapartes were back in Ajaccio in 1797 under the protection of General Napoleon. Soon after Napoleon became First Consul and then emperor, using his office to spread revolution throughout Europe. In 1811 he made Ajaccio the capital of the new Department of Corsica. Despite his subsequent defeat by the Prussians, Russians, and British, his exile and his death, no victorious power reversed that decision or tried to remove Corsica from France. Among the natives, though Corsican nationalism is strong, and feeling often runs high in favour of a union with Italy; loyalty to France, however, as evidenced by elections, remains stronger. In the 19th century Ajaccio became a popular winter resort of the high society of the time, especially for the English, in the same way as Monaco, Cannes, and Nice. An Anglican Church was even built. The first prison in France for children was built in Ajaccio in 1855: the Horticultural colony of Saint Anthony. It was a correctional colony for juvenile delinquents (from 8 to 20 years old), established under Article 10 of the Act of 5 August 1850. Nearly 1,200 children from all over France stayed there until 1866, when it was closed. Sixty percent of them perished, the victims of poor sanitation and malaria which infested the unhealthy areas that they were responsible to clean. On 9 September 1943, the people of Ajaccio rose up against the Nazi occupiers and became the first French town to be liberated from the domination of the Germans. General Charles de Gaulle went to Ajaccio on 8 October 1943 and said: "We owe it to the field of battle the lesson of the page of history that was written in French Corsica. Corsica to her fortune and honour is the first morsel of France to be liberated; which was done intentionally and willingly, in the light of its liberation, this demonstrates that these are the intentions and the will of the whole nation." Throughout this period, no Jew was executed or deported from Corsica through the protection afforded by its people and its government. This event now allows Corsica to aspire to the title "righteous among the nations", as no region except for the commune Le Chambon-sur-Lignon in Haute-Loire carries this title. Their case is being investigated . Since the middle of the 20th century, Ajaccio has seen significant development. The city has seen population growth and considerable urban sprawl. Today Ajaccio is the capital of Corsica and the main town of the island and seeks to establish itself as a true regional centre. The city is, with Bastia, the economic, commercial and administrative centre of Corsica. Its urban area of nearly 90,000 inhabitants is spread over a large part of the Corse-du-Sud, on either side of the Gulf of Ajaccio and up the valley of the Gravona. Its business is primarily oriented towards the services sector. The services sector is by far the main source of employment in the city. Ajaccio is an administrative centre comprising communal, intercommunal, departmental, regional, and prefectural services. It is also a shopping centre with the commercial streets of the city centre and the areas of peripheral activities such as that of "Mezzavia" (hypermarket "Géant Casino") and along the ring road (hypermarket Carrefour and E. Leclerc). Tourism is one of the most vital aspects of the economy, split between the seaside tourism of summer, cultural tourism, and fishing. A number of hotels, varying from one star to five star, are present across the commune. Ajaccio is the seat of the "Chamber of Commerce and Industry of Ajaccio and Corsica South". It manages the ports of Ajaccio, Bonifacio, Porto-Vecchio, Propriano and the Tino Rossi marina. It also manages Ajaccio airport and Figari airport as well as the convention centre and the "Centre of Ricanto". Secondary industry is underdeveloped, apart from the aeronautical company "Corsica Aerospace Composites CCA", the largest company on the island with 135 employees at two sites. The storage sites of GDF Suez (formerly Gaz de France) and Antargaz in the district of "Vazzio" are classified as high risk. The "Centrale EDF du Vazzio", a heavy oil power station, provides the south of the island with electricity. The Gravona Canal delivers water for consumption by the city. By road, the city is accessible from National Route NR194 from Bastia and NR193 via NR196 from Bonifacio. These two main axes, as well as the roads leading to suburban villages, connect Ajaccio from the north - the site of Ajaccio forming a dead end blocked by the sea to the south. Only the "Cours Napoleon" and the "Boulevard du Roi Jerome" cross the city. Along with the high urban density, this explains the major traffic and parking problems especially during peak hours and during the summer tourist season. A bypass through several neighbourhoods is nearing completion. The Muvistrada provide services on 21 urban routes, one "city" route for local links and 20 suburban lines. The frequency varies according to demand with intervals of 30 minutes for the most important routes: A park and ride with 300 spaces was built at "Mezzana" in the neighbouring commune of Sarrola-Carcopino in order to promote intermodality between cars and public transport. It was inaugurated on 12 July 2010. In addition, the municipality has introduced a Tramway between Mezzana station in the suburbs and Ajaccio station located in the centre. The city is served by an Ajaccio Napoleon Bonaparte Airport which is the headquarters of Air Corsica, a Corsican airline. It connects Ajaccio to a number of cities in mainland France (including Paris, Marseille, Nice, and Brive) and to places in Europe to serve the tourist industry. The airline CCM Airlines also has its head office on the grounds of the Airport. The port of Ajaccio is connected to the French mainland on an almost daily basis (Marseille, Toulon, Nice). There are also occasional links to the Italian mainland (Livorno) and to Sardinia, as well as a seasonal service serving Calvi and Propriano. The two major shipping companies providing these links are SNCM and Corsica Ferries. Ajaccio has also become a stopover for cruises with a total of 418,086 passengers in 2007by far the largest in Corsica and the second-largest in France (after Marseille, but ahead of Nice/Villefranche-sur-Mer and Cannes). The goal is for Ajaccio to eventually become the premier French port for cruises as well as being a main departure point. The Port function of the city is also served by the commercial, pleasure craft, and artisanal fisheries (3 ports). The railway station in Ajaccio belongs to "Chemins de Fer de la Corse" and is located near the port at the "Square Pierre Griffi". It connects Ajaccio to Corte, Bastia (3h 25min) and Calvi. There are two optional stops: Ajaccio was successively: Ajaccio remained (with some interruptions) an electoral stronghold of the Bonapartist (CCB) party until the municipal elections of 2001. The outgoing municipality was then beaten by a left-wing coalition led by Simon Renucci which gathered Social Democrats, Communists, and Charles Napoleon - the pretender to the imperial throne. List of Successive Mayors of Ajaccio 10 Quarters are recognized by the municipality. Since December 2001, Ajaccio has been part of the "Communauté d'agglomération du Pays Ajaccien" with nine other communes: Afa, Alata, Appietto, Cuttoli-Corticchiato, Peri, Sarrola-Carcopino, Tavaco, Valle-di-Mezzana, and Villanova. The geopolitical arrangements of the commune are slightly different from those typical of Corsica and France. Usually an arrondissement includes cantons and a canton includes one to several communes including the chef-lieu, "chief place", from which the canton takes its name. The city of Ajaccio is one commune, but it contains four cantons, Cantons 1–4, and a fraction of Canton 5. The latter contains three other communes: Bastelicaccia, Alata and Villanova, making a total of four communes for the five cantons of Ajaccio. Each canton contains a certain number of quartiers, "quarters". Cantons 1, 2, 3, 4 are located along the Gulf of Ajaccio from west to east, while 5 is a little further up the valleys of the Gravona and the Prunelli Rivers. These political divisions subdivide the population of Ajaccio into units that can be more democratically served but they do not give a true picture of the size of Ajaccio. In general language, "greater Ajaccio" includes about 100,000 people with all the medical, educational, utility and transportational facilities of a big city. Up until World War II it was still possible to regard the city as being a settlement of narrow streets localized to a part of the harbour or the Gulf of Ajaccio: such bucolic descriptions do not fit the city of today, and travelogues intended for mountain or coastal recreational areas do not generally apply to Corsica's few big cities. The arrondissement contains other cantons that extend generally up the two rivers into central Corsica. Ajaccio has twinning associations with: The demographic development of Ajaccio occurred mainly between 1945 and 1975 with a doubling of the population of the city in that period. This is explicable in the 1950s by the rural exodus. From the 1960s, the city saw the coming of "Pied-Noirs" (French Algerians) including immigrants from the Maghreb and French from mainland France. Ajaccio has three hospital sites: Ajaccio is the headquarters of the Academy of Corsica. The city of Ajaccio has: Higher education is undeveloped except for a few BTS and IFSI, the University of Corsica Pascal Paoli is located in Corte. A research facility of INRA is also located on Ajaccio. Ajaccio has a varied tourism potential, with both a cultural framework in the centre of the city and a natural heritage around the coves and beaches of the Mediterranean Sea, as well as the Natura 2000 reserve of the "îles Sanguinaires". The commune has many buildings and structures that are registered as historical monuments: The town is the seat of a bishopric dating at least from the 7th century. It has tribunals of first instance and of commerce, training colleges, a communal college, a museum and a library; the three latter are established in the Palais Fesch, founded by Cardinal Fesch, who was born at Ajaccio in 1763. The commune has several religious buildings and structures that are registered as historical monuments: There are various sports facilities developed throughout the city. Units that were stationed in Ajaccio:
https://en.wikipedia.org/wiki?curid=2640
Ajanta Caves The Ajanta Caves are approximately 30 rock-cut Buddhist cave monuments which date from the 2nd century BCE to about 480 CE in Aurangabad district of Maharashtra state of India. The caves include paintings and rock-cut sculptures described as among the finest surviving examples of ancient Indian art, particularly expressive paintings that present emotions through gesture, pose and form. According to UNESCO, these are masterpieces of Buddhist religious art that influenced the Indian art that followed. The caves were built in two phases, the first phase starting around the 2nd century BCE, while the second phase was built around 400–650 CE, according to older accounts, or in a brief period of 460–480 CE according to later scholarship. The site is a protected monument in the care of the Archaeological Survey of India, and since 1983, the Ajanta Caves have been a UNESCO World Heritage Site. The Ajanta Caves constitute ancient monasteries and worship-halls of different Buddhist traditions carved into a wall of rock. The caves also present paintings depicting the past lives and rebirths of the Buddha, pictorial tales from Aryasura's "Jatakamala", and rock-cut sculptures of Buddhist deities. Textual records suggest that these caves served as a monsoon retreat for monks, as well as a resting site for merchants and pilgrims in ancient India. While vivid colours and mural wall-painting were abundant in Indian history as evidenced by historical records, Caves 16, 17, 1 and 2 of Ajanta form the largest corpus of surviving ancient Indian wall-painting. The Ajanta Caves are mentioned in the memoirs of several medieval-era Chinese Buddhist travellers to India and by a Mughal-era official of Akbar era in the early 17th century. They were covered by jungle until accidentally "discovered" and brought to Western attention in 1819 by a colonial British officer Captain John Smith on a tiger-hunting party. The caves are in the rocky northern wall of the U-shaped gorge of the river Waghur, in the Deccan plateau. Within the gorge are a number of waterfalls, audible from outside the caves when the river is high. With the Ellora Caves, Ajanta is one of the major tourist attractions of Maharashtra. It is about The Ajanta Caves are generally agreed to have been made in two distinct periods, the first during the 2nd century BCE to 1st century CE, and a second several centuries later. The caves consist of 36 identifiable foundations, some of them discovered after the original numbering of the caves from 1 through 29. The later-identified caves have been suffixed with the letters of the alphabet, such as 15A, identified between originally numbered caves 15 and 16. The cave numbering is a convention of convenience, and does not reflect the chronological order of their construction. The earliest group consists of caves 9, 10, 12, 13 and 15A. The murals in these caves depict stories from the Jatakas. Later caves reflect the artistic influence of the Gupta period, but there are differing opinions on which century in which the early caves were built. According to Walter Spink, they were made during the period 100 BCE to 100 CE, probably under the patronage of the Hindu Satavahana dynasty (230 BCE – c. 220 CE) who ruled the region. Other datings prefer the period of the Maurya Empire (300 BCE to 100 BCE). Of these, caves 9 and 10 are stupa containing worship halls of "chaitya-griha" form, and caves 12, 13, and 15A are "vihāras" (see the architecture section below for descriptions of these types). The first Satavahana period caves lacked figurative sculpture, emphasizing the stupa instead. According to Spink, once the Satavahana period caves were made, the site was not further developed for a considerable period until the mid-5th century. However, the early caves were in use during this dormant period, and Buddhist pilgrims visited the site, according to the records left by Chinese pilgrim Faxian around 400 CE. The second phase of construction at the Ajanta Caves site began in the 5th century. For a long time it was thought that the later caves were made over an extended period from the 4th to the 7th centuries CE, but in recent decades a series of studies by the leading expert on the caves, Walter M. Spink, have argued that most of the work took place over the very brief period from 460 to 480 CE, during the reign of Hindu Emperor Harishena of the Vākāṭaka dynasty. This view has been criticised by some scholars, but is now broadly accepted by most authors of general books on Indian art, for example, Huntington and Harle. The second phase is attributed to the theistic Mahāyāna, or Greater Vehicle tradition of Buddhism. Caves of the second period are 1–8, 11, 14–29, some possibly extensions of earlier caves. Caves 19, 26, and 29 are "chaitya-grihas", the rest "viharas". The most elaborate caves were produced in this period, which included some refurbishing and repainting of the early caves. Spink states that it is possible to establish dating for this period with a very high level of precision; a fuller account of his chronology is given below. Although debate continues, Spink's ideas are increasingly widely accepted, at least in their broad conclusions. The Archaeological Survey of India website still presents the traditional dating: "The second phase of paintings started around 5th–6th centuries A.D. and continued for the next two centuries". According to Spink, the construction activity at the incomplete Ajanta Caves was abandoned by wealthy patrons in about 480 CE, a few years after the death of Harishena. However, states Spink, the caves appear to have been in use for a period of time as evidenced by the wear of the pivot holes in caves constructed close to 480 CE. The second phase of constructions and decorations at Ajanta corresponds to the very apogee of Classical India, or India's golden age. However, at that time, the Gupta Empire was already weakening from internal political issues and from the assaults of the Hūṇas, so that the Vakatakas were actually one of the most powerful empires in India. Some of the Hūṇas, the Alchon Huns of Toramana, were precisely ruling the neighbouring area of Malwa, at the doorstep of the Western Deccan, at the time the Ajanta caves were made. Through their control of vast areas of northwestern India, the Huns may actually have acted as a cultural bridge between the area of Gandhara and the Western Deccan, at the time when the Ajanta or Pitalkhora caves were being decorated with some designs of Gandharan inspiration, such as Buddhas dressed in robes with abundant folds. According to Richard Cohen, a description of the caves by 7th-century Chinese traveler Xuanzang and scattered medieval graffiti suggest that the Ajanta Caves were known and probably in use subsequently, but without a stable or steady Buddhist community presence. The Ajanta caves are mentioned in the 17th-century text "Ain-i-Akbari" by Abu al-Fazl, as twenty four rock-cut cave temples each with remarkable idols. On 28 April 1819, a British officer named John Smith, of the 28th Cavalry, while hunting tigers, discovered the entrance to Cave No. 10 when a local shepherd boy guided him to the location and the door. The caves were well known by locals already. Captain Smith went to a nearby village and asked the villagers to come to the site with axes, spears, torches, and drums, to cut down the tangled jungle growth that made entering the cave difficult. He then vandalised the wall by scratching his name and the date over the painting of a bodhisattva. Since he stood on a five-foot high pile of rubble collected over the years, the inscription is well above the eye-level gaze of an adult today. A paper on the caves by William Erskine was read to the Bombay Literary Society in 1822. Within a few decades, the caves became famous for their exotic setting, impressive architecture, and above all their exceptional and unique paintings. A number of large projects to copy the paintings were made in the century after rediscovery. In 1848, the Royal Asiatic Society established the "Bombay Cave Temple Commission" to clear, tidy and record the most important rock-cut sites in the Bombay Presidency, with John Wilson as president. In 1861 this became the nucleus of the new Archaeological Survey of India. During the colonial era, the Ajanta site was in the territory of the princely state of the Hyderabad and not British India. In the early 1920s, the Nizam of Hyderabad appointed people to restore the artwork, converted the site into a museum and built a road to bring tourists to the site for a fee. These efforts resulted in early mismanagement, states Richard Cohen, and hastened the deterioration of the site. Post-independence, the state government of Maharashtra built arrival, transport, facilities, and better site management. The modern Visitor Center has good parking facilities and public conveniences and ASI operated buses run at regular intervals from Visitor Center to the caves. The Ajanta Caves, along with the Ellora Caves, have become the most popular tourist destination in Maharashtra, and are often crowded at holiday times, increasing the threat to the caves, especially the paintings. In 2012, the Maharashtra Tourism Development Corporation announced plans to add to the ASI visitor centre at the entrance complete replicas of caves 1, 2, 16 & 17 to reduce crowding in the originals, and enable visitors to receive a better visual idea of the paintings, which are dimly-lit and hard to read in the caves. The caves are carved out of flood basalt rock of a cliff, part of the Deccan Traps formed by successive volcanic eruptions at the end of the Cretaceous geological period. The rock is layered horizontally, and somewhat variable in quality. This variation within the rock layers required the artists to amend their carving methods and plans in places. The inhomogeneity in the rock have also led to cracks and collapses in the centuries that followed, as with the lost portico to cave 1. Excavation began by cutting a narrow tunnel at roof level, which was expanded downwards and outwards; as evidenced by some of the incomplete caves such as the partially-built "vihara" caves 21 through 24 and the abandoned incomplete cave 28. The sculpture artists likely worked at both excavating the rocks and making the intricate carvings of pillars, roof, and idols; further, the sculpture and painting work inside a cave were integrated parallel tasks. A grand gateway to the site was carved, at the apex of the gorge's horseshoe between caves 15 and 16, as approached from the river, and it is decorated with elephants on either side and a nāga, or protective Naga (snake) deity. Similar methods and application of artist talent is observed in other cave temples of India, such as those from Hinduism and Jainism. These include the Ellora Caves, Ghototkacha Caves, Elephanta Caves, Bagh Caves, Badami Caves, Aurangabad Caves and Shivleni Caves. The caves from the first period seem to have been paid for by a number of different patrons to gain merit, with several inscriptions recording the donation of particular portions of a single cave. The later caves were each commissioned as a complete unit by a single patron from the local rulers or their court elites, again for merit in Buddhist afterlife beliefs as evidenced by inscriptions such as those in Cave 17. After the death of Harisena, smaller donors motivated by getting merit added small "shrinelets" between the caves or add statues to existing caves, and some two hundred of these "intrusive" additions were made in sculpture, with a further number of intrusive paintings, up to three hundred in cave 10 alone. The majority of the caves are "vihara" halls with symmetrical square plans. To each vihara hall are attached smaller square dormitory cells cut into the walls. A vast majority of the caves were carved in the second period, wherein a shrine or sanctuary is appended at the rear of the cave, centred on a large statue of the Buddha, along with exuberantly detailed reliefs and deities near him as well as on the pillars and walls, all carved out of the natural rock. This change reflects the shift from Hinayana to Mahāyāna Buddhism. These caves are often called monasteries. The central square space of the interior of the viharas is defined by square columns forming a more-or-less square open area. Outside this are long rectangular aisles on each side, forming a kind of cloister. Along the side and rear walls are a number of small cells entered by a narrow doorway; these are roughly square, and have small niches on their back walls. Originally they had wooden doors. The centre of the rear wall has a larger shrine-room behind, containing a large Buddha statue. The viharas of the earlier period are much simpler, and lack shrines. Spink places the change to a design with a shrine to the middle of the second period, with many caves being adapted to add a shrine in mid-excavation, or after the original phase. The plan of Cave 1 shows one of the largest viharas, but is fairly typical of the later group. Many others, such as Cave 16, lack the vestibule to the shrine, which leads straight off the main hall. Cave 6 is two viharas, one above the other, connected by internal stairs, with sanctuaries on both levels. The other type of main hall architecture is the narrower rectangular plan with high arched ceiling type "chaitya-griha" – literally, "the house of stupa". This hall is longitudinally divided into a nave and two narrower side aisles separated by a symmetrical row of pillars, with a stupa in the apse. The stupa is surrounded by pillars and concentric walking space for circumambulation. Some of the caves have elaborate carved entrances, some with large windows over the door to admit light. There is often a colonnaded porch or verandah, with another space inside the doors running the width of the cave. The oldest worship halls at Ajanta were built in the 2nd to 1st century BCE, the newest ones in the late 5th century CE, and the architecture of both resembles the architecture of a Christian church, but without the crossing or chapel chevette. The Ajanta Caves follow the Cathedral-style architecture found in still older rock-cut cave carvings of ancient India, such as the Lomas Rishi Cave of the Ajivikas near Gaya in Bihar dated to the 3rd century BCE. These chaitya-griha are called worship or prayer halls. The four completed "chaitya" halls are caves 9 and 10 from the early period, and caves 19 and 26 from the later period of construction. All follow the typical form found elsewhere, with high ceilings and a central "nave" leading to the stupa, which is near the back, but allows walking behind it, as walking around stupas was (and remains) a common element of Buddhist worship ("pradakshina"). The later two have high ribbed roofs carved into the rock, which reflect timber forms, and the earlier two are thought to have used actual timber ribs and are now smooth, the original wood presumed to have perished. The two later halls have a rather unusual arrangement (also found in Cave 10 at Ellora) where the stupa is fronted by a large relief sculpture of the Buddha, standing in Cave 19 and seated in Cave 26. Cave 29 is a late and very incomplete "chaitya" hall. The form of columns in the work of the first period is very plain and un-embellished, with both "chaitya" halls using simple octagonal columns, which were later painted with images of the Buddha, people and monks in robes. In the second period columns were far more varied and inventive, often changing profile over their height, and with elaborate carved capitals, often spreading wide. Many columns are carved over all their surface with floral motifs and Mahayana deities, some fluted and others carved with decoration all over, as in cave 1. The paintings in the Ajanta caves predominantly narrate the Jataka tales. These are Buddhist legends describing the previous births of the Buddha. These fables embed ancient morals and cultural lores that are also found in the fables and legends of Hindu and Jain texts. The Jataka tales are exemplified through the life example and sacrifices that the Buddha made in hundreds of his past incarnations, where he is depicted as having been reborn as an animal or human. Mural paintings survive from both the earlier and later groups of caves. Several fragments of murals preserved from the earlier caves (Caves 10 and 11) are effectively unique survivals of ancient painting in India from this period, and "show that by Sātavāhana times, if not earlier, the Indian painters had mastered an easy and fluent naturalistic style, dealing with large groups of people in a manner comparable to the reliefs of the Sāñcī toraņa crossbars". Some connections with the art of Gandhara can also be noted, and there is evidence of a shared artistic idiom. Four of the later caves have large and relatively well-preserved mural paintings which, states James Harle, "have come to represent Indian mural painting to the non-specialist", and represent "the great glories not only of Gupta but of all Indian art". They fall into two stylistic groups, with the most famous in Caves 16 and 17, and apparently later paintings in Caves 1 and 2. The latter group were thought to be a century or later than the others, but the revised chronology proposed by Spink would place them in the 5th century as well, perhaps contemporary with it in a more progressive style, or one reflecting a team from a different region. The Ajanta frescos are classical paintings and the work of confident artists, without cliches, rich and full. They are luxurious, sensuous and celebrate physical beauty, aspects that early Western observers felt were shockingly out of place in these caves presumed to be meant for religious worship and ascetic monastic life. The paintings are in "dry fresco", painted on top of a dry plaster surface rather than into wet plaster. All the paintings appear to be the work of painters supported by discriminating connoisseurship and sophisticated patrons from an urban atmosphere. We know from literary sources that painting was widely practised and appreciated in the Gupta period. Unlike much Indian mural painting, compositions are not laid out in horizontal bands like a frieze, but show large scenes spreading in all directions from a single figure or group at the centre. The ceilings are also painted with sophisticated and elaborate decorative motifs, many derived from sculpture. The paintings in cave 1, which according to Spink was commissioned by Harisena himself, concentrate on those Jataka tales which show previous lives of the Buddha as a king, rather than as deer or elephant or another Jataka animal. The scenes depict the Buddha as about to renounce the royal life. In general the later caves seem to have been painted on finished areas as excavating work continued elsewhere in the cave, as shown in caves 2 and 16 in particular. According to Spink's account of the chronology of the caves, the abandonment of work in 478 after a brief busy period accounts for the absence of painting in places including cave 4 and the shrine of cave 17, the later being plastered in preparation for paintings that were never done. Walter M. Spink has over recent decades developed a very precise and circumstantial chronology for the second period of work on the site, which unlike earlier scholars, he places entirely in the 5th century. This is based on evidence such as the inscriptions and artistic style, dating of nearby cave temple sites, comparative chronology of the dynasties, combined with the many uncompleted elements of the caves. He believes the earlier group of caves, which like other scholars he dates only approximately, to the period "between 100 BCE – 100 CE", were at some later point completely abandoned and remained so "for over three centuries". This changed during the Hindu emperor Harishena of the Vakataka Dynasty, who reigned from 460 to his death in 477, who sponsored numerous new caves during his reign. Harisena's rule extended the Central Indian Vakataka Empire to include a stretch of the east coast of India; the Gupta Empire ruled northern India at the same period, and the Pallava dynasty much of the south. According to Spink, Harisena encouraged a group of associates, including his prime minister Varahadeva and Upendragupta, the sub-king in whose territory Ajanta was, to dig out new caves, which were individually commissioned, some containing inscriptions recording the donation. This activity began in many caves simultaneously about 462. This activity was mostly suspended in 468 because of threats from the neighbouring Asmaka kings. Thereafter work continued on only Caves 1, Harisena's own commission, and 17–20, commissioned by Upendragupta. In 472 the situation was such that work was suspended completely, in a period that Spink calls "the Hiatus", which lasted until about 475, by which time the Asmakas had replaced Upendragupta as the local rulers. Work was then resumed, but again disrupted by Harisena's death in 477, soon after which major excavation ceased, except at cave 26, which the Asmakas were sponsoring themselves. The Asmakas launched a revolt against Harisena's son, which brought about the end of the Vakataka Dynasty. In the years 478–480 CE major excavation by important patrons was replaced by a rash of "intrusions" – statues added to existing caves, and small shrines dotted about where there was space between them. These were commissioned by less powerful individuals, some monks, who had not previously been able to make additions to the large excavations of the rulers and courtiers. They were added to the facades, the return sides of the entrances, and to walls inside the caves. According to Spink, "After 480, not a single image was ever made again at the site". However, there exists a Rashtrakuta inscription outside of cave 26 dateable to end of seventh or early 8th century, suggesting the caves were not abandoned until then. Spink does not use "circa" in his dates, but says that "one should allow a margin of error of one year or perhaps even two in all cases". The Ajanta Caves were built in a period when both the Buddha and the Hindu gods were simultaneously revered in Indian culture. According to Spink and other scholars, the royal Vakataka sponsors of the Ajanta Caves probably worshipped both Hindu and Buddhist gods. This is evidenced by inscriptions in which these rulers, who are otherwise known as Hindu devotees, made Buddhist dedications to the caves. According to Spink, A terracotta plaque of Mahishasuramardini, also known as Durga, was also found in a burnt-brick vihara monastery facing the caves on the right bank of the river Waghora that has been recently excavated. This suggest that the deity was possibly under worship by the artisans. According to Yuko Yokoschi and Walter Spink, the excavated artifacts of the 5th century near the site suggest that the Ajanta caves deployed a huge number of builders. Cave 1 was built on the eastern end of the horseshoe-shaped scarp and is now the first cave the visitor encounters. This cave, when first made, would have been a less prominent position, right at the end of the row. According to Spink, it is one of the last caves to have been excavated, when the best sites had been taken, and was never fully inaugurated for worship by the dedication of the Buddha image in the central shrine. This is shown by the absence of sooty deposits from butter lamps on the base of the shrine image, and the lack of damage to the paintings that would have happened if the garland-hooks around the shrine had been in use for any period of time. Spink states that the Vākāṭaka Emperor Harishena was the benefactor of the work, and this is reflected in the emphasis on imagery of royalty in the cave, with those Jataka tales being selected that tell of those previous lives of the Buddha in which he was royal. The cliff has a more steep slope here than at other caves, so to achieve a tall grand facade it was necessary to cut far back into the slope, giving a large courtyard in front of the facade. There was originally a columned portico in front of the present facade, which can be seen "half-intact in the 1880s" in pictures of the site, but this fell down completely and the remains, despite containing fine carvings, were carelessly thrown down the slope into the river, from where they have been lost. This cave (35.7 m × 27.6 m) has one of the most elaborate carved façades, with relief sculptures on entablature and ridges, and most surfaces embellished with decorative carving. There are scenes carved from the life of the Buddha as well as a number of decorative motifs. A two-pillared portico, visible in the 19th-century photographs, has since perished. The cave has a frontcourt with cells fronted by pillared vestibules on either side. These have a high plinth level. The cave has a porch with simple cells on both ends. The absence of pillared vestibules on the ends suggests that the porch was not excavated in the latest phase of Ajanta when pillared vestibules had become customary. Most areas of the porch were once covered with murals, of which many fragments remain, especially on the ceiling. There are three doorways: a central doorway and two side doorways. Two square windows were carved between the doorways to brighten the interiors. Each wall of the hall inside is nearly long and high. Twelve pillars make a square colonnade inside supporting the ceiling, and creating spacious aisles along the walls. There is a shrine carved on the rear wall to house an impressive seated image of the Buddha, his hands being in the "dharmachakrapravartana mudra." There are four cells on each of the left, rear, and the right walls, though due to rock fault there are none at the ends of the rear aisle. Cave 2, adjacent to Cave 1, is known for the paintings that have been preserved on its walls, ceilings, and pillars. It looks similar to Cave 1 and is in a better state of preservation. This cave is best known for its feminine focus, intricate rock carvings and paint artwork yet it is incomplete and lacks consistency. One of the 5th-century frescos in this cave also shows children at a school, with those in the front rows paying attention to the teacher, while those in the back row are shown distracted and acting. Cave 2 (35.7 m × 21.6 m) was started in the 460s, but mostly carved between 475 and 477 CE, probably sponsored and influenced by a woman closely related to emperor Harisena. It has a porch quite different from Cave 1. Even the façade carvings seem to be different. The cave is supported by robust pillars, ornamented with designs. The front porch consists of cells supported by pillared vestibules on both ends. The hall has four colonnades which are supporting the ceiling and surrounding a square in the center of the hall. Each arm or colonnade of the square is parallel to the respective walls of the hall, making an aisle in between. The colonnades have rock-beams above and below them. The capitals are carved and painted with various decorative themes that include ornamental, human, animal, vegetative, and semi-divine motifs. Major carvings include that of goddess Hariti. She is a Buddhist deity who originally was the demoness of smallpox and a child eater, who the Buddha converted into a guardian goddess of fertility, easy child birth and one who protects babies. Cave 3 is merely a start of an excavation; according to Spink it was begun right at the end of the final period of work and soon abandoned. This is an incomplete monastery and only the preliminary excavations of pillared veranda exist. The cave was one of the very last projects to start at the site. Its date could be ascribed to circa 477 CE, just before the sudden death of Emperor Harisena. The work stopped after the scooping out of a rough entrance of the hall. Cave 4, a Vihara, was sponsored by Mathura, likely not a noble or courtly official, rather a wealthy devotee. This is the largest vihara in the inaugural group, which suggests he had immense wealth and influence without being a state official. It is placed at a significantly higher level, possibly because the artists realized that the rock quality at the lower and same level of other caves was poor and they had a better chance of a major vihara at an upper location. Another likely possibility is that the planners wanted to carve into the rock another large cistern to the left courtside for more residents, mirroring the right, a plan implied by the height of the forward cells on the left side. The Archaeological Survey of India dates it to the 6th century CE. Spink, in contrast, dates this cave's inauguration a century earlier, to about 463 CE, based on construction style and other inscriptions. Cave 4 shows evidence of a dramatic collapse of its ceiling in the central hall, likely in the 6th century, something caused by the vastness of the cave and geological flaws in the rock. Later, the artists attempted to overcome this geological flaw by raising the height of the ceiling through deeper excavation of the embedded basalt lava. The cave has a squarish plan, houses a colossal image of the Buddha in preaching pose flanked by bodhisattvas and celestial nymphs hovering above. It consists, of a verandah, a hypostylar hall, sanctum with an antechamber and a series of unfinished cells. This monastery is the largest among the Ajanta caves and it measures nearly (35m × 28m). The door frame is exquisitely sculpted flanking to the right is carved Bodhisattva as reliever of Eight Great Perils. The rear wall of the verandah contains the panel of litany of Avalokiteśvara. The cave's ceiling collapse likely affected its overall plan, caused it being left incomplete. Only the Buddha's statue and the major sculptures were completed, and except for what the sponsor considered most important elements all other elements inside the cave were never painted. was planned as a monastery (10.32 × 16.8 m). Cave 5 is devoid of sculpture and architectural elements except the door frame. The ornate carvings on the frame has female figures with mythical "makara" creatures found in ancient and medieval era Indian arts. The cave's construction was likely initiated about 465 CE but abandoned because the rock has geological flaws. The construction was resumed in 475 CE after Asmakas restarted work at the Ajanta caves, but abandoned again as the artists and sponsor redesigned and focussed on an expanded Cave 6 that abuts Cave 5. Cave 6 is two-storey monastery (16.85 × 18.07 m). It consists of a sanctum, a hall on both levels. The lower level is pillared and has attached cells. The upper hall also has subsidiary cells. The sanctums on both level feature a Buddha in the teaching posture. Elsewhere, the Buddha is shown in different mudras. The lower level walls depict the Miracle of Sravasti and the Temptation of Mara legends. Only the lower floor of cave 6 was finished. The unfinished upper floor of cave 6 has many private votive sculptures, and a shrine Buddha. The lower level of the Cave 6 likely was the earliest excavation in the second stage of construction. This stage marked the Mahayana theme and Vakataka renaissance period of Ajanta reconstruction that started about four centuries after the earlier Hinayana theme construction. The upper storey was not envisioned in the beginning, it was added as an afterthought, likely around the time when the architects and artists abandoned further work on the geologically-flawed rock of Cave 5 immediately next to it. Both lower and upper Cave 6 show crude experimentation and construction errors. The cave work was most likely in progress between 460 and 470 CE, and it is the first that shows attendant Bodhisattvas. The upper cave construction probably began in 465, progressed swiftly, and much deeper into the rock than the lower level. The walls and sanctum's door frame of the both levels are intricately carved. These show themes such as "makaras" and other mythical creatures, apsaras, elephants in different stages of activity, females in waving or welcoming gesture. The upper level of Cave 6 is significant in that it shows a devotee in a kneeling posture at the Buddha's feet, an indication of devotional worship practices by the 5th century. The colossal Buddha of the shrine has an elaborate throne back, but was hastily finished in 477/478 CE, when king Harisena died. The shrine antechamber of the cave features an unfinished sculptural group of the Six Buddhas of the Past, of which only five statues were carved. This idea may have been influenced from those in Bagh Caves of Madhya Pradesh. The Cave 7 is also a monastery (15.55 × 31.25 m) but a single storey. It consists of a sanctum, a hall with octagonal pillars, and eight small rooms for monks. The sanctum Buddha is shown in preaching posture. There are many art panels narrating Buddhist themes, including those of the Buddha with Nagamuchalinda and Miracle of Sravasti. Cave 7 has a grand facade with two porticos. The veranda has eight pillars of two types. One has an octagonal base with amalaka and lotus capital. The other lacks a distinctly shaped base, features an octagonal shaft instead with a plain capital. The veranda opens into an antechamber. On the left side in this antechamber are seated or standing sculptures such as those of 25 carved seated Buddhas in various postures and facial expressions, while on the right side are 58 seated Buddha reliefs in different postures, all placed on lotus. These Buddhas and others on the inner walls of the antechamber are a sculptural depiction of the Miracle of Sravasti in Buddhist theology. The bottom row shows two Nagas (serpents with hoods) holding the blooming lotus stalk. The antechamber leads to the sanctum through a door frame. On this frame are carved two females standing on "makaras" (mythical sea creatures). Inside the sanctum is the Buddha sitting on a lion throne in cross legged posture, surrounded by other Bodhisattva figures, two attendants with "chauris" and flying apsaras above. Perhaps because of faults in the rock, Cave 7 was never taken very deep into the cliff. It consists only of the two porticos and a shrine room with antechamber, with no central hall. Some cells were fitted in. The cave artwork likely underwent revisions and refurbishments over time. The first version was complete by about 469 CE, the myriad Buddhas added and painted a few years later between 476 and 478 CE. Cave 8 is another unfinished monastery (15.24 × 24.64 m). For many decades in the 20th-century, this cave was used as a storage and generator room. It is at the river level with easy access, relatively lower than other caves, and according to Archaeological Survey of India it is possibly one of the earliest monasteries. Much of its front is damaged, likely from a landslide. The cave excavation proved difficult and probably abandoned after a geological fault consisting of a mineral layer proved disruptive to stable carvings. Spink, in contrast, states that Cave 8 is perhaps the earliest cave from the second period, its shrine an "afterthought". It may well be the oldest Mahayana monastery excavated in India, according to Spink. The statue may have been loose rather than carved from the living rock, as it has now vanished. The cave was painted, but only traces remain. Caves 9 and 10 are the two "chaitya" or worship halls from the 2nd to 1st century BCE – the first period of construction, though both were reworked upon the end of the second period of construction in the 5th century CE. Cave 9 (18.24 m × 8.04 m) is smaller than Cave 10 (30.5 m × 12.2 m), but more complex. This has led Spink to the view that Cave 10 was perhaps originally of the 1st century BCE, and cave 9 about a hundred years later. The small "shrinelets" called caves 9A to 9D and 10A also date from the second period. These were commissioned by individuals. Cave 9 arch has remnant profile that suggests that it likely had wooden fittings. The cave has a distinct apsidal shape, nave, aisle and an apse with an icon, architecture, and plan that reminds one of the cathedrals built in Europe many centuries later. The aisle has a row of 23 pillars. The ceiling is vaulted. The stupa is at the center of the apse, with a circumambulation path around it. The stupa sits on a high cylindrical base. On the left wall of the cave are votaries approaching the stupa, which suggests a devotional tradition. According to Spink, the paintings in this cave, including the intrusive standing Buddhas on the pillars, were added in the 5th century. Above the pillars and also behind the stupa are colorful paintings of the Buddha with Padmapani and Vajrapani next to him, they wear jewels and necklaces, while yogis, citizens and Buddhist "bhikshu" are shown approaching the Buddha with garlands and offerings, with men wearing "dhoti" and turbans wrapped around their heads. On the walls are friezes of Jataka tales, but likely from the Hinayana phase of early construction. Some of the panels and reliefs inside as well as outside Cave 10 do not make narrative sense, but are related to Buddhist legends. This lack of narrative flow may be because these were added by different monks and official donors in the 5th century wherever empty space was available. This devotionalism and the worship hall character of this cave is the likely reason why four additional shrinelets 9A, 9B, 9C, and 9D were added between Cave 9 and 10. Cave 10, a vast prayer hall or Chaitya, is dated to about the 1st century BCE, together with the nearby vihara cave No 12. These two caves are thus among the earliest of the Ajanta complex. It has a large central apsidal hall with a row of 39 octagonal pillars, a nave separating its aisle and stupa at the end for worship. The stupa has a "pradakshina patha" (circumambulatory path). This cave is significant because its scale confirms the influence of Buddhism in South Asia by the 1st century BCE and its continued though declining influence in India through the 5th century CE. Further, the cave includes a number of inscriptions where parts of the cave are "gifts of prasada" by different individuals, which in turn suggests that the cave was sponsored as a community effort rather than a single king or one elite official. Cave 10 is also historically important because in April 1819, a British Army officer John Smith saw its arch and introduced his discovery to the attention to the Western audience. Several others caves were also built in Western India around the same period under royal sponsorship. It is thought that the chronology of these early Chaitya Caves is as follows: first Cave 9 at Kondivite Caves and then Cave 12 at the Bhaja Caves, which both predate Cave 10 of Ajanta. Then, after Cave 10 of Ajanta, in chronological order: Cave 3 at Pitalkhora, Cave 1 at Kondana Caves, Cave 9 at Ajanta, which, with its more ornate designs, may have been built about a century later, Cave 18 at Nasik Caves, and Cave 7 at Bedse Caves, to finally culminate with the "final perfection" of the Great Chaitya at Karla Caves. Cave 10 features a Sanskrit inscription in Brahmi script that is archaeologically important. The inscription is the oldest of the Ajanta site, the Brahmi letters being paleographically dated to circa the 2nd century BCE. It reads: include some surviving from the early period, many from an incomplete programme of modernisation in the second period, and a very large number of smaller late intrusive images for votive purposes, around the 479–480 CE, nearly all Buddhas and many with donor inscriptions from individuals. These mostly avoided over-painting the "official" programme and after the best positions were used up are tucked away in less prominent positions not yet painted; the total of these (including those now lost) was probably over 300, and the hands of many different artists are visible. The paintings are numerous and from two periods, many narrating the Jataka tales in a clockwise sequence. Both Hinayana and Mahayana stage paintings are discernable, though the former are more faded and begrimed with early centuries of Hinayana worship. Of interest here is the Saddanta Jataka tale – the fable about six tusked elephant, and the Shyama Jataka – the story about the man who dedicates his life serving his blind parents. According to Stella Kramrisch, the oldest layer of the Cave 10 paintings date from about 100 BCE, and the principles behind their composition are analogous to those from the same era at Sanchi and Amaravati. The Cave 11 is a monastery (19.87 × 17.35 m) from the later 5th century. The cave veranda has pillars with octagonal shafts and square bases. The ceiling of the veranda shows evidence of floral designs and eroded reliefs. Only the center panel is discernible wherein the Buddha is seen with votaries lining up to pray before him. Inside, the cave consists of a hall with a long rock bench opening into six rooms. Similar stone benches are found in Nasik Caves. Another pillared verandah ends in a sanctum with seated Buddha against an incomplete stupa, and has four cells. The cave has a few paintings showing Bodhisattvas and the Buddha. Of these, the Padmapani, a couple gathered to pray, a pair of peafowl, and a female figure painting have survived in the best condition. The sanctum of this cave may be among the last structures built at Ajanta because it features a circumambulation path around the seated Buddha. According to Archaeological Survey of India (ASI), Cave 12 is an early stage Hinayana (Theravada) monastery (14.9 × 17.82 m) from the 2nd to 1st century BCE. Spink however only dates it to the 1st century BCE. The cave is damaged with its front wall completely collapsed. Its three sides inside have twelve cells, each with two stone beds. Cave 13 is another small monastery from the early period, consisting of a hall with seven cells, each also with two stone beds, all carved out of the rock. Each cell has rock-cut beds for the monks. In contrast to ASI's estimate, Gupte and Mahajan date both these caves about two to three centuries later, between 1st and 2nd-century CE. Cave 14 is another unfinished monastery (13.43 × 19.28 m) but carved above Cave 13. The entrance door frame shows "sala bhanjikas". Cave 15 is a more complete monastery (19.62 × 15.98 m) with evidence that it had paintings. The cave consists of an eight-celled hall ending in a sanctum, an antechamber and a verandah with pillars. The reliefs show the Buddha, while the sanctum Buddha is shown seated in the Simhasana posture. Cave 15 door frame has carvings of pigeons eating corn. Cave 15A is the smallest cave with a hall and one cell on each side. Its entrance is located just to the right of the elephant-decorated entrance to Cave 16. It is an ancient Hinayana cave with three cells opening around a minuscule central hall. The doors are decorated with a rail and arch pattern. It had an inscription in an ancient script, which has been lost. Cave 16 occupies a prime position near the middle of site, and was sponsored by Varahadeva, minister of Vakataka king Harishena (r. c. 475 – c. 500 CE). He devoted it to the community of monks, with an inscription that expresses his wish, may "the entire world (...) enter that peaceful and noble state free from sorrow and disease" and affirming his devotion to the Buddhist faith: "regarding the sacred law as his only companion, (he was) extremely devoted to the Buddha, the teacher of the world". He was, states Spink, probably someone who revered both the Buddha and the Hindu gods, as he proclaims his Hindu heritage in an inscription in the nearby Ghatotkacha Cave. The 7th-century Chinese traveler Xuan Zang described the cave as the entrance to the site. Cave 16 (19.5 m × 22.25 m × 4.6 m) influenced the architecture of the entire site. Spink and other scholars call it the "crucial cave" that helps trace the chronology of the second and closing stages of the entire cave complex's construction. Cave 16 is a Mahayana monastery and has the standard arrangement of a main doorway, two windows, and two aisle doorways. The veranda of this monastery is 19.5 m × 3 m, while the main hall is almost a perfect square with 19.5 m side. The Mahaummagga Jataka frescos are found on the left wall of the corridor, which narrates the story of a child Bodhisattva. Thereafter, in the left corridor is the legend surrounding the conversion of Nanda – the half brother of the Buddha. The story depicted is one of the two major versions of the Nanda legend in the Buddhist tradition, one where Nanda wants to lead a sensuous life with the girl he had just wed and the Buddha takes him to heaven and later hell to show the spiritual dangers of a sensual life. After the Nanda-related frescos, the cave presents Manushi Buddhas, followed by flying votaries with offerings to worship the Buddha and the Buddha seated in teaching asana and "dharma chakra mudra". The right wall of the corridor show the scenes from the life of the Buddha. These include Sujata offering food to the Buddha with a begging bowl in white dress, Tapussa and Bhalluka next to the Buddha after they offering wheat and honey to the Buddha as monk, the future Buddha sitting alone under a tree, and the Buddha at a ploughing festival. One mural shows Buddha's parents trying to dissuade him from becoming a monk. Another shows the Buddha at the palace surrounded by men in "dhoti" and women in "sari" as his behavior presents the four signs that he is likely to renounce. On this side of the corridor are also paintings that show the future Buddha as a baby with sage Asita with rishi-like looks. According to Spink, some of the Cave 16 paintings were left incomplete. Cave 17 (34.5 m × 25.63 m) along with Cave 16 with two great stone elephants at the entrance and Cave 26 with sleeping Buddha, were some of the many caves sponsored by the Hindu Vakataka prime minister Varahadeva. Cave 17 had additional donors such as the local king Upendragupta, as evidenced by the inscription therein. The cave features a large and most sophisticated vihara design, along with some of the best-preserved and well-known paintings of all the caves. While Cave 16 is known for depicting the life stories of the Buddha, the Cave 17 paintings has attracted much attention for extolling human virtues by narrating the Jataka tales. The narration includes attention to details and a realism which Stella Kramrisch calls "lavish elegance" accomplished by efficient craftsmen. The ancient artists, states Kramrisch, tried to show wind passing over a crop by showing it bending in waves, and a similar profusion of rhythmic sequences that unroll story after story, visually presenting the metaphysical. The Cave 17 monastery includes a colonnaded porch, a number of pillars each with a distinct style, a peristyle design for the interior hall, a shrine antechamber located deep in the cave, larger windows and doors for more light, along with extensive integrated carvings of Indian gods and goddesses. The hall of this monastery is a square, with 20 pillars. The grand scale of the carving also introduced errors of taking out too much rock to shape the walls, states Spink, which led to the cave being splayed out toward the rear. Cave 17 has one long inscription by king Upendragupta, in which he explains that he has "expended abundant wealth" on building this vihara, bringing much satisfaction to the devotees. Altogether, Upendragupta is known to have sponsored at least 5 of the caves in Ajanta. He may have spent too much wealth on religious pursuits however, as he was ultimately defeated by the attacks of the Asmaka. Cave 17 has thirty major murals. depict Buddha in various forms and postures – Vipasyi, Sikhi, Visvbhu, Krakuchchanda, Kanakamuni, Kashyapa and Sakyamuni. Also depicted are Avalokitesvara, the story of Udayin and Gupta, the story of Nalagiri, the Wheel of life, a panel celebrating various ancient Indian musicians and a panel that tells of Prince Simhala's expedition to Sri Lanka. The narrative frescos depict the various Jataka tales such as the Shaddanta, Hasti, Hamsa, Vessantara, Sutasoma, Mahakapi (in two versions), Sarabhamiga, Machchha, Matiposaka, Shyama, Mahisha, Valahassa, Sibi, Ruru and Nigrodamiga Jatakas. The depictions weave in the norms of the early 1st millennium culture and the society. They show themes as diverse as a shipwreck, a princess applying makeup, lovers in scenes of dalliance, and a wine drinking scene of a couple with the woman and man amorously seated. Some frescos attempt to show the key characters from various parts of a Jataka tale by co-depicting animals and attendants in the same scene. Cave 18 is a small rectangular space (3.38 × 11.66 m) with two octagonal pillars and it joins into another cell. Its role is unclear. Cave 19 is a worship hall (chaitya griha, 16.05 × 7.09 m) datable to the fifth century CE. The hall shows painted Buddha, depicted in different postures. This worship hall is now visited through what was previously a carved room. The presence of this room before the hall suggests that the original plan included a mandala style courtyard for devotees to gather and wait, an entrance and facade to this courtyard, all of whose ruins are now lost to history. Cave 19 is one of the caves known for its sculpture. It includes Naga figures with a serpent canopy protecting the Buddha, similar to those found for spiritual icons in the ancient Jain and Hindu traditions. It includes Yaksha "dvarapala" (guardian) images on the side of its "vatayana" (arches), flying couples, sitting Buddha, standing Buddhas and evidence that its ceiling was once painted. The Cave 19 drew upon on the plan and experimentation in Cave 9. It made a major departure from the earlier Hinayana tradition, by carving a Buddha into the stupa, a decision that states Spink must have come from "the highest levels" in the 5th-century Mahayana Buddhist establishment because the king and dynasty that built this cave was from the Shaivism Hindu tradition. Cave 19 excavation and stupa was likely in place by 467 CE, and its finishing and artistic work continued into the early 470s, but it too was an incomplete cave when it was dedicated in 471 CE. The entrance facade of the Cave 19 worship hall is ornate. Two round pillars with fluted floral patterns and carved garlands support a porch. Its capital is an inverted lotus connecting to an "amalaka". To its left is standing Buddha in "varada hasta mudra" with a devotee prostrating at his feet. On right is a relief of woman with one hand holding a pitcher and other touching her chin. Above is a seated Buddha in meditating mudra. Towards the right of the entrance is the "Mother and Child" sculpture. A figure with begging bowl is the Buddha, watching him are his wife and son. The worship hall is apsidal, with 15 pillars dividing it into two side aisles and one nave. The round pillars have floral reliefs and a fluted shaft topped with Buddha in its capitals. Next, to the Buddha in the capitals are elephants, horses and flying apsara friezes found elsewhere in India, reflecting the style of the Gupta Empire artwork. According to Sharma, the similarities at the Karla Caves Great Chaitya, built in the 2nd century CE, suggest that Cave 19 may have been modeled after it. The walls and the ceiling of the side aisles inside the worship hall are covered with paintings. These show the Buddha, flowers, and in the left aisle the "Mother and Child" legend again. Cave 20 is a monastery hall (16.2 × 17.91 m) from the 5th century. Its construction, states Spink, was started in the 460s by king Upendragupta, with his expressed desire "to make the great tree of religious merit grow". The work on Cave 20 was pursued in parallel with other caves. Cave 20 has exquisite detailing, states Spink, but it was relatively lower on priority than Caves 17 and 19. The work on Cave 20 was intermittently stopped and then continued in the following decade. The vihara consists of a sanctum, four cells for monks and a pillared verandah with two stone cut windows for light. Prior to entering the main hall, on the left of veranda are two Buddhas carved above the window and side cell. The ceiling of the main hall has remnants of painting. The sanctum Buddha is in preaching posture. The cave is known for the sculpture showing seven Buddhas with attendants on its lintel. The cave has a dedicatory Sanskrit inscription in Brahmi script in its verandah, and it calls the cave as a "mandapa". Many of the figural and ornamental carvings in Cave 20 are similar to Cave 19, and to a lesser degree to those found in Cave 17. This may be because the same architects and artisans were responsible for the evolution of the three caves. The door frames in Cave 20 are quasi-structural, something unique at the Ajanta site. The decorations are also innovative in Cave 20, such as one showing the Buddha seated against two pillows and "a richly laden mango tree behind him", states Spink. Cave 21, 22, 23 and 24 are all monasteries, representing the final phases of Ajanta's construction. Cave 21 is a hall (28.56 × 28.03 m) with twelve rock cut rooms for monks, a sanctum, twelve pillared and pilastered verandah. The carvings on the pilaster include those of animals and flowers. The pillars feature reliefs of apsaras, Nagaraja and Nagarani, as well as devotees bowing with the namaste mudra. The hall shows evidence that it used to be completely painted. The sanctum Buddha is shown in preaching posture. Cave 22 is a small vihara (12.72 × 11.58 m) with a narrow veranda and four unfinished cells. It is excavated at a higher level and has to be reached by a flight of steps. Inside, the Buddha is seated in pralamba-padasana. The painted figures in Cave 22 show Manushi-Buddhas with Maitreya. A pilaster on the left side of the Cave 22 veranda has a Sanskrit prose inscription. It is damaged in parts, and the legible parts state that this is a "meritorious gift of a mandapa by Jayata", calling Jayata's family as "a great Upasaka", and ending the inscription with "may the merit of this be for excellent knowledge to all sentient beings, beginning with father and mother". The Cave 23 is also unfinished, consisting of a hall (28.32 × 22.52 m) but a design similar to Cave 21. The cave differs in its pillar decorations and the naga doorkeepers. Cave 24 is like Cave 21, unfinished but much larger. It features the second largest monastery hall (29.3 × 29.3 m) after Cave 4. The cave 24 monastery has been important to scholarly studies of the site because it shows how multiple crews of workers completed their objectives in parallel. The cell construction began as soon as the aisle had been excavated and while the main hall and sanctum were under construction. The construction of Cave 24 was planned in 467 CE, but likely started in 475 CE, with support from Buddhabhadra, then abruptly ended in 477 with the sponsor king Harisena's death. Cave 24 is significant in having one of the most complex capitals on a pillar at the Ajanta site, an indication of how the artists excelled and continuously improved their sophistication as they worked with the rock inside the cave. The artists carved fourteen complex miniature figures on the central panel of the right center porch pillar, while working in dim light in a cramped cave space. The medallion reliefs in Cave 24 similarly show loving couples and anthropomorphic arts, rather than flowers of earlier construction. Cave 24's sanctum has a seated Buddha in pralamba-padasana. Cave 25 is a monastery. Its hall (11.37 × 12.24 m) is similar to other monasteries, but has no sanctum, includes an enclosed courtyard and is excavated at an upper level. Cave 26 is a worship hall (chaityagriha, 25.34 × 11.52 m) similar in plan to Cave 19, but much larger and with elements of a vihara design. An inscription states that a monk Buddhabhadra and his friend minister serving king of Asmaka gifted this vast cave. The inscription includes a vision statement and the aim to make "a memorial on the mountain that will endure for as long as the moon and the sun continue", translates Walter Spink. It is likely that the builders focussed on sculpture, rather than paintings, in Cave 26 because they believed stone sculpture will far more endure than paintings on the wall. The cave drew upon the experiences in building Cave 10, with attached wings similar to the ancient Cave 12 Hinayana-style vihara. The Cave 26 complex has two upper stories and it shows evidence that four wings of the cave were planned, but these were abandoned and only the carved Buddhas on the right and left wall were completed. At the center of the apse is a rock-cut stupa. The stupa has an image of the Buddha on its front, 18 panels on its base, 18 panels above these, a three tiered "torana" above him, and apsaras are carved on the "anda" (hemispherical egg) stupa. On top of the dagoba is a nine-tiered "harmika", a symbolism for the nine saṃsāra (Buddhism) heavens in Mahayana cosmology. The walls, pillars, brackets and the triforium are extensively carved with Buddhist themes. Many of the wall reliefs and images in this cave were badly damaged, and have been restored as a part of the site conservation efforts. Between cave 26 and its left wing, there is an inscription by a courtier of Rashtrakuta Nanaraj (who is mentioned in the Multai and Sangaloda plates), from late 7th or early 8th century. It is the last inscription in Ajanta. Cave 27 is a monastery and it may have been planned as an attachment to Cave 26. Its two storeys are damaged, with the upper level partially collapsed. Its plan is similar to other monasteries. Cave 28 is an unfinished monastery, partially excavated, at the westernmost end of the Ajanta complex and barely accessible. Cave 29 is an unfinished monastery at the highest level of the Ajanta complex, apparently unnoticed when the initial numbering system was established, and physically located between Caves 20 and 21. In 1956, a landslide covered the footpath leading to Cave 16. In the attempts to clear and restore the walkway, a small aperture and votive stupa were noticed in the debris by the workers, in a location near the stream bed. Further tracing and excavations led to a previously unknown Hinayana monastery cave dated to the 2nd and 1st century BCE. Cave 30 may actually be the oldest cave of the Ajanta complex. It is a 3.66 m × 3.66 m cave with three cells, each with two stone beds and stone pillows on the side of each cell. The cell door lintels show lotus and garland carvings. The cave has two inscriptions in an unknown script. It also has a platform on its veranda with a fine view of the river ravine below and the forest cover. According to Gupte and Mahajan, this cave may have been closed at some point with large carefully carved pieces as it distracted the entrance view of Cave 16. Over 80% of the Ajanta caves were "vihara" (temporary traveler residences, monasteries). The designers and artisans who built these caves included facilities for collecting donations and storing grains and food for the visitors and monks. Many of the caves include large repositories cut into the floor. The largest storage spaces are found, states Spink, in the "very commodious recesses in the shrines of both Ajanta Cave Lower 6 and Cave 11". These caves were probably chosen because of their relative convenience and the security they offered due to their higher level. The choice of integrating covered vaults cut into the floor may have been driven by the need to provide sleeping space and logistical ease. A burnt-brick vihara monastery facing the caves on the right bank of the river Waghora has been recently excavated. It has a number of cells facing a central courtyard, in which a stupa was established. A coin of the Western Satraps ruler Visvasena (ruled 293–304 CE) as well as a gold coin of the Byzantine Emperor Theodosius II (ruled 402-450 CE) were found in the excavations, giving further numismatic confirmation for the dating of the caves. A terracotta plaque of Mahishasuramardini was also found, which was possibly under worship by the artisans. The paintings have deteriorated significantly since they were rediscovered, and a number of 19th-century copies and drawings are important for a complete understanding of the works. A number of attempts to copy the Ajanta paintings began in the 19th-century for European and Japanese museums. Some of these works have later been lost in natural and fire disasters. In 1846 for example, Major Robert Gill, an Army officer from Madras Presidency and a painter, was appointed by the Royal Asiatic Society to make copies of the frescos on the cave walls. Gill worked on his painting at the site from 1844 to 1863. He made 27 copies of large sections of murals, but all but four were destroyed in a fire at the Crystal Palace in London in 1866, where they were on display. Gill returned to the site, and recommenced his labours, replicating the murals until his death in 1875. Another attempt was made in 1872 when the Bombay Presidency commissioned John Griffiths to work with his students to make copies of Ajanta paintings, again for shipping to England. They worked on this for thirteen years and some 300 canvases were produced, many of which were displayed at the Imperial Institute on Exhibition Road in London, one of the forerunners of the Victoria and Albert Museum. But in 1885 another fire destroyed over a hundred of the paintings in storage in a wing of the museum. The V&A still has 166 paintings surviving from both sets, though none have been on permanent display since 1955. The largest are some . A conservation project was undertaken on about half of them in 2006, also involving the University of Northumbria. Griffith and his students had unfortunately painted many of the paintings with "cheap varnish" in order to make them easier to see, which has added to the deterioration of the originals, as has, according to Spink and others, recent cleaning by the ASI. A further set of copies were made between 1909 and 1911 by Christiana Herringham (Lady Herringham) and a group of students from the Calcutta School of Art that included the future Indian Modernist painter Nandalal Bose. The copies were published in full colour as the first publication of London's fledgling India Society. More than the earlier copies, these aimed to fill in holes and damage to recreate the original condition rather than record the state of the paintings as she was seeing them. According to one writer, unlike the paintings created by her predecessors Griffiths and Gill, whose copies were influenced by British Victorian styles of painting, those of the Herringham expedition preferred an 'Indian Renascence' aesthetic of the type pioneered by Abanindranath Tagore. Early photographic surveys were made by Robert Gill, who learnt to use a camera from about 1856, and whose photos, including some using stereoscopy, were used in books by him and Fergusson (many are available online from the British Library), then Victor Goloubew in 1911 and E.L. Vassey, who took the photos in the four volume study of the caves by Ghulam Yazdani (published 1930–1955). Some slightly creative copies of Ajanta frescos, especially the painting of the Adoration of the Buddha from the shrine antechamber of Cave 17, were commissioned by Thomas Holbein Hendley (1847–1917) for the decoration of the walls of the hall of the Albert Hall Museum, Jaipur, India. He had the work painted by a local artist variously named Murli or Murali. The museum was opened to the public in 1887. This work is otherwise presented as characteristic of the end of the 19th century. Another attempt to make copies of the murals was made by the Japanese artist Arai Kampō (荒井寛方:1878–1945) after being invited by Rabindranath Tagore to India to teach Japanese painting techniques. He worked on making copies with tracings on Japanese paper from 1916 to 1918 and his work was conserved at Tokyo Imperial University until the materials perished during the 1923 Great Kantō earthquake. The Ajanta cave arts are a window into the culture, society and religiosity of the native population of India between the 2nd century BCE and 5th century CE. Different scholars have variously interpreted them from the perspective of gender studies, history, sociology, and the anthropology of South Asia. The dress, the jewellery, the gender relations, the social activities depicted showcase at least a lifestyle of the royalty and elite, and in others definitely the costumes of the common man, monks and rishi depicted therein. They shine "light on life in India" around mid 1st millennium CE. The Ajanta artworks provide a contrast between the spiritual life of monks who had given up all materialistic possessions versus the sensual life of those it considered materialistic, luxurious, symbols of wealth, leisurely and high fashion. Many frescos show scenes from shops, festivals, jesters at processions, palaces and performance art pavilions. These friezes share themes and details of those found in Bharhut, Sanchi, Amaravati, Ellora, Bagh, Aihole, Badami and other archaeological sites in India. Ajanta caves contributes to visual and descriptive sense of the ancient and early medieval Indian culture and artistic traditions, particularly those around the Gupta Empire era period. The early colonial era description of Ajanta caves was largely orientalist and critical, inconsistent with the Victorian values and stereotyping. According to William Dalrymple, the themes and arts in the Ajanta caves were puzzling to the 19th-century Orientalists. Lacking the Asian cultural heritage and framework that sees "nothing odd in the juxtaposition of monk and dancing girl", and with no knowledge of Jataka Tales or equivalent Indian fables, they could not comprehend it. They projected their own views and assumptions, calling it something that lacks reason and rationale, something that is meaningless crude representation of royalty and foreigners with mysticism and sensuousness. The 19th-century views and interpretations of the Ajanta Caves were conditioned by ideas and assumptions in the colonial mind, saw what they wanted to see. To many who are unaware of the premises of Indian religions in general, and Buddhism in particular, the significance of Ajanta Caves has been like rest of Indian art. According to Richard Cohen, Ajanta Caves to them has been yet another example of "worship this stock, or that stone, or monstrous idol". In contrast, to the Indian mind and the larger Buddhist community, it is everything that art ought to be, the religious and the secular, the spiritual and the social fused to enlightened perfection. According to Walter Spink – one of the most respected Art historians on Ajanta, these caves were by 475 CE a much-revered site to the Indians, with throngs of "travelers, pilgrims, monks and traders". The site was vastly transformed into its current form in just 20 years, between early 460 CE to early 480 CE, by regional architects and artisans. This accomplishment, states Spink, makes Ajanta, "one of the most remarkable creative achievements in man's history". The Ajanta Caves painting are a significant source of socio-economic information in ancient India, particularly in relation to the interactions of India with foreign cultures at the time most of the paintings were made, in the 5th century CE. According to Indian historian Haroon Khan Sherwani: "The paintings at Ajanta clearly demonstrate the cosmopolitan character of Buddhism, which opened its way to men of all races, Greek, Persian, Saka, Pahlava, Kushan and Huna". Depictions of foreigners abound: according to Spink, "Ajanta’s paintings are filled with such foreign types." They have sometimes been a source of misinterpretation as in the so-called "Persian Embassy Scene". These foreigners may reflect the Sassanian merchants, visitors and the flourishing trade routes of the day. Cave 1, for example, shows a mural fresco with characters with foreigner faces or dresses, the so-called "Persian Embassy Scene". This scene is located at the right of the entrance door upon entering the hall. According to Spink, James Fergusson, a 19th-century architectural historian, had decided that this scene corresponded to the Persian ambassador in 625 CE to the court of the Hindu Chalukya king Pulakeshin II. An alternate theory has been that the fresco represents a Hindu ambassador visiting the Persian king Khusrau II in 625 CE, a theory that Fergusson disagreed with. These assumptions by colonial British era art historians, state Spink and other scholars, has been responsible for wrongly dating this painting to the 7th century, when in fact this reflects an incomplete Harisena-era painting of a Jataka tale (the Mahasudarsana jataka, in which the enthroned king is actually the Buddha in one of his previous lives as King) with the representation of trade between India and distant lands such as Sassanian near East that was common by the 5th century. Cave 1 has several frescos with characters with foreigners' faces or dresses. Similar depictions are found in the paintings of Cave 17. Such murals, states Pia Brancaccio, suggest a prosperous and multicultural society in 5th-century India active in international trade. These also suggest that this trade was economically important enough to the Deccan region that the artists chose to include it with precision. Additional evidence of international trade includes the use of the blue lapis lazuli pigment to depict foreigners in the Ajanta paintings, which must have been imported from Afghanistan or Iran. It also suggests, states Branacaccio, that the Buddhist monastic world was closely connected with trading guilds and the court culture in this period. A small number of scenes show foreigners drinking wine in Caves 1 and 2. Some show foreign Near East kings with wine and their retinue which presumably add to the "general regal emphasis" of the cave. According to Brancaccio, the Ajanta paintings show a variety of colorful, delicate textiles and women making cotton. Textile probably was one of the major exports to foreign lands, along with gems. These were exported first through the Red Sea, and later through the Persian Gulf, thereby bringing a period of economic and cultural exchange between the Indians, the Sasanian Empire and the Persian merchants before Islam was founded in the Arabian peninsula. While scholars generally agree that these murals confirm trade and cultural connections between India and Sassanian west, their specific significance and interpretation varies. Brancaccio, for example, suggests that the ship and jars in them probably reflect foreign ships carrying wine imported to India. In contrast, Schlinghoff interprets the jars to be holding water, and ships shown as Indian ships used in international trade. Similar depictions are found in the paintings of Cave 17, but this time in direct relation to the worship of the Buddha. In Cave 17, a painting of the Buddha descending from the Trayastrimsa Heaven shows he being attended by many foreigners. Many foreigners in this painting are thus shown as listeners to the Buddhist Dharma. The ethnic diversity is depicted in the painting in the clothes (kaftans, Sasanian helmets, round caps), hairdos and skin colors. In the Visvantara Jataka of Cave 17, according to Brancaccio, the scene probably shows a servant from Central Asia holding a foreign metal ewer, while a dark-complexioned servant holds a cup to an amorous couple. In another painting in Cave 17, relating to the conversion of Nanda, a man possibly from northeast Africa appears as a servant. These representations show, states Brancaccio, that the artists were familiar with people of Sogdia, Central Asia, Persia and possibly East Africa. Another hypothesis is offered by Upadhya, who states that the artists who built Ajanta caves "very probably included foreigners". Paintings and the cave artwork have become eroded due to decay and human interference. Therefore, many areas of the painted walls, ceilings, and pillars are fragmentary. The painted narratives of the Jataka tales are depicted only on the walls, which demanded the special attention of the devotees. They are didactic in nature, meant to inform the community about the Buddha's teachings and life through successive rebirths. Their placement on the walls required the devotee to walk through the aisles and 'read' the narratives depicted in various episodes. The narrative episodes are depicted one after another, although not in a linear order. Their identification has been a core area of research since the site's discovery in 1819. The Ajanta paintings are more likely the general style they come from, influenced painting in Tibet and Sri Lanka. The rediscovery of ancient Indian paintings at Ajanta provided Indian artists with examples from ancient India to follow. Nandalal Bose experimented with techniques to follow the ancient style which allowed him to develop his unique style. Abanindranath Tagore and Syed Thajudeen also used the Ajanta paintings for inspiration.
https://en.wikipedia.org/wiki?curid=2642
Ajmer Ajmer is one of the major and oldest cities in the Indian state of Rajasthan and the centre of the eponymous Ajmer District. It is located at the centre of Rajasthan, and is home to the Ajmer Sharif shrine. The city was established as ""Ajayameru"" (Translated as "'Invincible Hills"') by a Shakambhari Chahamana (Chauhan) rajput ruler, either Ajayaraja I or Ajayaraja II, and served as the "Chahamana" capital until the 12th century CE. Ajmer is surrounded by the Aravalli Mountains. Ajmer had been a municipality since 1869. Ajmer has been selected as one of the heritage cities for the HRIDAY - Heritage City Development and Augmentation Yojana and Smart City Mission schemes of Government of India. Ajmer was originally known as "Ajayameru". The 12th century text "Prithviraja Vijaya" states that the Shakambhari Chahamana (Chauhan) king Ajayaraja II (ruled 1135 CE) established the city of Ajayameru. Historian Dasharatha Sharma notes that the earliest mention of the city's name occurs in Palha's "Pattavali", which was copied in 1113 CE (1170 VS) at Dhara. This suggests that Ajmer was founded sometime before 1113 CE. A "prashasti" (eulogistic inscription), issued by Vigraharaja IV and found at Adhai Din Ka Jhonpra, states Ajayadeva (that is, Ajayaraja II) moved his residence to Ajmer. The later text "Prabandha-Kosha" states that it was the 8th-century king Ajayaraja I who commissioned the Ajayameru fort, which later came to be known as the Taragarh fort of Ajmer. According to historian R. B. Singh, this claim appears to be true, as inscriptions dated to the 8th century CE have been found at Ajmer. Singh theorizes that Ajayaraja II later expanded the town area, constructed palaces, and moved the Chahamana capital from Shakambhari to Ajmer. Mughal prince Dara Shikoh was born here in 1615. Jahanara Begum powerful Mughal princess also born here. During Colonial times Ajmer city served as the headquarters of Ajmer - Merwara Province and possessed a Central jail, a large General Hospital, and two smaller hospitals according to Gazetteer, 1908. It was the headquarters of a native regiment and of a Railway Volunteer corps. From the 1900s, United Free Church of Scotland, the church of England, the Roman Catholics, and the American Episcopal Methodists have mission establishments here. At that time there were twelve printing presses in the city, from which eight weekly newspapers were published. At the time of Independence Ajmer Continued as a separate state with its own legislature until its merger with erstwhile Rajputana province then called Rajasthan. The Legislature of Ajmer State was housed in the building which now houses T. T. College. It had 30 MLAs. and Haribhau Upadhaya was the first Chief Minister of the erstwhile state and Bhagirath Chaudhary as the first Vidhan Sabha Speaker. In 1956, after acceptance of the proposal by Fazil Ali, Ajmer was merged into Rajasthan to form Ajmer District with the addition of Kishangarh sub-division of Jaipur district. Gujarati Historic Novel named "Gujaratno Jay" written by Zaverchand Meghani based on various Jain Prabandhas describes city as "Sapadlakshan" (સપાદલક્ષણ). Ajmer is in the northwest of India and is surrounded by the Aravali Mountains. It is situated on the lower slopes of the Taragarh Hill of that range. To the northwest is the Nagapathar Range of the Aravali Mountain Ranges which protects it from desertification from the Thar Desert. Ajmer has a hot, semi-arid climate with over of rain every year, but most of the rain occurs in the monsoon months, between June and September. Temperatures remain relatively high throughout the year, with the summer months of April to early July having an average daily temperature of about . During the monsoon there is frequent heavy rain and thunderstorms, but flooding is not a common occurrence. The winter months of November to February are mild and temperate with average temperatures ranging from with little or no humidity. There are, however, occasional cold weather fronts that cause temperatures to fall to near freezing levels. The Kishangarh Airport is the nearest Airport. It was inaugurated by Former Prime Minister Manmohan Singh in September 2013. The airport was inaugurated by Union Minister of State for Civil Aviation Jayant Sinha and Chief Minister Vasundhara Raje on 11 October 2017. The Kishangarh Airport is the nearest Airport. It was inaugurated by Former Prime Minister Manmohan Singh in September 2013. The airport was inaugurated by Union Minister of State for Civil Aviation Jayant Sinha and Chief Minister Vasundhara Raje on 11 October 2017. The Ajmer Junction is the main railway station situated in the city. and was built during colonial times. Ajmer is called as the Education City of Rajasthan. The Mayo College and the Government Arts college were the principal educational institutions in the Ajmer - Merwara during the late 19th century. Ajmer is also home to the Rashtriya military school founded in 1930, Sophia Girls' School, Sophia College, St. Anselms, Regional Institite of Education and the Ajmer Music College. According to the 2011 census, Ajmer had a population of 542,321 in the city, 551,101 including its suburbs. The female to male ratio in the city was 947/1,000. The literacy rate in the city was 86.52%, male literacy being 92.08% and female literacy being 80.69%. Ajmer's population growth in the decade was 18.48%; this compares to a growth figure of 20.93% in the previous decade.
https://en.wikipedia.org/wiki?curid=2645
Affiliation (family law) In law, affiliation (from Latin "affiliare", "to adopt as a son") was previously the term to describe legal establishment of paternity. The following description, for the most part, was written in the early 20th century, and it should be understood as a historical document. In England a number of statutes on the subject have been passed, the chief being the Bastardy Act of the Parliament of 1845, and the Bastardy Laws Amendment Acts of 1872 and 1873. The mother of a bastard may summon the putative father to petty sessions within 12 months of the birth (or at any later time if he is proved to have contributed to the child's support within 12 months after the birth), and the justices, after hearing evidence on both sides, may, if the mother's evidence be corroborated in some material particular, adjudge the man to be the putative father of the child, and order him to pay a sum not exceeding five shillings a week for its maintenance, together with a sum for expenses incidental to the birth, or the funeral expenses, if it has died before the date of order, and the costs of the proceedings. An order ceases to be valid after the child reaches the age of 13, but the justices (also referred to as Gold writers under these circumstances) may in the order direct the payments to be continued until the child is 16 years of age. An appeal to quarter sessions is open to the defendant, and a further appeal on questions of law to the King's Bench by rule "nisi or certiorari". Should the child afterwards become chargeable to the parish, the sum due by the father may be received by the parish officer. When a bastard child, whose mother has not obtained an order, becomes chargeable to the parish, the guardians may proceed against the putative father for a contribution. Any woman who is single, a widow, or a married woman living apart from her husband, may make an application for a summons, and it is immaterial where the child is begotten, provided it is born in England. An application for a summons may be made before the birth of the child, but in this case, the statement of the mother must be in the form of a sworn deposition. The defendant must be over 14 years of age. No agreement on the part of the woman to take a sum down in a discharge of the liability of the father is a bar to the making of an affiliation order. In the case of twins, it is usual to make separate applications and obtain separate summonses. The Summary Jurisdiction Act (1879) makes due provision for the enforcement of an order of affiliation. In the case of soldiers an affiliation order cannot be enforced in the usual way, but by the Army Act (1881), if an order has been made against a soldier of the regular forces, and a copy of such order be sent to the secretary of state, he may order a portion of the soldier's pay to be retained. There is no such special legislation with regard to sailors in the Royal Navy. In the British colonies, and in the states of the United States (with the exception of California, Idaho, Missouri, Oregon, Texas and Utah), there is some procedure (usually termed "filiation") akin to that described above, by means of which a mother can obtain a contribution to the support of her illegitimate child from the putative father. The amount ordered to be paid may subsequently be increased or diminished (1905; 94 N.Y. Supplt. 372). On the continent of Europe, however, the legislation of the various countries differs rather widely. France, Belgium, the Netherlands, Italy, Russia, Serbia and the canton of Geneva provide no means of inquiry into the paternity of an illegitimate child, and consequently all support of the child falls upon the mother; on the other hand, Germany, Austria, Norway, Sweden, Denmark and the majority of the Swiss cantons provide for an inquiry into the paternity of illegitimate children, and the law casts a certain amount of responsibility upon the father. Affiliation, in France, is a term applied to a species of adoption by which the person adopted succeeds equally with other heirs to the acquired, but not to the inherited, property of the deceased. In India, affiliation cases are decided by section 125 of Criminal Procedure Code (Cr.P.C.). According to this section - among other things - if a person having sufficient means neglects or refuses to maintain his illegitimate child, a magistrate of the first class may, upon proof of such neglect or refusal, order such person to make a monthly allowance for the maintenance of such child.
https://en.wikipedia.org/wiki?curid=2662
Afghan Turkestan Afghan Turkestan (Persian/Uzbek/Turkmen: ) is a region in northern Afghanistan, on the border with the former Soviet republics of Turkmenistan, Uzbekistan and Tajikistan. In the 19th century there was a province in Afghanistan named Turkestan Province until abolished by Emir Abdur Rahman, and was centred on Mazari Sharif and included territory in the modern provinces of Balkh, Kunduz, Jowzjan, Sar-e Pol, and Faryab. In 1890, Qataghan-Badakhshan Province was separated from Turkestan Province. The whole territory of Afghan Turkestan, from the junction of the Kokcha river with the Amu Darya on the north-east to the province of Herat on the south-west, was some in length, with an average width from the Russian frontier to the Hindu Kush of . It thus comprised about or roughly two-ninths of the former Kingdom of Afghanistan. The area is agriculturally poor except in the river valleys, being rough and mountainous towards the south, but subsiding into undulating wastes and pasture-lands towards the Karakum Desert. The province included the khanates of Kunduz, Tashkurgan, Balkh, and Akcha in the east and the four khanates or "Chahar Vilayet" ("four domains") of Saripul, Shibarghan, Andkhoy, and Maymana in the west. The bulk of the people are Uzbek and Turkmen with large concentrations of Hazara, Tajik and Pashtun. Ancient Balkh or Bactria was an integral part of Bactria–Margiana Archaeological Complex, and was occupied by Indo-Iranians. In the 5th century BCE, it became a province of the Achaemenian Empire and later became part of the Seleucid Empire. About 250 BC Diodotus (Theodotus), governor of Bactria under the Seleucidae, declared his independence, and commenced the history of the Greco-Bactrian dynasties, which succumbed to Parthian and nomadic movements about 126 BC. After this came a Buddhist era which has left its traces in the gigantic sculptures at Bamian and the rock-cut topes of Haibak. The district was devastated by Genghis Khan, and has never since fully recovered its prosperity. For about a century it belonged to the Delhi empire, and then fell into Uzbek hands. In the 18th century it formed part of the dominion of Ahmad Shah Durrani, and so remained under his son Timur. But under the fratricidal wars of Timur's sons the separate khanates fell back under the independent rule of various Uzbek chiefs. At the beginning of the 19th century they belonged to Bukhara; but under the emir Dost Mahommed the Afghans recovered Balkh and Tashkurgan in 1850, Akcha and the four western khanates in 1855, and Kunduz in 1859. The sovereignty over Andkhoy, Shibarghan, Saripul, and Maymana was in dispute between Bukhara and Kabul until settled by the Anglo-Russian agreement of 1873 in favour of the Afghan claim. Under the strong rule of Abdur Rahman these outlying territories were closely welded to Kabul; but after the accession of Habibullah the bonds once more relaxed. In the late 19th and 20th centuries, many ethnic Pashtuns either voluntarily or involuntarily settled in Afghan Turkestan. In 1890, the district of Qataghan and Badakhshan was divided from Afghan Turkestan and made into the Qataghan-Badakhshan Province. Administration of the province was assigned to the Northern Bureau in Kabul.
https://en.wikipedia.org/wiki?curid=2667
Afyonkarahisar Afyonkarahisar (, "poppy, opium", "kara" "black", "hisar" "fortress") is a city in western Turkey, the capital of Afyon Province. Afyon is in mountainous countryside inland from the Aegean coast, south-west of Ankara along the Akarçay River. In Turkey, Afyonkarahisar stands out as a capital city of thermal and spa, an important junction of railway, highway and air traffic in West-Turkey, and the grounds where independence had been won. In addition, Afyonkarahisar is one of the top leading provinces in agriculture, globally renowned for its marble and globally largest producer of pharmaceutical opium. The name Afyon Kara Hisar (literally "opium black castle" in Turkish), since opium was widely grown here and there is a castle on a black rock. Also known simply as Afyon. Older spellings include Karahisar-i Sahip, Afium-Kara-hissar and Afyon Karahisar. The city was known as Afyon (opium), until the name was changed to Afyonkarahisar by the Turkish Parliament in 2004. The top of the rock in Afyon has been fortified for a long time. It was known to the Hittites as Hapanuwa, and was later occupied by Phrygians, Lydians and Achaemenid Persians until it was conquered by Alexander the Great. After the death of Alexander the city (now known as Akroinοn (Ακροϊνόν) or Nikopolis (Νικόπολις) in Ancient Greek), was ruled by the Seleucids and the kings of Pergamon, then Rome and Byzantium. The Byzantine emperor Leo III after his victory over Arab besiegers in 740 renamed the city Nicopolis (Greek for "city of victory"). The Seljuq Turks then arrived in 1071 and changed its name to Kara Hissar ("black castle") after the ancient fortress situated upon a volcanic rock 201 meters above the town. Following the dispersal of the Seljuqs the town was occupied by the Sâhib Ata and then the Germiyanids. The castle was much fought over during the Crusades and was finally conquered by the Ottoman Sultan Beyazid I in 1392 but was lost after the invasion of Timur Lenk in 1402. It was recaptured in 1428 or 1429. The area thrived during the Ottoman Empire, as the centre of opium production and Afyon became a wealthy city. In 1902, a fire burning for 32 hours destroyed parts of the city. During the 1st World War British prisoners of war who had been captured at Gallipoli were housed here in an empty Armenian church at the foot of the rock. During the Greco-Turkish War (1919-1922) campaign (part of the Turkish War of Independence) Afyon and the surrounding hills were occupied by Greek forces. However, it was recovered on 27 August 1922, a key moment in the Turkish counter-attack in the Aegean region. After 1923 Afyon became a part of the Republic of Turkey. The region was a major producer of raw opium (hence the name "Afyon") until the late 1960s when under international pressure, from the US in particular, the fields were burnt and production ceased. Now poppies are grown under a strict licensing regime. They do not produce raw opium any more but derive Morphine and other opiates using the poppy straw method of extraction. Afyon was depicted on the reverse of the Turkish 50 lira banknote of 1927–1938. The economy of Afyonkarahisar is based on agriculture, industries and thermal tourism. Especially its agriculture is strongly developed from the fact, a large part of its population living in the countrysides. Which stimulated agricultural activities greatly. Afyonkarahisar produces an important chunk of Turkish processed marbles, it ranks second on processed marble exports and fourth on travertine. Afyon holds an important share of Turkish marble reserves, with some 12,2% of total Turkish reserves. Afyon has unique marble types and colors, which were historically very renown and are unique to Afyon. Like "Afyon white", historically known as "Synnadic white". "Afyon Menekse", historically known as "Pavonazzetto" and "Afyon kaplan postu", this type was not popular. Historically marble from Afyon was generally referred to as "Docimeaen marble". Docimian marble was highly admired and valued for its unique colors and fine grained quality, by ancients such as Romans. When the Romans took control over Docimian quarries, they were blown away about the beautiful color combinations of Docimian Pavonazzetto, which is a type of white marble with purple veins. A trend started about it right away. Emperors Augustus, Trajan, Hadrian, all made extensive use of Docimian marble to all their major building projects. Docimian Pavonazzetto was extensively used in major building projects in the very heart of Rome aka Forum Romanum and all over the empire. Pavonazzetto was used on the most eye catching places such as, columns, wall and floor veneer and wall reliefs. Other marbles from all corners of the empire were used in combination, whenever Pavonazzetto was used as floor cover, it was usually in combination with other decorative marbles. But, the Pavonazzetto being a white marble was mostly the dominant color and gave the buildings a freshening white color. One of the greatest Roman architectural piece, the Pantheon contains Docimian Pavonazzetto as floor pavement along with other marble types. The dominant white color is the Pavonazzetto, also some of the interior main columns and pilasters are made from Docimian marble. Other buildings in Roman capital which contains or contained Docimian marble were, Forum of Augustus, Forum of Trajan (floor and 184 column shafts), Temple of Mars Ultor (floor), Temple of Apollo (floor), Basilica Aemelia (20 statues), Basilica Julia (floor and some columns), Basilica Ulpia (some of the columns), Basilica San Paolo Fuori Le Mura (24 columns, destroyed by fire in 1823), The eight statues on the Arch of Constantine, The greatest Roman bath, Baths of Caracalla (some of the columns and wall veneer. Other major buildings outside Rome, Italy were: The Hagia Sophia, one of the greatest buildings ever built, has Docimian marble as veneer on the aisles and galleries. The heart of Catholic Christianity, Saint Peter's Basilica, as veneer. Lepcis Magna, former limestone columns were replaced with Pavonazzetto. Library of Celsus, the columns on the famous wall. Ancient City of Sagalassos, as wall and floor covering, 40 tons of veneer were recovered. Temple of Zeus and Hera in Greece, 100 columns and wall. Docimian marble was also preferred for sarcophagi sculpting, some emperors preferred this marble for its high value and majestic looks. As a result, one of the greatest masterpieces were made from this material. Such as sarcophagus of, Sidamara, Silifkeh, Seleukeia, Eudocia, Heraclius...several hundreds sarcophagi were constructed. The geography of Afyon has great geothermal activity. Hence, the place has plenty thermal springs, there are about 5 main springs. Each of them have high mineral contents and high degree's, 40 °C-100 °C. The waters have strong healing properties to some diseases. As a result, plenty of thermals commenced over time. In time Afyon has developed its thermal sector with more capacity, comfort and innovation. Afyon combined the traditional bath houses with 5 star resorts, the health benefits of the natural springs have put the thermal resorts further then a mere attraction. Hospitals and universities have come in association with thermal resorts, to utilize the full health potentials of the thermals. As such, Kocatepe University Physical Therapy and Rehabilitation Hospital opened for that purpose. Afyon now has the largest residence capacity of thermal resorts, of which a large part are 5 star thermal hotels which give medical care with qualified personnel. Kızılay, was the first mineral water factory in Turkey which opened in Afyon, in 1926 by Atatürk. After the mineral water from Gazligöl springs, healed Atatürk's kidneys and proved its health benefits. Since its foundation, "Kızılay Spa Water" grew as the biggest spa water distributor in Turkey, Middle-East and Balkans. Almost a third of all the morphine produced in the world derives from alkaloids factory in Afyon, named as "Afyon Alkaloids". this large capacity is the byproduct of Afyon's poppy plantations. The pharmaceuticals derive from the opium of the poppy capsules. "Afyon Alkaloids" factory is the largest of its kind in the world, with high capacity processing ability and modern laboratories. The raw opium is put through a chain of biochemical processes, resulting into several types of morphine. In the Alkaloid Extraction Unit only base morphine is produced. In the adjacent Derivatives Unit half of the morphine extracted is converted to morphine hydrochloride, codeine, codeine phosphate, codeine sulphate, codeine hydrochloride, morphine sulphate, ethylmorphine hydrochloride. Livestocks Afyon breeds a large amount of livestocks, its landscape and demography is suitable for this field. As such it ranks in the top 10 within Turkey in terms of amounts of sheep and cattle it has. Meat and meat products As a result of being an important source of livestock, related sectors such as meat and meat products are also very productive in afyon. Its one of the leading provinces in red meat production and has very prestigious brand marks of sausages, such as "Cumhuriyet Sausages". Eggs Afyon is the sole leader in egg production within Turkey. It has the largest amount of laying hens, with a figure of 12,7 million. And produces a record amount of 6 million eggs per day. Cherries and sour cherries Sour cherries are cultivated in Afyon in very large numbers, so much so that it became very iconic to Afyon. Every year, a sour cherry festival takes place in the Cay district. It is the largest producer of sour cherries in Turkey. The sour cherries grown in Afyon are of excellent quality because of the ideal climate they're grown in. For the same reason Afyon is also an ideal place for cherry cultivation. First quality cherries known as "Napolyon Cherries" are grown in abundance, its one of the top 5 leading provinces. Poppy One of the iconic agricultural practices of Afyon is the cultivation of poppy. Afyon's climate is ideal for the cultivation of this plant, hence a large amount of poppy plantation occurs in this region. Though, a strong limitation came some decades ago from international laws, cause of the opium content of poppy plants peels. Nevertheless, Afyon is the largest producer of poppy in Turkey and accounts for a large amount of global production. Potatoes and sugar-beets Afyon has a durable reputation in potato production, it produces around 8% of Turkish potato need. It ranks in the top 5 in potato, sugar-beets, cucumber and barley production. Afyonkarahisar has a hot dry-summer continental climate (Dsa) under the Köppen classification and a hot summer continental (Dca) or hot summer oceanic climate (Doa) under the Trewartha classification. The winters are cold and snowy winters and the summers are hot and dry with cool nights. Rainfall occurs mostly during the spring and autumn. Afyon is the centre of an agricultural area and the city has a country town feel to it. There is little in the way of bars, cafes, live music or other cultural amenities, and the standards of education are low for a city in the west of Turkey. Afyon Kocatepe University. Afyon is known for its marble (in 2005 there were 355 marble quarries in the province of Afyon producing high quality white stone), its "sucuk" (spiced sausages), its "kaymak" (meaning either "cream" or a white Turkish Delight) and various handmade weavings. There is also a large cement factory. This is a natural crossroads, the routes from Ankara to İzmir and from Istanbul to Antalya intersect here and Afyon is a popular stopping-place on these journeys. There are a number of well-established roadside restaurants for travellers to breakfast on the local cuisine. Some of these places are modern well-equipped hotels and spas; the mineral waters of Afyon are renowned for their healing qualities. There is also a long string of roadside kiosks selling the local Turkish delight. Afyon is also an important rail junction between İzmir, Konya, Ankara and Istanbul. Afyon is on the route of the planned high-speed rail line between Ankara and Izmir. With its rich architectural heritage, the city is a member of the European Association of Historic Towns and Regions . Following list is alphabetically sorted after family name.
https://en.wikipedia.org/wiki?curid=2668
Abba Arikha Abba Arikha (175–247 CE; Jewish Babylonian Aramaic: ; born: "Rav Abba bar Aybo", ), commonly known as Rav (), was a Jewish amora of the 3rd century. He was born and lived in Kafri, Asoristan, in the Sasanian Empire. Abba Arikha established at Sura the systematic study of the rabbinic traditions, which, using the Mishnah as text, led to the compilation of the Talmud. With him began the long period of ascendancy of the great Talmudic Academies in Babylonia around the year 220. In the Talmud, he is frequently associated with Samuel of Nehardea, with whom he debated many issues. His surname, Arikha (English, "Long"—that is, "Tall"; it occurs only once—"Hullin" 137b), he owed to his height, which exceeded that of his contemporaries. Others, reading Arekha, consider it an honorary title, "Lecturer". In the traditional literature he is referred to almost exclusively as Rav, "the Master", (both his contemporaries and posterity recognizing in him a master), just as his teacher, Judah HaNasi, was known simply as "Rabbi". He is called Rabbi Abba only in the "tannaitic" literature, where a number of his sayings are preserved. He occupies a middle position between the "Tannaim" and the "Amoraim", and is accorded the right, rarely conceded to one who is only an "amora", of disputing the opinion of a "tanna". Rav was a descendant of a distinguished Babylonian family which claimed to trace its origin to Shimei, brother of King David. His father, Aibo, was a brother of Hiyya the Great who lived in Palestine, and was a highly esteemed scholar in the collegiate circle of the patriarch Judah haNasi. From his associations in the house of his uncle, and later as his uncle's disciple and as a member of the academy at Sepphoris, Rav acquired such knowledge of the tradition as to make him its foremost exponent in Babylonia. While Judah haNasi was still living, Rav, having been ordained as teacher (with certain restrictions), returned to Babylonia, where he at once began a career that was destined to mark an epoch in the development of Babylonian Judaism. In the annals of the Babylonian schools, the year of his arrival is recorded as the starting-point in the chronology of the Talmudic age. It was the 530th year of the Seleucidan and the 219th year of the common era. As the scene of his activity, Rav first chose Nehardea, where the exilarch appointed him "agoranomos", or market-master, and Rabbi Shela made him lecturer ("amora") of his college. Then he moved to Sura, on the Euphrates, where he established a school of his own, which soon became the intellectual center of the Babylonian Jews. As a renowned teacher of the Law and with hosts of disciples, who came from all sections of the Jewish world, Rav lived and worked in Sura until his death. Samuel, another disciple of Judah haNasi, at the same time brought to the academy at Nehardea a high degree of prosperity; in fact, it was at the school of Rav that Jewish learning in Babylonia found its permanent home and center. Rav's activity made Babylonia independent of Palestine, and gave it that predominant position which it was destined to occupy for several centuries. Little is known of Rav's personal life. That he was rich seems probable; for he appears to have occupied himself for a time with commerce and afterward with agriculture. He is referred to as the son of noblemen, but it is not clear if this is an affectionate term or a true description of his status. Rashi does tell us that he is being described as the son of great men. He was highly respected by the Gentiles as well as by the Jews of Babylonia, as shown by the friendship which existed between him and the last Parthian king, Artaban. He was deeply affected by the death of Artaban (226) and the downfall of the Arsacid dynasty, and does not appear to have sought the friendship of Ardeshir, founder of the Sassanian dynasty, although Samuel of Nehardea probably did so. Rav became closely related, through the marriage of one of his daughters, to the family of the exilarch. Her sons, Mar Ukba and Nehemiah, were considered types of the highest aristocracy. Rav had many sons, several of whom are mentioned in the Talmud, the most distinguished being the eldest, Chiyya. Chiyya did not, however, succeed his father as head of the academy: this post fell to Rav's disciple Rav Huna. Two of his grandsons occupied in succession the office of exilarch. Rav died at an advanced age, deeply mourned by numerous disciples and the entire Babylonian Jewry, which he had raised from comparative insignificance to the leading position in Judaism. The method of treatment of the traditional material to which the Talmud owes its origin was established in Babylonia by Rav. That method takes the Mishnah of Judah haNasi as a text or foundation, adding to it the other "tannaitic" traditions, and deriving from all of them the theoretical explanations and practical applications of the religious Law. The legal and ritual opinions recorded in Rav's name and his disputes with Samuel constitute the main body of the Babylonian Talmud. His numerous disciples—some of whom were very influential and who, for the most part, were also disciples of Samuel—amplified and, in their capacity as instructors and by their discussions, continued the work of Rav. In the Babylonian schools, Rav was rightly referred to as "our great master." Rav also exercised a great influence for good upon the moral and religious conditions of his native land, not only indirectly through his disciples, but directly by reason of the strictness with which he repressed abuses in matters of marriage and divorce, and denounced ignorance and negligence in matters of ritual observance. Rav, says tradition, found an open, neglected field and fenced it in. He gave special attention to the liturgy of the synagogue. The Aleinu prayer first appeared in the manuscript of the Rosh Hashana liturgy by Rav. He included it in the Rosh Hashana mussaf service as a prologue to the Kingship portion of the Amidah. For that reason some attribute to Rav the authorship, or at least the revising, of Aleinu. In this noble prayer are evinced profound religious feeling and exalted thought, as well as ability to use the Hebrew language in a natural, expressive, and classical manner. The many homiletic and ethical sayings recorded of him show similar ability. The greatest aggadist among Babylonian "Amoraim", he is the only one of them whose aggadic utterances approach in number and contents those of the Palestinian haggadists. The Jerusalem Talmud has preserved a large number of his halakhic and aggadic utterances; and the Palestinian "Midrashim" also contain many of his "aggadot". Rav delivered homiletic discourses, both in the beit midrash and in the synagogues. He especially loved to discuss in his homilies the events and personages of Biblical history; and many beautiful and genuinely poetic embellishments of the Biblical record, which have become common possession of the aggadah, are his creations. His "aggadah" is particularly rich in thoughts concerning the moral life and the relations of human beings to one another. A few of these teachings may be quoted here: Rav loved the "Book of Ecclesiasticus" (Sirach), and warned his disciple Hamnuna against unjustifiable asceticism by quoting its advice that considering the transitoriness of human life, one should not despise the good things of this world. To the celestial joys of the future he was accustomed to refer in the following poetic words: Rav also devoted much attention to mystical and transcendental speculations regarding Maaseh Bereshit, Maaseh Merkabah, and the Divine Name. Many of his important utterances testify to his tendency in this direction.
https://en.wikipedia.org/wiki?curid=2670
Abbahu Rabbi Abbahu () was a Jew and Talmudist of the Talmudic Academies in Syria Palaestina from about 279-320 and is counted a member of the third generation of amoraim. He is sometimes cited as "Rabbi Abbahu of Kisrin" (Caesarea). His rabbinical education was acquired mainly at Tiberias in the academy presided over by Johanan bar Nappaha, with whom his relations were almost those of a son. He frequently made pilgrimages to Tiberias, even after he had become well known as rector of the Caesarean academy. Abbahu was an authority on weights and measures. He encouraged the study of Koine Greek by Jews. He learned Greek himself in order to become useful to his people, then under the Roman "proconsuls", that language having become, to a considerable extent, the rival of Hebrew even in prayer. In spite of the bitter protest of Shimon bar Abba, he also taught his daughters Greek. Indeed, it was said of Abbahu that he was a living illustration of the Biblical maxim, "It is good that you should take hold of this [the study of the Law]; yea, also from that [other branches of knowledge] withdraw not your hand: for he that fears God shall come forth of them all". Being wise, handsome, and wealthy, Abbahu became not only popular with his coreligionists, but also influential with the proconsular government. On one occasion, when his senior colleagues, Hiyya bar Abba, Rabbi Ammi, and Rabbi Assi, had punished a certain woman, and feared the wrath of the proconsul, Abbahu was deputed to intercede for them. He had, however, anticipated the rabbis' request, and wrote them that he had appeased the informers but not the accuser. The witty enigmatic letter describing this incident, preserved in the Talmud, is in the main pure Hebrew, and even includes Hebrew translations of Greek proper names, to avoid the danger of possible exposure should the letter have fallen into the hands of enemies and informers. After his ordination he declined a teacher's position, recommending in his stead a more needy friend, Abba of Acre, as worthier than himself. He thereby illustrated his own doctrine that it is a divine virtue to sympathize with a friend in his troubles as well as to partake of his joys. Later he assumed the office of rector in Caesarea, the former seat of Hoshaiah Rabbah, and established himself at the so-called Kenishta Maradta (Insurrectionary Synagogue); from which some of the most prominent teachers of the next generation issued. In Caesarea he originated several ritual rules, one of which (that regulating the sounding of the shofar) has since been universally adopted, and is referred to by rishonim as "the Enactment of R. Abbahu". He did not confine his activity to Caesarea, but also visited and taught in many other Jewish towns. On these journeys, Abbahu gathered so many "halakhot" that scholars turned to him for information on mooted questions. In the course of these travels he made a point of complying with all local enactments, even where such compliance laid him open to the charge of inconsistency. On the other hand, where circumstances required it, he did not spare even the princes of his people. Where, however, the rigorous exposition of laws worked hardship on the masses, he did not scruple to modify the decisions of his colleagues for the benefit of the community. As for himself, he was very strict in the observance of the laws. Once he ordered some Samaritan wine, but subsequently hearing that the Samaritans no longer strict observed the dietary laws, with the assistance of his colleagues (Hiyya bar Abba, Rabbi Ammi, and Rabbi Assi) he investigated the report, and, ascertaining it to be well founded, ruled the Samaritans to be equivalent to Gentiles for all ritual purposes. Abbahu's chief characteristic seems to have been modesty. While lecturing in different towns, he met R. Hiyya bar Abba, who was lecturing on intricate halakhic themes. As Abbahu delivered popular sermons, the masses naturally crowded to hear him, and deserted the halakhist. At this apparent slight, Hiyya manifested chagrin, and Abbahu hastened to comfort him by comparing himself to the peddler of glittering fineries that always attracted the eyes of the masses, while his rival was a trader in precious stones, the virtues and values of which were appreciated only by the connoisseur. This speech not having the desired effect, R. Abbahu showed special respect for his slighted colleague by following him for the remainder of that day. "What," said Abbahu, "is my modesty as compared with that of Abba of Acre, who does not even remonstrate with his interpreter for interpolating his own comments in the lecturer's expositions." When his wife reported to him that his interpreter's wife had boasted of her own husband's greatness, Abbahu simply said, "What difference does it make which of us is really the greater, so long as through both of us heaven is glorified?" His principle of life he expressed in the maxim, "Let man ever be of the persecuted, and not of the persecutors; for there are none among the birds more persecuted than turtle-doves and pigeons, and the Scriptures declare them worthy of the altar." Abbahu had two sons, Zeira and Hanina. Some writers ascribe to him a third son, Abimi. Abbahu sent Hanina to the academy at Tiberias, where he himself had studied, but the youth occupied himself with the burial of the dead, and on hearing of this, the father sent him a reproachful message in this laconic style: "Is it because there are no graves in Caesarea that I have sent you off to Tiberias? Study must precede practice". Abbahu left behind him a number of disciples, the most prominent among whom were the leaders of the 4th amoraic generation, R. Jonah and R. Jose. At Abbahu's death the mourning was so great that it was said, "Even the statues of Caesarea shed tears". R. Abbahu, though eminent as a halakhist, was more distinguished as an aggadist and controversialist. He had many interesting disputes with the Christians of his day. Sometimes these disputes were of a humorous nature. Thus, a heretic bearing the name of Sason (=Joy) once remarked to him, "In the next world your people will have to draw water for me; for thus it is written in the Bible, 'With joy shall ye draw water.'" To this R. Abbahu replied, "Had the Bible said 'for joy' ["le-sason"], it would mean as you say, but since it says 'with joy' ["be-sason"], it means that we shall make bottles of your skin and fill them with water". These controversies, though forced on him, provoked resentment, and it is even related that his physician, Jacob the Schismatic ("Minaah"), was slowly poisoning him, but Rabbi Ammi and Rabbi Assi discovered the crime in time. A Christian ("Minaah") once asked Abbahu "When does your Messiah come?" in a tone of mockery. Abbahu replied: "When you will be wrapped in darkness, for it says, 'Behold, darkness shall cover the earth, and gross darkness the nations; then shall the Lord rise upon you and His glory shall be seen on you'." A Christian came to Abbahu with the quibbling question: "How could your God in His priestly holiness bury Moses without providing for purificatory rites, yet oceans are declared insufficient?" Abbahu replied: "Does it not say, 'The Lord comes with fire'? Fire is the true element of purification, according to Numbers 31:23." Another question of the same character: "Why the boastful claim, 'What nation on earth is like Your people Israel', since we read, 'All the nations are as nothing before Him'?" Abbahu replied: "Do we not read of Israel, he 'shall not be reckoned among the nations'?" Abbahu made a notable exception with reference to the Tosefta's statement that the "Gilyonim" (Gospels) and other books of the heretics ("Minnin") are not to be saved from a fire on Shabbat: "the books of those [written by "Minnin" for the purpose of debating with Jews] at Abidan may or may not be saved." In regard to the line ""Barukh Shem Kevod Malkhuto"" (Blessed be the Name of His glorious Kingdom) recited after the Shema, Abbahu says that in Palestine, where the Christians look for points of controversy, the words should be recited aloud (lest the Jews be accused of silently tampering with the unity of God proclaimed in the "Shema"), whereas in the Babylonian city of Nehardea, where there are no Christians, the words are recited with a low voice. Preaching directly against the Christian dogma, Abbahu says: "A king of flesh and blood may have a father, a brother, or a son to share in or dispute his sovereignty, but the Lord says, 'I am the Lord your God! I am the first - that is, I have no father; and I am the last - that is, I have no brother; and besides me there is no God - that is, I have no son'". His comment on Numbers 23:19 has a still more polemical tone: "God is not a man that he should lie; neither the son of man, that he should repent; if a man says: 'I am God,' he is a liar; if he says: 'I am a son of man,' he will have cause to regret it; and if he says, 'I will go up to heaven,' he has said ["something"] but will not keep his word". Some of his controversies on Christian theological subjects, as on Adam, on Enoch, and on the resurrection, are less clear and direct. There are several other Abbahus mentioned in the Talmudim and Midrashim, prominent among whom is Abbahu (Abuha, Aibut) b. Ihi (Ittai), a Babylonian halakhist, contemporary of Samuel and Anan, and brother of Minyamin (Benjamin) bar Ihi. While this Abbahu repeatedly applied to Samuel for information, Samuel in return learned many halakhot from him. It has the following bibliography:
https://en.wikipedia.org/wiki?curid=2671
Abbreviator An Abbreviator (plural "Abbreviators" in English and "Abbreviatores" in Latin) or Breviator was a writer of the Papal Chancery who adumbrated and prepared in correct form Papal bulls, briefs, and consistorial decrees before these were written out "in extenso" by the "scriptores". They are first mentioned in the Papal bull "Extravagantes" of Pope John XXII and in a Papal bull of Pope Benedict XII. After the protonotaries left the adumbration of the minutes to the Abbreviators, those "de Parco majori" of the dignity of prelate were the most important officers of the Papal Chancery. By the pontificate of Pope Martin V their signature was essential to the validity of the acts of the Chancery. Over time they obtained many important privileges. Abbreviators make an abridgment or abstract of a long writing or discourse by contracting the parts, i. e., the words and sentences; an abbreviated form of writing common among the ancient Romans. Abbreviations were of two kinds: the use of a single letter for a single word and the use of a sign, note, or mark for a word or phrase. The Emperor Justinian forbade the use of abbreviations in the compilation of the "Digest" and afterward extended his prohibition to all other writings. This prohibition was not universally obeyed. The Abbreviators found it convenient to use the abbreviated form, and this was especially the case in Rome. The early Christians practised the abbreviated mode, no doubt as an easy and safe way of communicating with one another and safeguarding their secrets from enemies and false brethren. In course of time the Papal Chancery adopted this mode of writing as the "curial" style, still further abridging by omitting the diphthongs "ae" and "oe", and likewise all lines and marks of punctuation. The "Abbreviatores" were officials of the Roman Curia. The scope of its labour, as well as the number of its officials, varied over time. Up to the twelfth or thirteenth century, the duty of the Apostolic – or Roman – Chancery was to prepare and expedite the Papal letters and writs for collation of ecclesiastical dignitaries and other matters of grave importance which were discussed and decided in Papal consistory. About the thirteenth or fourteenth century, the Popes, then residing in Avignon, France, began to reserve the collation of a great many benefices, so that all the benefices, especially the greater ones, were to be conferred through the Roman Curia (Lega, "Praelectiones Jur. Can.", 1, 2, 287). As a consequence, the labour was immensely augmented, and the number of "Abbreviatores" necessarily increased. To regulate the proper expedition of these reserved benefices, Pope John XXII instituted the rules of chancery to determine the competency and mode of procedure of the Chancery. Afterwards the establishment of the "Dataria Apostolica" and the Secretariate of Briefs lightened the work of the Chancery and led to a reduction in the number of "Abbreviatores". According to Ciampini ("Lib. de abbreviatorum de parco majore etc.", Cap. 1) the institution of curial abbreviators was very ancient, succeeding after the persecutions to the notaries who recorded the acts of the martyrs. Other authors reject this early institution and ascribe it to Pope John XXII in 1316. It is certain that he uses the name ""abbreviatores"", but speaks as if they had existed before his time, and had, by over-taxation of their labour, caused much complaint and protest. He ("Extravag. Joan.", Tit. 13, "Cum ad Sacrosanctae Romanae Ecclesiae") prescribed their work, determined how much they could charge for their labour, fixed a certain tax for an abstract or abridgment of twenty-five words or their equivalent at 150 letters, forbade them to charge more, even though the abstract was over twenty-five words but less than fifty words, enacted that the basis of the tax was the labour employed in writing, expediting, etc. the bulls, and by no means the emoluments that accrued to the recipient of the favour or benefice conferred by the bull, and declared that whoever charged more than the tax fixed by him was suspended for six months from office, and upon a second violation of the law, was deprived of it altogether, and if the delinquent was an abbreviator, he was excommunicated. Should a large letter have to be rewritten, owing to the inexact copy of the abbreviator, the abbreviator and not the receiver of the bull had to pay the extra charge for the extra labour to the Apostolic writer. Whatever may be the date of the institution of the office of abbreviator, it is certain that it became of greater importance and more highly privileged upon its erection into a college of prelates. Pope Martin V (Constit. 3 "In Apostolicae", 2 and 5) fixed the manner for their examination and approbation and also the tax they could demand for their labour and the punishment for overcharge. He also assigned to them certain remunerations. The Abbreviators of the lower, or lesser, were to be promoted to the higher, or greater, bar or presidency. Their offices were compatible with other offices, i. e. they could hold two benefices or offices simultaneously, some conferred by the Cardinal Vice Chancellor, others by the Pope. In the pontificate of Pope Pius II, their number, which had been fixed at twenty-four, had overgrown to such an extent as to diminish considerably the individual remuneration, and, as a consequence, competent men no longer sought the office, and hence the old style of writing and expediting the bulls was no longer used, to the great injury of justice, the interested parties, and the dignity of the Apostolic See. To remedy this and to restore the old established chancery style, the Pope selected out of the many then living Abbreviators seventy, and formed them into a college of prelates denominated the "College of Abbreviators", and decreed that their office should be perpetual, that certain remunerations should be attached to it, and granted certain privileges to the possessors of the same. He ordained further that some should be called "Abbreviators of the Upper Bar" ("Abbreviatores de Parco Majori"; the name derived from a place in the Chancery that was surrounded by a grating, in which the officials sat, which is called higher or lower (major or minor) according to the proximity of the seats to that of the Vice Chancellor), the others of the Lower Bar ("Abbreviatores de Parco Minori"); that the former should sit upon a slightly raised portion of the chamber, separated from the rest of the chamber by lattice work, assist the Cardinal Vice-Chancellor, subscribe the letters and have the principal part in examining, revising, and expediting the Apostolic letters to be issued with the leaden seal; that the latter, however, should sit among the Apostolic writers upon benches in the lower part of the chamber, and their duty was to carry the signed schedules or supplications to the prelates of the Upper Bar. Then one of the prelates of the Upper Bar made an abstract, and another prelate of the same bar revised it. Prelates of the Upper Bar formed a quasi-tribunal, in which as a college they decided all doubts that might arise about the form and quality of the letters, of the clauses and decrees to be adjoined to the Apostolic letters, and sometimes about the payment of the remunerations and other contingencies. Their opinion about questions concerning Chancery business was held in the highest estimation by all the Roman tribunals. Pope Paul II suppressed the College; but Pope Sixtus IV ("Constitutio" 16, "Divina") re-instituted it. He appointed seventy-two abbreviators, of whom twelve were of the upper, or greater, and twenty-two of the lower, or lesser, presidency ("parco"), and thirty-eight examiners on first appearance of letters. They were bound to be in attendance on certain days under penalty of fine, and sign letters and diplomas. Ciampini mentions a decree of the Vice Chancellor by which absentees were mulcted in the loss of their share of the remuneration of the following session of the Chancery. The same Pope also granted many privileges to the College of Abbreviators, but especially to the members of the greater presidency. Pope Pius VII suppressed many of the offices of the Chancery, and so the Tribunal of Correctors and the Abbreviators of the lower presidency disappeared. Of the Tribunal of Correctors, a substitute-corrector alone remains. Bouix ("Curia Romana", edit. 1859) chronicled the suppression of the lower presidency and put the number of Abbreviators at that date at eleven. Later the College consisted of seventeen prelates, six substitutes, and one sub-substitute, all of whom, except the prelates, were clerics or laity. Although the duty of Abbreviators was originally to make abstracts and abridgments of the Apostolic letters, diplomas, et cetera, using the legal abbreviations, clauses, and formularies, in course of time, as their office grew in importance they delegated that part of their office to their substitute and confined themselves to overseeing the proper expedition of the Apostolic letters. Prior to 1878, all Apostolic letters and briefs requiring for their validity the leaden seal were engrossed upon rough parchment in Gothic characters or round letters, also called "Gallicum" and commonly "Bollatico", but in Italy "Teutonic", without lines, diphthongs, or marks of punctuation. Bulls engrossed on a different parchment, or in different characters with lines and punctuation marks, or without the accustomed abbreviations, clauses, and formularies, were rejected as spurious. Pope Leo XIII in his "Constitutio Universae Eccles." of 29 December 1878 ordained that they should be written henceforth in ordinary Latin characters upon ordinary parchment and that no abbreviations were to be used except those easily understood. Many great privileges were conferred upon Abbreviators. By decree of Pope Leo X they were elevated as Papal nobles, ranking as "Comes palatinus" ("Count Palatine"), familiars and members of the Papal household, so that they might enjoy all the privileges of domestic prelates and of prelates in actual attendance on the Pope, as regards plurality of benefices as well as expectatives. They and their clerics and their properties were exempt from all jurisdiction except the immediate jurisdiction of the Pope, and they were not subject to the judgments of the Auditor of Causes or the Cardinal Vicar. He also empowered them to confer, later within strict limitations, the degree of Doctor, with all university privileges, institute notaries (later abrogated), legitimize children so as to make them eligible to receive benefices vacated by their fathers (later revoked), also to ennoble three persons and to make Knights of the Order of St. Sylvester ("Militiae Aureae"), the same to enjoy and to wear the insignia of nobility. Pope Gregory XVI rescinded this privilege and reserved to the Pope the right of institution of such knights ("Acta Pont. Greg. XVI", Vol. 3, 178-179-180). Pope Paul V, who in early manhood was a member of the College (Const. 2, "Romani"), made them Referendaries of Favours, and after three years of service, Referendaries of Justice, enjoying the privileges of Referendaries and permitting one to assist in the signatures before the Pope, giving all a right to a portion in the Papal palace and exempting them from the registration of favours as required by Pope Pius IV (Const., 98) with regard to matters pertaining to the Apostolic Chamber. They followed immediately after the twelve voting members of the Signature "in capella". Abbreviators of the greater presidency were permitted to wear the purple cassock and "cappa", as also rochet "in capella". Abbreviators of the lower presidency before their suppression were simple clerics, and according to permission granted by Pope Sixtus IV (loc. cit.) might be even married. These offices becoming vacant by death of the Abbreviator, no matter where the death occurred, were reserved to the Roman Curia. The prelates could resign their office in favour of others. Formerly these offices as well as those of the other Chancery officers from the Regent down were occasions of venality, until Popes, especially Pope Benedict XIV and Pope Pius VII, gradually abolished that. Pope Leo XIII in a motu proprio of 4 July 1898 most solemnly decreed the abolition of all venality in the transfer or collation of the said offices. As domestic prelates, prelates of the Roman Curia, they had personal preeminence in every diocese of the world. They were addressed as "Reverendissimus", "Right Reverend", and "Monsignor". As prelates, and therefore possessing the legal dignity, they were competent to receive and execute Papal commands. Pope Benedict XIV (Const. 3, "Maximo") granted prelates of the greater presidency the privilege of wearing a hat with a purple band, which right they held even after they ceased to be abbreviators. Pope Pius X abrogated the College in 1908 and their obligations were transferred to the "protonotarii apostolici participantes".
https://en.wikipedia.org/wiki?curid=2673
Abd al-Latif al-Baghdadi ʿAbd al-Laṭīf al-Baghdādī (, 1162 Baghdad–1231 Baghdad), short for Muwaffaq al-Dīn Muḥammad ʿAbd al-Laṭīf ibn Yūsuf al-Baghdādī (), was a physician, philosopher, historian, Arabic grammarian and traveler, and one of the most voluminous writers in the Near East of his time. Many details of ʿAbd al-Laṭīf al-Baghdādī's life are known from his autobiography as present in Ibn Abī Uṣaybiʿah's literary history of medicine. As a young man, he studied grammar, law, tradition, medicine, alchemy and philosophy. He focused his studies on ancient authors, in particular Aristotle, after first adopting Avicenna [Ibn Sīnā] as his philosophical mentor at the suggestion of a wandering scholar from the Maghreb. He traveled extensively and resided for a while in Mosul (in 1189) where he studied the works of al-Suhrawardi before traveling on to Damascus (1190) and the camp of Saladin outside Acre (1191). It was at the latter location that he met Baha’ al-Din Ibn Shaddad and ‘Imad al-Din al-Isfahani and acquired the Qadi al-Fadil's patronage. He went on to Cairo, where he met Abu'l-Qasim al-Shari'i, who introduced him to the works of al-Farabi, Alexander of Aphrodisias, and Themistius and (according to al-Latif) turned him away from Avicenna and alchemy. He met Saladin himself in 1192 in Jerusalem and enjoyed his patronage, then went to Damascus again before returning to Cairo. In later years he again journeyed to Jerusalem and to Damascus in 1207-8, and eventually made his way via Aleppo to Erzindjan, where he remained at the court of the Mengujekid Ala’-al-Din Da’ud [Dāwūd Shāh] until the city was conquered by the Rūm Seljuk ruler Kayqubād Ibn Kaykhusraw [Kayqubād II]. ‘Abd al-Latif returned to Baghdad in 1229, travelling back via Erzerum, Kamakh, Divriği and Malatya. He died in Baghdad two years later. ʿAbd al-Laṭīf was undoubtedly a man of great knowledge and of an inquisitive and penetrating mind. Of the numerous works (mostly on medicine) which Ibn Abī Uṣaybiʿah ascribes to him, one only, his graphic and detailed "Account of Egypt" (in two parts), appears to be known in Europe. ʿAbd-al-Laṭīf was well aware of the value of ancient monuments. He praised Muslim rulers for preserving and protecting pre-Islamic artifacts and monuments, but he also criticized them for failing to do this. He noted that the preservation of antiquities presented a number of benefits for Muslims: While discussing the profession of treasure hunting, he notes that poorer treasure hunters were often sponsored by rich businessmen to go on archeological expeditions. In some cases, an expedition could turn out to be fraud, with the treasure hunter disappearing with large amounts of money extracted from sponsors. This fraudulent practice continues to the present day, with rich businessmen in Egypt still being deceived by local treasure hunters. This work was one of the earliest works on Egyptology. It contains a vivid description of a famine caused, during the author's residence in Egypt, by the Nile failing to overflow its banks. He also wrote detailed descriptions on ancient Egyptian monuments. Al-Baghdādī wrote that during the famine in Egypt in 597 AH (1200 AD), he had the opportunity to observe and examine a large number of skeletons. This was one of the earliest examples of a postmortem autopsy, through which he discovered that Galen was incorrect regarding the formation of the bones of the lower jaw [mandible], coccyx, and sacrum. The Arabic manuscript was discovered in 1665 by Edward Pococke the orientalist, and preserved in the Bodleian Library {under siglum 230}. He then published the Arabic manuscript in the 1680s. His son, Edward Pococke the Younger, translated the work into Latin, though he was only able to publish less than half of his work. Thomas Hunt attempted to publish Pococke's complete translation in 1746, though his attempt was unsuccessful. Pococke's complete Latin translation was eventually published by Joseph White of Oxford in 1800. The work was then translated into French, with valuable notes, by Silvestre de Sacy in 1810. As far as philosophy is concerned, one may adduce that ʿAbd al-Laṭīf al-Baghdādī regarded philosophers as paragons of real virtue and therefore he refused to accept as a true philosopher one lacking not only true insight, but also a truly moral personality as true philosophy was in the service of religion, verifying both belief and action. Apart from this he regarded the philosophers’ ambitions as vain (Endress, in Martini Bonadeo, Philosophical journey, xi). ʿAbd al-Laṭīf composed several philosophical works, among which is an important and original commentary on Aristotle’s Metaphysics (Kitāb fī ʿilm mā baʿd al-ṭabīʿa). This is a critical work in the process of the Arabic assimilation of Greek thought, demonstrating its author’s acquaintance with the most important Greek metaphysical doctrines, as set out in the writings of al-Kindī (d. circa 185-252/801-66) and al-Fārābī (d. 339/950). The philosophical section of his Book of the Two Pieces of Advice [Kitāb al-Naṣīḥatayn] contains an interesting and challenging defence of philosophy and illustrates the vibrancy of philosophical debate in the Islamic colleges. It moreover emphasises the idea that Islamic philosophy did not decline after the twelfth century CE (Martini Bonadeo, Philosophical journey; Gutas). ʿAbd al-Laṭīf al-Baghdādī may therefore well be an exponent of what Gutas calls the “golden age of Arabic philosophy” (Gutas, 20). ʿAbd al-Laṭīf also penned two passionate and somewhat grotesque pamphlets against the art of alchemy in all its facets. Although he engaged himself with alchemy for a short while, he later abandoned the art completely by rejecting not only its practice, but also its theory. In ʿAbd al-Laṭīf’s view alchemy could not be placed in the system of the sciences, and its false presumptions and pretensions must be distinguished from true scientific knowledge, which can be given a rational basis (Joosse, Rebellious intellectual, 29-62; Joosse, Unmasking the craft, 301-17; Martini Bonadeo, Philosophical journey, 5-6 and 203-5; Stern, 66-7; Allemann). During the years following the Great War [i.e. the First World War, 1914-18] ʿAbd al-Laṭīf al-Baghdādī’s name miraculously reappeared within the spiritualistic movement in the United Kingdom. He was introduced to the public by the Irish medium Eileen J. Garrett, the author Sir Arthur Conan Doyle and the spiritualist R.H. Saunders and became known as Abduhl Latif, the great Persian physician. He is said to have acted as a control of mediums until the mid 1960’s (Joosse, Geest, 221-9). The Bodleian Library MS Pococke 230 and the interpretation of the Videans (Zand-Videan, 8-9) may also have prompted the whimsical short-story ‘Ghost Writer’, as told to Tim Mackintosh-Smith, in which ʿAbd al-Laṭīf al-Baghdādī speaks in the first person.
https://en.wikipedia.org/wiki?curid=2674
Abd al-Rahman I Abd al-Rahman I, more fully Abd al-Rahman ibn Mu'awiya ibn Hisham ibn Abd al-Malik ibn Marwan (731–788) (), was the founder of a Muslim dynasty that ruled the greater part of Iberia for nearly three centuries (including the succeeding Caliphate of Córdoba). Abd al-Rahman was a member of the Umayyad dynasty in Damascus, and his establishment of a government in Iberia represented a break with the Abbasids, who had overthrown the Umayyads in 750. He was also known by the surnames "al-Dakhil" ("the Entrant"), "Saqr Quraish" ("the Falcon of Quraysh") and the "Falcon of Andalus". Variations of the spelling of his name include Abd ar-Rahman I, Abdul Rahman I, Abdar Rahman I, and Abderraman I. Born near Damascus in Syria, Abd al-Rahman was the son of the Umayyad prince Mu'awiya ibn Hisham and his concubine Ra'ha, a Berber woman from the Nafza tribe, and thus the grandson of Hisham ibn Abd al-Malik, caliph from 724 to 743. He was twenty when his family, the ruling Umayyads, were overthrown by the Abbasid Revolution in 748–750. Abd al-Rahman and a small part of his family fled Damascus, where the center of Umayyad power had been; people moving with him include his brother Yahya, his four-year-old son Sulayman, and some of his sisters, as well as his Greek freedman, Bedr. The family fled from Damascus to the River Euphrates. All along the way the path was filled with danger, as the Abbasids had dispatched horsemen across the region to try to find the Umayyad prince and kill him. The Abbasids were merciless with all Umayyads that they found. Abbasid agents closed in on Abd al-Rahman and his family while they were hiding in a small village. He left his young son with his sisters and fled with Yahya. Accounts vary, but Bedr likely escaped with Abd ar-Rahman. Some histories indicate that Bedr met up with Abd al-Rahman at a later date. Abd al-Rahman, Yahya, and Bedr quit the village, narrowly escaping the Abbasid assassins. On the way south, Abbasid horsemen again caught up with the trio. Abd al-Rahman and his companions then threw themselves into the River Euphrates. The horsemen urged them to return, promising that no harm would come to them; and Yahya, perhaps from fear of drowning, turned back. The 17th-century historian Ahmed Mohammed al-Maqqari poignantly described Abd al-Rahman's reaction as he implored Yahya to keep going: "O brother! Come to me, come to me!" Yahya returned to the near shore, and was quickly dispatched by the horsemen. They cut off his head and left his body to rot. Al-Maqqari quotes earlier historians reporting that Abd al-Rahman was so overcome with fear that from the far shore he ran until exhaustion overcame him. Only he and Bedr were left to face the unknown. After barely escaping with their lives, Abd al-Rahman and Bedr continued south through Palestine, the Sinai, and then into Egypt. Abd al-Rahman had to keep a low profile as he traveled. It may be assumed that he intended to go at least as far as northwestern Africa (Maghreb), the land of his mother, which had been partly conquered by his Umayyad predecessors. The journey across Egypt would prove perilous. At the time, Abd al-Rahman ibn Habib al-Fihri was the semi-autonomous governor of Ifriqiya (roughly, modern Tunisia) and a former Umayyad vassal. The ambitious Ibn Habib, a member of the illustrious Fihrid family, had long sought to carve out Ifriqiya as a private dominion for himself. At first, he sought an understanding with the Abbasids, but when they refused his terms and demanded his submission, Ibn Habib broke openly with the Abbasids and invited the remnants of the Umayyad dynasty to take refuge in his dominions. Abd al-Rahman was only one of several surviving Umayyad family members to make their way to Ifriqiya at this time. But Ibn Habib soon changed his mind. He feared the presence of prominent Umayyad exiles in Ifriqiya, a family more illustrious than his own, might become a focal point for intrigue among local nobles against his own usurped powers. Around 755, believing he had discovered plots involving some of the more prominent Umayyad exiles in Kairouan, Ibn Habib turned against them. At the time, Abd al-Rahman and Bedr were keeping a low profile, staying in Kabylie, at the camp of a Nafza Berber chieftain friendly to their plight. Ibn Habib dispatched spies to look for the Umayyad prince. When Ibn Habib's soldiers entered the camp, the Berber chieftain's wife Tekfah hid Abd al-Rahman under her personal belongings to help him go unnoticed. Once they were gone, Abd al-Rahman and Bedr immediately set off westwards. In 755, Abd al-Rahman and Bedr reached modern-day Morocco near Ceuta. Their next step would be to cross the sea to al-Andalus, where Abd al-Rahman could not have been sure whether or not he would be welcomed. Following the Berber Revolt of the 740s, the province was in a state of confusion, with the Muslim community torn by tribal dissensions among the Arabs (the Qays–Yemeni feud) and racial tensions between the Arabs and Berbers. At that moment, the nominal ruler of al-Andalus, emir Yusuf ibn Abd al-Rahman al-Fihri—another member of the Fihrid family and a favorite of the old Arab settlers ("baladiyun"), mostly of south Arabian or "Yemeni" tribal stock—was locked in a contest with his vizier (and son-in-law) al-Sumayl ibn Hatim al-Kilabi, the head of the "Syrians"—the "shamiyum", drawn from the "junds" or military regiments of Syria, mostly of north Arabian Qaysid tribes—who had arrived in 742. Among the Syrian "junds" were contingents of old Umayyad clients, numbering perhaps 500, and Abd al-Rahman believed he might tug on old loyalties and get them to receive him. Bedr was dispatched across the straits to make contact. Bedr managed to line up three Syrian commanders—Ubayd Allah ibn Uthman and Abd Allah ibn Khalid, both originally of Damascus, and Yusuf ibn Bukht of Qinnasrin. The trio approached the Syrian arch-commander al-Sumayl (then in Zaragoza) to get his consent, but al-Sumayl refused, fearing Abd al-Rahman would try to make himself emir. As a result, Bedr and the Umayyad clients sent out feelers to their rivals, the Yemenite commanders. Although the Yemenites were not natural allies (the Umayyads are a Qaysid tribe), their interest was piqued. The emir Yusuf al-Fihri had proven himself unable to keep the powerful al-Sumayl in check and several Yemenite chieftains felt their future prospects were poor, whether in a Fihrid or Syrian-dominated Spain, so that they had a better chance of advancement if they hitched themselves to the glitter of the Umayyad name. Although the Umayyads did not have a historical presence in the region (no member of the Umayyad family was known to have ever set foot in al-Andalus before) and there were grave concerns about young Abd al-Rahman's inexperience, several of the lower-ranking Yemenite commanders felt they had little to lose and much to gain, and agreed to support the prince. Bedr returned to Africa to tell Abd al-Rahman of the invitation of the Umayyad clients in al-Andalus. Shortly thereafter, they set off with a small group of followers for Europe. When some local Berber tribesmen learned of Abd al-Rahman's intent to set sail for al-Andalus, they quickly rode to catch up with him on the coast. The tribesmen might have figured that they could hold Abd al-Rahman as hostage, and force him to buy his way out of Africa. He did indeed hand over some amount of dinars to the suddenly hostile local Berbers. Just as Abd al-Rahman launched his boat, another group of Berbers arrived. They also tried to obtain a fee from him for leaving. One of the Berbers held on to Abd al-Rahman's vessel as it made for al-Andalus, and allegedly had his hand cut off by one of the boat's crew. Abd al-Rahman landed at Almuñécar in al-Andalus, to the east of Málaga, in September 755; however, his landing site was unconfirmed. Upon landing in al-Andalus, Abd al-Rahman was greeted by clients Abu Uthman and Ibn Khalid and an escort of 300 cavalry. During his brief time in Málaga, he was able to amass local support quickly. Waves of people made their way to Málaga to pay respect to the prince they thought was dead, including many of the aforementioned Syrians. One famous story that persisted through history related to a gift Abd al-Rahman was given while in Málaga. The gift was a beautiful young slave girl, but Abd al-Rahman humbly returned her to her previous master. News of the prince's arrival spread like wildfire throughout the peninsula. During this time, emir al-Fihri and the Syrian commander al-Sumayl pondered what to do about the new threat to their shaky hold on power. They decided to try to marry Abd al-Rahman into their family. If that did not work, then Abd al-Rahman would have to be killed. Abd al-Rahman was apparently sagacious enough to expect such a plot. In order to help speed his ascension to power, he was prepared to take advantage of the feuds and dissensions. However, before anything could be done, trouble broke out in northern al-Andalus. Zaragoza, an important trade city on the Upper March of al-Andalus, made a bid for autonomy. Al-Fihri and al-Sumayl rode north to quash the rebellion. This might have been fortunate timing for Abd al-Rahman, since he was still getting a solid foothold in al-Andalus. By March 756, Abd al-Rahman and his growing following of Umayyad clients and Yemenite "junds", were able to take Sevilla without violence. He managed to break the rebellion attempt in Zaragoza, but just about that time the Cordovan governor received news of a Basque rebellion in Pamplona. An important detachment was sent by Yusuf ibn 'Abd al-Rahman to quash it, but his troops were annihilated. After the setback, al-Fihri turned his army back south to face the "pretender". The fight for the right to rule al-Andalus was about to begin. The two contingents met on opposite sides of the River Guadalquivir, just outside the capital of Córdoba on the plains of Musarah. The river was, for the first time in years, overflowing its banks, heralding the end of a long drought. Nevertheless, food was still scarce, and Abd al-Rahman's army suffered from hunger. In an attempt to demoralize Abd al-Rahman's troops, al-Fihri ensured that his troops not only were well fed, but also ate gluttonous amounts of food in full view of the Umayyad lines. An attempt at negotiations soon followed in which it is likely that Abd al-Rahman was offered the hand of al-Fihri's daughter in marriage and great wealth. Abd al-Rahman, however, would settle for nothing less than control of the emirate, and an impasse was reached. Even before the fight began, dissension spread through some of Abd al-Rahman's lines. Specifically, the Yemeni Arabs were unhappy that the prince was mounted on a fine Spanish steed and that his mettle was untried in battle. The Yemenis observed significantly that such a fine horse would provide an excellent mount to escape from battle. Being the ever-wary politician, Abd al-Rahman acted quickly to regain Yemeni support, and rode to a Yemeni chief who was mounted on a mule named "Lightning". Abd al-Rahman averred that his horse proved difficult to ride and was wont to buck him out of the saddle. He offered to exchange his horse for the mule, a deal to which the surprised chief readily agreed. The swap quelled the simmering Yemeni rebellion. Soon both armies were in their lines on the same bank of the Guadalquivir. Abd al-Rahman had no banner, and so one was improvised by unwinding a green turban and binding it round the head of a spear. Subsequently, the turban and the spear became the banner and symbol of the Andalusian Umayyads. Abd al-Rahman led the charge toward al-Fihri's army. Al-Sumayl in turn advanced his cavalry out to meet the Umayyad threat. After a long and difficult fight "Abd ar-Rahman obtained a most complete victory, and the field was strewn with the bodies of the enemy.". Both al-Fihri and al-Sumayl managed to escape the field (probably) with parts of the army too. Abd al-Rahman triumphantly marched into the capital, Córdoba. Danger was not far behind, as al-Fihri planned a counterattack. He reorganized his forces and set out for the capital Abd al-Rahman had usurped from him. Again Abd al-Rahman met al-Fihri with his army; this time negotiations were successful, although the terms were somewhat changed. In exchange for al-Fihri's life and wealth, he would be a prisoner and not allowed to leave the city limits of Córdoba. Al-Fihri would have to report once a day to Abd al-Rahman, as well as turn over some of his sons and daughters as hostages. For a while al-Fihri met the obligations of the one-sided truce, but he still had many people loyal to him--people who would have liked to see him back in power. Al-Fihri eventually did make another bid for power. He quit Córdoba and quickly started gathering supporters. While at large, al-Fihri managed to gather an army allegedly numbering 20,000. It is doubtful, however, that his troops were "regular" soldiers, but rather a hodge-podge of men from various parts of al-Andalus. Abd al-Rahman's appointed governor in Sevilla took up the chase, and after a series of small fights, managed to defeat al-Fihri's army. Al-Fihri himself managed to escape to the former Visigoth capital of Toledo in central al-Andalus; once there, he was promptly killed. Al-Fihri's head was sent to Córdoba, where Abd al-Rahman had it nailed to a bridge. With this act, Abd al-Rahman proclaimed himself the emir of al-Andalus. However, in order to take over southern Iberia, al-Fihri's general, al-Sumayl, had to be dealt with, and he was garroted in Córdoba's jail. Still, most of central and northern al-Andalus (Toledo, Zaragoza, Barcelona, etc.) was out of his rule, with large swathes remaining in the hands of Yusuf ibn 'Abd al-Rahman al-Fihri's supporters until 779 (submission of Zaragoza). It is unclear whether Abd al-Rahman proclaimed himself caliph. There are documents in the archives of Cordoba that state that this was his first act upon entering the city. He himself believed he was destined to be Caliph because of prophesies he had heard as a boy, so it seems likely he would. However, historically he is recorded as Emir and not Caliph. Abd al-Rahman's 7th descendant, Abd al-Rahman III, would, however, take up the title of caliph. In the meantime, a call went out through the Muslim world that al-Andalus was a safe haven for friends of the house of Umayya, if not for Abd al-Rahman's scattered family that managed to evade the Abbasids. Abd al-Rahman probably was quite happy to see his call answered by waves of Umayyad faithful and family. He was finally reacquainted with his son Sulayman, whom he last saw weeping on the banks of the Euphrates with his sisters. Abd ar-Rahman's sisters were unable to make the long voyage to al-Andalus. Abd al-Rahman placed his family members in high offices across the land, as he felt he could trust them more than non-family. The Umayyad family would again grow large and prosperous over successive generations. One of these kinsmen, Abd al-Malik ibn Umar ibn Marwan, persuaded Abd al-Rahman in 757 to drop the name of the Abbasid caliph from the Friday prayers (a traditional recognition of sovereignty in medieval Islam), and became one of his top generals and his governor in Seville. By 763 Abd ar-Rahman had to get back to the business of war. Al-Andalus had been invaded by an Abbasid army. Far away in Baghdad, the current Abbasid caliph, al-Mansur, had long been planning to depose the Umayyad who dared to call himself emir of al-Andalus. Al-Mansur installed al-Ala ibn-Mugith as governor of Africa (whose title gave him dominion over the province of al-Andalus). It was al-Ala who headed the Abbasid army that landed in al-Andalus, possibly near Beja (in modern-day Portugal). Much of the surrounding area of Beja capitulated to al-Ala, and in fact rallied under the Abbasid banners against Abd al-Rahman. Abd al-Rahman had to act quickly. The Abbasid contingent was vastly superior in size, said to have numbered 7,000 men. The emir quickly made for the redoubt of Carmona with his army. The Abbasid army was fast on their heels, and laid siege to Carmona for approximately two months. Abd al-Rahman must have sensed that time was against him as food and water became scarce, and his troops morale likely came into question. Finally Abd al-Rahman gathered his men as he was "resolved on an audacious sally". Abd al-Rahman hand-picked 700 fighters from his army and led them to Carmona's main gate. There, he started a great fire and threw his scabbard into the flames. Abd al-Rahman told his men that time had come to go down fighting rather than die of hunger. The gate lifted and Abd ar-Rahman's men fell upon the unsuspecting Abbasids, thoroughly routing them. Most of the Abbasid army was killed. The heads of the main Abbasid leaders were cut off, preserved in salt, identifying tags pinned to their ears, and then bundled together in a gruesome package and sent to the Abbasid caliph, who was on pilgrimage at Mecca. Upon receiving the evidence of al-Ala's defeat in al-Andalus, al-Mansur is said to have gasped, "God be praised for placing a sea between us!" Al-Mansur hated, and yet apparently respected Abd al-Rahman to such a degree that he dubbed him the "Hawk of Quraysh" (the Umayyads were from a branch of the Quraysh tribe). Despite such a tremendous victory, Abd al-Rahman had to continuously put down rebellions in al-Andalus. Various Arab and Berber tribes fought each other for varying degrees of power, some cities tried to break away and form their own state, and even members of Abd al-Rahman's family tried to wrest power from him. During a large revolt, dissidents marched on Córdoba itself; However, Abd al-Rahman always managed to stay one step ahead, and crushed all opposition; as he always dealt severely with dissidents in al-Andalus. However, this assumption needs to be put in perspective, since in 756 he held a limited number of southern strongholds and faced resistance from other towns short of outright revolt during the next 25 years. Despite all this turmoil in al-Andalus, Abd al-Rahman wanted to take the fight back east to Baghdad. Revenge for the massacre of his family at the hands of the Abbasids must surely have been the driving factor in Abd al-Rahman's war plans. However, his war against Baghdad was put on hold by more internal problems. The city of Zaragoza on the Upper March remained out of reach of the Umayyad leader since the time of Yusuf ibn 'Abd al-Rahman al-Fihri, bidding for autonomy. Little could Abd al-Rahman have known that, as he set off to settle matters in that northern city, his hopes of waging war against Baghdad would be indefinitely put on hold. Zaragoza proved to be a most difficult city to reign over for not only Abd ar-Rahman, but his successors as well. In the year 777–778, several notable men including Sulayman ibn Yokdan al-Arabi al-Kelbi, the self-appointed governor of Zaragoza, met with delegates of the leader of the Franks, Charlemagne. "[Charlemagne's] army was enlisted to help the Muslim governors of Barcelona and Zaragoza against the Umayyad [emir] in Cordoba..." Essentially Charlemagne was being hired as a mercenary, even though he likely had other plans of acquiring the area for his own empire. After Charlemagne's columns arrived at the gates of Zaragoza, Sulayman got cold feet and refused to let the Franks into the city, after his subordinate, al-Husayn ibn Yahiya, had successfully defeated and captured Abd al-Rahman's most trusted general, Thalaba Ibn Ubayd. It is possible that he realized that Charlemagne would want to usurp power from him. After capturing Sulayman, Charlemagne's force eventually headed back to France via a narrow pass in the Pyrenees, where his rearguard was wiped out by Basque and Gascon rebels (this disaster inspired the epic Chanson de Roland). Charlemagne was also attacked by Sulayman's relatives, who had freed Sulayman. Now Abd al-Rahman could deal with Sulayman and the city of Zaragoza without having to fight a massive Christian army. In 779 Abd al-Rahman offered Husayn, one of Sulayman's allies, the job of Zaragoza's governorship. The temptation was too much for al-Husayn, who murdered his colleague Sulayman. As promised, al-Husayn was awarded Zaragoza with the expectation that he would always be a subordinate of Córdoba. However, within two years al-Husayn broke off relations with Abd al-Rahman and announced that Zaragoza would be an independent city-state. Once again Abd al-Rahman had to be concerned with developments in the Upper March. He was intent on keeping this important northern border city within the Umayyad fold. By 783 Abd al-Rahman's army advanced on Zaragoza. It appeared as though Abd al-Rahman wanted to make clear to this troublesome city that independence was out of the question. Included in the arsenal of Abd al-Rahman's army were thirty-six siege engines. Zaragoza's famous white granite defensive walls were breached under a torrent of ordnance from the Umayyad lines. Abd al-Rahman's warriors spilled into the city's streets, quickly thwarting al-Husayn's desires for independence. After the aforementioned period of conflict, Abd al-Rahman continued in his improvement of al-Andalus' infrastructure. He ensured roadways were begun, aqueducts were constructed or improved, and that a new mosque was well funded in his capital at Córdoba. Construction on what would in time become the world-famous Great Mosque of Córdoba was started circa the year 786. Abd al-Rahman knew that one of his sons would one day inherit the rule of al-Andalus, but that it was a land torn by strife. In order to successfully rule in such a situation, Abd al-Rahman needed to create a reliable civil service and organize a standing army. He felt that he could not always rely on the local populace in providing a loyal army; and therefore bought a massive standing army consisting mainly of Berbers from North Africa as well as slaves from other areas. The total number of soldiers under his command was nearly 40,000. As was common during the years of Islamic expansion from Arabia, religious tolerance was practiced. Abd al-Rahman continued to allow Jews and Christians and other monotheistic religions to retain and practice their faiths, in exchange for a tribute tax. Abd al-Rahman's policy of taxing non-Muslims, often continued by later rulers, changed the religious dynamic of al-Andalus. Possibly because of excessive tribute taxes, "the bulk of the country's population must have become Muslim". However, other scholars have argued that though 80% of al-Andalus converted to Islam, it did not truly occur until near the 10th century. Christians more often converted to Islam than Jews although there were converted Jews among the new followers of Islam. There was a great deal of freedom of interaction among the groups: for example, Sarah, the granddaughter of the Visigoth king Wittiza, married a Muslim man and bore two sons who were later counted among the ranks of the highest Arab nobility. Abd al-Rahman I was able to forge a new Umayyad dynasty by standing successfully against Charlemagne, the Abbasids, the Berbers, and other Muslim Spaniards. There is some dispute as to whether what he created was an extension of the Umayyad Dynasty or a new Caliphate of Cordoba. Abd al-Rahman died c. 788 in his adopted city of Córdoba, and was supposedly buried under the site of the Mezquita. Abd al-Rahman's alleged favorite son was his choice for successor, and would later be known as Hisham I. Abd ar-Rahman's progeny would continue to rule al-Andalus in the name of the house of Umayyad for several generations, with the zenith of their power coming during the reign of Abd al-Rahman III. Abd al-Rahman was the son of Mu'awiya, son of Hisham, son of Abd al-Malik, according to Abd el-Wahid Merrakechi when reciting his ancestry. Abd al-Rahman's mother was a member of the Nafza Berbers with whom he found refuge after the murder of his family in 750. Abd al-Rahman married a woman named Hulal. She is said to be the mother of Hisham. Abd al-Rahman was the father of several sons, but the identity of their mother(s) is not clear: In his lifetime, Abd al-Rahman was known as "al Dakhil" ("the Entrant"), but he was also known as "Saqr Quraish" ("The Falcon of the Quraish"), bestowed on him by one of his greatest enemies, the Abbasid caliph al-Mansur. According to the chroniclers, al-Mansur once asked his courtiers who deserved the exalted title of "Falcon of the Quraysh" ("Saqr Quraish", foremost of the Prophet's tribe). The obsequious courtiers naturally replied "You, O Commander of the Faithful!", but the Caliph said no. Then they suggested Mu'awiya (founder of the Umayyad Caliphate), but the Caliph again said no. Then they suggested Abd al-Malik ibn Marwan (one of the greatest of the Umayyad caliphs), but again no. Then who, they asked, and al-Mansur replied:
https://en.wikipedia.org/wiki?curid=2676
Abd al-Rahman III Abd al-Raḥmān III, also known as ′Abd al-Rahmān ibn Muḥammad ibn ′Abd Allāh ibn Muḥammad ibn ′Abd al-Raḥmān ibn al-Ḥakam al-Rabdī ibn Hishām ibn ′Abd al-Raḥmān al-Dākhil () (889/91 - 961), was the Umayyad Emir of Córdoba from 912-929, at which point he founded the Caliphate of Córdoba, serving as its first Caliph until his death. ′Abd al-Raḥmān won the "laqab" (sobriquet) "al-Nasir li-Dīn Allāh"Defender of God's Faith in his early 20s when he supported the Maghrawa in North Africa against Fatimid expansion and rose to the Caliphate. His half-century reign (912–961) of al-AndalusMuslim Iberian Spain was known for its religious tolerance. Abd al-Rahman was born in Córdoba, on 7/11 January 889/91. He was the grandson of Abdullah ibn Muhammad al-Umawi, seventh independent Umayyad emir of al-Andalus. His parents were Abdullah's son Muhammad and Muzna (or Muzayna), a Christian concubine. His paternal grandmother was also a Christian, the royal "infanta" Onneca Fortúnez, daughter of the captive king Fortún Garcés of Pamplona. Abd al-Rahman was thus nephew in the half-blood of queen Toda of Pamplona. He is described as having "white skin, blue eyes and attractive face; good looking, although somewhat sturdy and stout. His legs were short, to the point that the stirrups of his saddle were mounted just one palm under it. When mounted, he looked tall, but on his feet he was quite short. He dyed his beard black." Muhammad was assassinated by his brother Al-Mutarrif, who had allegedly grown jealous of the favour Muhammad had gained in the eyes of their father Abdallah. Al-Mutarrif had accused Muhammad of plotting with the rebel Umar ibn Hafsun, and Muhammad had been imprisoned. According to some sources, the emir himself was behind Muhammad's fall, as well as Al-Mutarrif's death in 895. Abd al-Rahman spent his youth in his mother's harem. Al-Mutarrif's sister, known as "al-Sayyida" ("the Lady"), was entrusted with his education. She made sure that Abd al-Rahman's education was conducted with some rigor. Emir Abdallah died at the age of 72. Despite the fact that four of his sons (Aban, Abd al Rahman, Muhammad and Ahmad) were alive at the time of his death, all of them were passed over for succession. Abdallah instead chose as his successor his grandson, Abd al-Rahman III (the son of his first son). This came as no surprise, since Abdallah had already demonstrated his affection for his grandson in many ways, namely by allowing him to live in his own tower (something he did not allow for any of his sons), and allowing him to sit on the throne on some festive occasions. Most importantly Abdallah gave Abd al-Rahman his ring, the symbol of power, when Abdallah fell ill prior to his death. Abd al-Rahman succeeded Abdallah the day after his death, 16 October 912. Historiographers of the time, such as "Al-Bayan al-Mughrib" and the "Crónica anónima de Abd al-Rahman III", state that his succession was "without incident". At the time, Abd al-Rahman was about 21 or 22 years old. He inherited an emirate on the verge of dissolution, his power extending not far beyond the vicinity of Córdoba. To the north, the Christian Kingdom of Asturias was continuing its program of "Reconquista" in the Douro valley. To the south in Ifriqiya, the Fatimids had created an independent caliphate that threatened to attract the allegiance of the Muslim population, who had suffered under the harsh rule of Abdullah. On the internal front the discontented Muladi families (Muslims of Iberian origin) represented a constant danger for the Córdoban emir. The most powerful of the latter was Umar ibn Hafsun, who, from his impregnable fortress of Umar ibn Hafsun, controlled much of eastern Al-Andalus. From the very early stages of his reign, Abd al-Rahman showed a firm resolve to quash the rebels of al-Andalus, consolidate centralized power, and reestablish internal order within the emirate. Within 10 days of taking the throne, he exhibited the head of a rebel leader in Cordoba. From this point on he led annual expeditions against the northern and southern tribes to maintain control over them. To accomplish his aims he introduced into the court the "saqalibah", slaves of East European origin. The "saqalibah" represented a third ethnic group that could neutralize the endless strife between his subjects of Muslim Arab heritage, and those of Muslim Berber heritage. Hasdai ibn Shaprut, a Jewish courtier of the king's court who served as financier to the king, wrote of the king's revenues on this wise: During the first 20 years of his rule, Abd al-Rahman avoided military action against the northern Christian kingdoms, Asturias and the Kingdom of Navarre. The Muladi rebels were the first problem he confronted. Those powerful families were supported by Iberians who were openly or secretly Christians and had acted with the rebels. These elements, which formed the bulk of the population, were not averse to supporting a strong ruler who would protect them against the Arab aristocracy. Abd al-Rahman moved to subdue them by means of a mercenary army that included Christians. He first had to suppress the rebel Umar ibn Hafsun. On 1 January 913 an army, led by the eunuch Badr, conquered the fortress of Écija, at some from the capital. All the city's fortifications were destroyed, aside from the citadel, which was left as residence of the governor and a garrison for the emiral troops. In the following spring, after sixty-five days of meticulous preparations, Abd al-Rahman personally led an expedition to the south of his realm. His troops were able to recover the "coras" (provinces) of Jaén and Granada, while a cavalry detachment was sent to free Málaga from ibn Hafsun's siege. He also obtained the capitulation of Fiñana (in the modern province of Almería), after setting fire to its suburbs. Subsequently, he moved against the castle of Juviles in the Alpujarras. After devastating its countryside to deprive it of any resource, he encircled it. Finding it difficult to bombard with catapults, he ordered the construction of a platform where his siege engines could be mounted to greater effect, and cut the water supply. The Muladi defenders surrendered after a few days: their lives, apart from fifty-five die-hards who were beheaded, were spared in exchange for their allegiance to the emir. The campaign continued in a similar vein, lasting for a total of ninety days. Abd al-Rahman forced the defeated Muladi to send hostages and treasures to Córdoba, in order to secure their continued submission. In the first year of his reign, Abd al-Rahman took advantage of the rivalries between the Banu Hajjaj lords of Seville and Carmona to force them to submit. He initially sent a special corps ("hasam") under Ahmad ibn Muhammad ibn Hudayr, governor of Écija, to Seville, to obtain their submission. This attempt failed, but gained him the support of Muhammad ibn Ibrahim ibn Hayyay, lord of Carmona, and a cousin of the Sevillan lord, Ahmad ibn Maslama. When the latter was surrounded by Umayyad troops, he sued for help to Ibn Hafsun, but the latter was defeated by the besiegers and returned to Bobastro. He next went after the forts in the provinces of Elvira, Granada, and Jaén, all of which were either directly or indirectly controlled by Hafsun. Seville finally capitulated on 20 December 913. Ibn al-Mundhir al-Qurays, a member of the royal family, was named governor of the city, while the Lord of Carmona obtained the title of vizier. Muhammad ibn Ibrahim enjoyed his office for only a single day, for Abd al-Rahman soon discovered his collusion with the rebel governor of Carmona. Muhammad was sent to prison, where he later met his death. The region of Valencia submitted peacefully in 915. Abd al-Rahman's next objective was to quash the longstanding rebellion of Umar ibn Hafsun. His troops left Córdoba on 7 May 914 and, after a few days, encamped before the walls of Balda (identified with today's Cuevas de San Marcos). His cavalry ravaged the nearby woods and the countryside, while the rest of the troops moved to Turrus, a castle located in the present municipality of Algarinejo, which was surrounded within five days, while its environs were also devastated. The Umayyad army then moved to the citadel of ʿUmar ibn Hafsun, while the cavalry was sent to the castle of Sant Batir, which was abandoned by the defenders, allowing Abd al-Rahman's troops to secure a large booty. Then it was the turn of the castles of Olías and Reina. The latter fell after a violent fight, leaving the road open to the major city and provincial capital of Málaga, which he captured after one day. Abd al-Rahman then turned and followed the coast by Montemayor, near Benahavís, Suhayl (Fuengirola) and another castle called "Turrus" or "Turrus Jusayn" (identified by Évariste Lévi-Provençal as Ojén). He finally arrived at Algeciras on 1 June 914. He ordered a patrol of the coast to destroy the boats that supplied the citadel of Umar ibn Hafsun from the Maghreb. Many of them were captured and set afire in front of the emir. The rebellious castles near Algeciras surrendered as soon as the Cordoban army manifested itself. Abd al-Rahman launched three different campaigns against Ibn Hafsun (who died in 917) and his sons. Among them, Jaʿfar ibn Hafsun held the stronghold of Toledo. Abd al-Rahman ravaged the countryside around the city. Jafar, after two years of siege, escaped the city to ask for help in the northern Christian kingdoms. In the meantime Abd al-Rahman obtained the surrender of the city from its population, after promising them immunity, although 4,000 rebels escaped in a night sally. The city surrendered on 2 August 932, after a siege of two years. In 921 the Banu Muhallab of Guadix submitted, followed by those of Jerez de la Frontera and Cádiz, as well as the trading republic of Pechina (922). In 927, Abd al-Rahman also launched a campaign against the rebel Banu Qasi, but was forced to break it off by the intervention of Jimeno Garcés of Pamplona. The last of the sons of Ibn Hafsun to fall was Hafs, who commanded his powerful fortress of Umar ibn Hafsun. Surrounded by Abd al-Rahman's vizier, Said ibn al-Mundhir, who had ordered the construction of bastions around the city, he resisted the siege for six months, until he surrendered in 928 and had his life spared. The continued expeditions against the Hafsunids did not distract Abd al-Rahman III from the situation of other regions in al-Andalus, which recognized him only nominally, if not being in open revolt. Most of the loyal governors of the cities were in a weak position, such as the governor of Évora, who could not prevent the attack of the king of Galicia (future king of León), Ordoño II, who captured the city in the summer of 913, taking back a sizable booty and 4,000 prisoners and massacring many Muslims. In much of the eastern and western provinces, Abd al-Rahman's authority was completely unrecognized. The lord of Badajoz, Abd Allah ibn Muhammad, grandson of Abd al-Rahman ibn Marwan al-Yilliqi, not only fortified his city against a possible attack from Ordoño, but also acted in complete independence from Córdoba. To avoid the fall of Évora into the hands of the Berber groups of the region, he ordered the destruction of its defensive towers and lowered the walls, though a year later he decided to reconstruct it, giving its control to his ally Masud ibn Sa' dun al-Surunbaqi. The Algarve was dominated completely by a muladí coalition led by Saʿid ibn Mal, who had expelled the Arabs from Beja, and the lords of Ocsónoba, Yahya ibn Bakr, and of Niebla, Ibn Ufayr. Alcácer do Sal and Lisbon were under control of the Banu Dānis. The absence of royal authority enabled Ordoño II to easily campaign in this area, his main objective being the city of Mérida, in the summer of 915. Abd al-Rahman III did not send an army and only several local Berber "jefes" offered some resistance which was ineffective. In the next year, despite having defeated only some of the rebels, Abd al-Rahman III considered himself powerful enough to declare himself Caliph of Córdoba (16 January 929), effectively breaking his allegiance to, and ties with, the Fatimid and Abbasid caliphs. The caliphate was thought only to belong to the Emperor who ruled over the sacred cities of Mecca and Medina, and his ancestors had until then been content with the title of emir. But the force of this tradition had weakened over time; and the title increased Abd al-Rahman's prestige with his subjects, both in Iberia and Africa. He based his claim to the caliphate on his Umayyad ancestors who had held undisputed control of the caliphate until they were overthrown by the Abbasids. Abd al-Rahman's move made him both the political and the religious leader of all the Muslims in al-Andalus, as well as the protector of his Christian and Jewish subjects. The symbols of his new caliphal power were a scepter ("jayzuran") and the throne ("sarir"). In the mint he had founded in November 928, Abd al-Rahman started to mint gold dinars and silver dirhams, replacing the "al-Andalus" title with his name. In his new role as caliph, he achieved the surrender of Ibn Marwan of Badajoz in 930 as well as the surrender of the Banu Dānis of Alcácer do Sal. On the southern front, to counter the increasing Fatimid power in North Africa, abd al-Rahmad ordered the construction of a fleet based in Almeria. The caliph helped the Maghrawa Berbers conquer Melilla (927), Ceuta (931) and Tangiers (951), which accepted his suzerainty. However, he was unable to defeat Jawhar al-Siqilli of the Fatimids. In 951 he signed a peace with the new king of León, Ordoño III, in order to have a free hand against the Fatimids whose ships were harassing caliphal shipping in the Mediterranean and had even launched an assault against Almeria. Abd al-Rahman's force, led by prime minister Ahmad ibn Said, besieged the Fatimid port of Tunis, which bought its safety by paying a huge sum. In the end he was able to create a protectorate covering the northern and central Maghreb, supporting the Idrisid dynasty; the Caliphate's influence in the area disappeared after a Fatimid offensive in 958, after which abd al-Rahman kept only the strongholds of Ceuta and Tangiers. Even before al-Andalus was firmly under his rule, he had restarted the war against King Ordoño II of León, who had taken advantage of the previous troublesome situation to capture some boundary areas and menace the Umayyad territory. In 917 the then emir had sent a large army under his general Ahmad ibn Abi Abda against León, but this force was destroyed at the Battle of San Esteban de Gormaz in September of that year. Recognizing he had underestimated the power of Ordoño II, in 920 Abd al-Rahman mustered another powerful army to reclaim the territories lost after the previous campaign. He captured the forts of Osma and San Esteban de Gormaz. After defeating King Sancho Garcés I of Navarre and the king of León at Valdejunquera on 26 July, he penetrated into Navarre, overcoming Aragon by the classic route of the invasions from the south. Abd al-Rahman reached the Basque city of Pamplona, which was sacked and its cathedral church demolished. In 924 Abd al-Rahman felt obliged to avenge the massacre of Viguera castle perpetrated by King Sancho Ordóñez of Navarre one year earlier. He launched a counter offensive against Sancho in which Abd al-Rahman devastated a large area of Basque territory. The succession crisis which struck León after Ordoño II's death in the same year caused hostilities to cease until Ramiro II obtained the throne in 932; a first attempt by him to assist the besieged rebels in Toledo was repelled in 932, despite the Christian king capturing Madrid and scoring a victory at Osma. In 934, after reasserting supremacy over Pamplona and Álava, Abd al-Rahmad forced Ramiro to retreat to Burgos, and forced the Navarrese queen Toda, his aunt, to submit to him as a vassal and withdraw from direct rule as regent for her son García Sánchez I. In 937 Abd al-Rahman conquered some thirty castles in León. Next he turned to Muhammad ibn Hashim al-Tugib, governor of Zaragoza, who had allied with Ramiro but was pardoned after the capture of his city. Despite early defeats, Ramiro and García were able to crush the caliphal army in 939 at the Battle of Simancas, and almost kill Abd al-Rahman, due to treason by Arab elements in the caliph's army. After this defeat, Abd al-Rahman stopped taking personal command of his military campaigns. His cause was helped, however, by Fernán González of Castile, one of the Christian leaders at Simancas, who subsequently launched a sustained rebellion against Ramiro. The victory of Simancas enabled the Christian kingdom to maintain the military initiative in the peninsula until the defeat of Ramiro's successor, Ordoño III of León, in 956. However, they did not press this advantage as civil war broke out in the Christian territories. In 950 Abd al-Rahman received in Córdoba an embassy from count Borrell II of Barcelona, by which the northern county recognized caliphal supremacy in exchange for peace and mutual support. In 958, Sancho, the exiled king of León, King García Sánchez of Pamplona, and his mother Queen Toda all paid homage to Abd al-Rahman in Córdoba. Until 961, the caliphate played an active role in the dynastic strife characterizing the Christian kingdoms during the period. Ordoño III's half-brother and successor, Sancho the Fat, had been deposed by his cousin Ordoño IV. Together with his grandmother Toda of Pamplona, Sancho sought an alliance with Córdoba. In exchange for some castles, Abd al-Rahman helped them to take back Zamora (959) and Oviedo (960) and to overthrow Ordoño IV. Abd al-Rahman was accused of having sunk in his later years into the self-indulgent habits of the harem. He is known to have openly kept a male as well as a female harem. This likely influenced the polemical story of his sexual attraction for a 13-year-old boy (later enshrined as a Christian martyr and canonised as Saint Pelagius of Córdoba) who refused the Caliph's advances. This story of paedophilia may have been a construct on top of an original tale, however, in which he ordered the boy-slave to convert to Islam. Either way, enraged, he had the boy tortured and dismembered, thus contributing to the Christian perception of Muslim brutality. Abd al-Rahman spent the rest of his years in his new palace outside Córdoba. He died 15 October 961 and was succeeded by his son al-Hakam II. Abd al-Rahman was a great humanist and patron of arts, especially architecture. A third of his revenue sufficed for the ordinary expenses of government, a third was hoarded, and a third was spent on buildings. After declaring the caliphate, he had a massive palace complex, known as the Medina Azahara, built some five kilometers north of Córdoba. The Medina Azahara was modeled after the old Umayyad palace in Damascus and served as a symbolic tie between the new caliph and his ancestors. It was said that Córdoba contained 3000 mosques and 100,000 shops and homes during his reign. Under his reign, Córdoba became the most important intellectual centre of Western Europe. He expanded the city's library, which would be further enriched by his successors. He also reinforced the Iberian fleet, which became the most powerful in Mediterranean Europe. Iberian raiders moved up to Galicia, Asturias, and North Africa. The colonizers of Fraxinetum came from al-Andalus as well. Due to his consolidation of power, Muslim Iberia became a power for a few centuries. It also brought prosperity, and with this he created mints where pure gold and silver coins were created. He renovated and added to the Mosque–Cathedral of Córdoba. He was very wary of losing control and kept tight reins in his family. In 949, he executed one of his sons for conspiring against him. He was tolerant of non-Muslims, Jews and Christians who were treated fairly. European nations sent emissaries such as from Otto I of Germany, and the Byzantine emperor. Abd al-Rahman III's mother Muzna was a Christian captive, possibly from the Pyrenean region. His paternal grandmother Onneca Fortúnez was a Christian princess from the Kingdom of Pamplona. In his immediate ancestry, Abd al-Rahman III was Arab and Hispano–Basque. Abd al-Rahman has a prominent place in the novel "The Long Ships", by the Swedish writer Frans G. Bengtsson. He and the society of his realm are depicted from the point of view of Danish Vikings who are captured by Abd al-Rahman's navy, reduced to slavery, later promoted to mercenary soldiers and finally escaping back to Denmark. In "The Eastern King, a Pilgrim's Tale" from Letitia Elizabeth Landon's "The Golden Violet", she invokes Abd al-Rahman's saying: And yet amid all that fell to my share, How much was weariness, how much was care! I numbered years of pain and distress, And but fourteen days of happiness.
https://en.wikipedia.org/wiki?curid=2678
Abdulaziz Abdulaziz ( / "ʻAbdü'l-ʻAzīz", ; 8 February 18304 June 1876) was the 32nd Sultan of the Ottoman Empire and reigned between 25 June 1861 and 30 May 1876. He was the son of Sultan Mahmud II and succeeded his brother Abdulmejid I in 1861. Born at Eyüp Palace, Constantinople (present-day Istanbul), on 8 February 1830, Abdülaziz received an Ottoman education but was nevertheless an ardent admirer of the material progress that was made in the West. He was the first Ottoman Sultan who travelled to Western Europe, visiting a number of important European capitals including Paris, London and Vienna in the summer of 1867. Apart from his passion for the Ottoman Navy, which had the world's third largest fleet in 1875 (after the British and French navies), the Sultan took an interest in documenting the Ottoman Empire. He was also interested in literature and was a talented classical music composer. Some of his compositions, together with those of the other members of the Ottoman dynasty, have been collected in the album "European Music at the Ottoman Court" by the London Academy of Ottoman Court Music. He was deposed on grounds of mismanaging the Ottoman economy on 30 May 1876, and was found dead six days later under unnatural and mysterious circumstances. His parents were Mahmud II and Pertevniyal Sultan (1812–1883), originally named Besime, a Circassian. In 1868 Pertevniyal was residing at Dolmabahçe Palace. That year Abdülaziz took the visiting Eugénie de Montijo, Empress of France, to see his mother. Pertevniyal considered the presence of a foreign woman within her private quarters of the seraglio to be an insult. She reportedly slapped Eugénie across the face, which almost caused an international incident. According to another account, Pertevniyal was outraged by the forwardness of Eugénie in taking the arm of one of her sons while he gave a tour of the palace garden, and she gave the Empress a slap on the stomach as a possibly more subtly intended reminder that they were not in France. The Pertevniyal Valide Sultan Mosque was built under the patronage of his mother. The construction work began in November 1869 and the mosque was finished in 1871. His paternal grandparents were Sultan Abdul Hamid I and Sultana Nakşidil Sultan. Several accounts identify his paternal grandmother with Aimée du Buc de Rivéry, a cousin of Empress Joséphine. Pertevniyal was a sister of Khushiyar Qadin, third wife of Ibrahim Pasha of Egypt. Khushiyar and Ibrahim were the parents of Isma'il Pasha. Between 1861 and 1871, the Tanzimat reforms which began during the reign of his brother Abdulmejid I were continued under the leadership of his chief ministers, Mehmed Fuad Pasha and Mehmed Emin Âli Pasha. New administrative districts ("vilayets") were set up in 1864 and a Council of State was established in 1868. Public education was organized on the French model and Istanbul University was reorganised as a modern institution in 1861. He was also integral in establishing the first Ottoman civil code. Abdülaziz cultivated good relations with the Second French Empire and the British Empire. In 1867 he was the first Ottoman sultan to visit Western Europe; his trip included a visit to the Exposition Universelle (1867) in Paris and a trip to the United Kingdom, where he was made a Knight of the Garter by Queen Victoria and shown a Royal Navy Fleet Review with Ismail Pasha. He travelled by a private rail car, which today can be found in the Rahmi M. Koç Museum in Istanbul. His fellow Knights of the Garter created in 1867 were Charles Gordon-Lennox, 6th Duke of Richmond, Charles Manners, 6th Duke of Rutland, Henry Somerset, 8th Duke of Beaufort, Prince Arthur, Duke of Connaught and Strathearn (a son of Queen Victoria), Franz Joseph I of Austria and Alexander II of Russia. Also in 1867, Abdülaziz became the first Ottoman Sultan to formally recognize the title of Khedive (Viceroy) to be used by the Vali (Governor) of the Ottoman Eyalet of Egypt and Sudan (1517–1867), which thus became the autonomous Ottoman Khedivate of Egypt and Sudan (1867–1914). Muhammad Ali Pasha and his descendants had been the governors (Vali) of Ottoman Egypt and Sudan since 1805, but were willing to use the higher title of Khedive, which was unrecognized by the Ottoman government until 1867. In return, the first Khedive, Ismail Pasha, had agreed a year earlier (in 1866) to increase the annual tax revenues which Egypt and Sudan would provide for the Ottoman treasury. Between 1854 and 1894, the revenues from Egypt and Sudan were often declared as a surety by the Ottoman government for borrowing loans from British and French banks. After the Ottoman government declared a sovereign default on its foreign debt repayments on 30 October 1875, which triggered the Great Eastern Crisis in the empire's Balkan provinces that led to the devastating Russo-Turkish War (1877–78) and the establishment of the Ottoman Public Debt Administration in 1881, the importance for Britain of the sureties regarding the Ottoman revenues from Egypt and Sudan increased. Combined with the much more important Suez Canal which was opened in 1869, these sureties were influential in the British government's decision to occupy Egypt and Sudan in 1882, with the pretext of helping the Ottoman-Egyptian government to put down the ʻUrabi Revolt (1879–1882). Egypt and Sudan (together with Cyprus) nominally remained Ottoman territories until 5 November 1914, when the British Empire declared war against the Ottoman Empire during World War I. In 1869, Abdülaziz received visits from Eugénie de Montijo, Empress consort of Napoleon III of France and other foreign monarchs on their way to the opening of the Suez Canal. The Prince of Wales, the future Edward VII, twice visited Istanbul. By 1871 both Mehmed Fuad Pasha and Mehmed Emin Âli Pasha were dead. The Second French Empire, his Western European model, had been defeated in the Franco-Prussian War by the North German Confederation under the leadership of the Kingdom of Prussia. Abdülaziz turned to the Russian Empire for friendship, as unrest in the Balkan provinces continued. In 1875, the Herzegovinian rebellion was the beginning of further unrest in the Balkan provinces. In 1876, the April Uprising saw insurrection spreading among the Bulgarians. Ill feeling mounted against Russia for its encouragement of the rebellions. While no one event led to his being deposed, the crop failure of 1873 and his lavish expenditures on the Ottoman Navy and on new palaces which he had built, along with mounting public debt, helped to create an atmosphere conducive to his being overthrown. Abdülaziz was deposed by his ministers on 30 May 1876. Abdülaziz's death at Çırağan Palace in Istanbul a few days later was documented as a suicide at the time, but suspicions of murder promptly erupted. A claim states that on the morning of June 5, Abdülaziz asked for a pair of scissors with which to trim his beard. Shortly after this he was found dead in a pool of blood flowing from two wounds in his arms. His body wanted to be examined by 17 physicians, but they were only allowed to examine his wrists: "Dr. Marco, Nouri, A. Sotto, Physician attached to the Imperial and Royal Embassy of Austria‐Hungary; Dr. Spagnolo, Marc Markel, Jatropoulo, Abdinour, Servet, J. de Castro, A. Marroin, Julius Millingen, C. Caratheodori; E. D. Dickson, Physician of the British Embassy; Dr. O. Vitalis, Physician of the Sanitary Board; Dr. E. Spadare, J. Nouridjian, Miltiadi Bey, Mustafa, Mehmed" who certified that the death had been “caused by the loss of blood produced by the wounds of the blood‐vessels at the joints of the arms” and that “the direction and nature of the wounds, together with the instrument which is said to have produced them, lead us to conclude that suicide had been committed.” One of those physicians also stated that “His skin was very pale, and entirely free from bruises, marks or spots of any kind whatever. There was no lividity of the lips indicating suffocation nor any sign of pressure having been applied to the throat.” These claims turned out to be false, as a photograph surfaced in the book "Hatıra-i Uhuvvet: Portre Fotoğraflarının Cazibesi 1846-1950" by Bahattin Öztuncay’, which revealed the sultan's last taken picture after being deposed: The royal sultan is sitting on a chair in regular street clothes with two palace servants casually leaning on his shoulders - one with the Masonic hidden hand gesture - while the sultan is staring solemnly. Following Sultan Abdülaziz's dethronement, he was forced into a room at Topkapi Palace. This room happened to be the same room that Sultan Selim III was murdered in. Six men entered the room and had a hard time overpowering Sultan Abdülaziz as he was a wrestler too. He was left with bruises on his body. Then they cut both of his wrists, wrapped him and left him to bleed to death. In Sultan Abdulhamid II's recently surfaced memoirs, the event is described as an assassination by the order of Hüseyin Avni Pasha and Midhat Pasha. Hüseyin Avni Pasha was used by the French and Midhat Pasha was used by the English to topple Sultan Abdulaziz. According to the memoirs, Sultan Murad V began to show signs of paranoia, madness, and continuous fainting and vomiting even on the day of his coronation, and threw himself into a pool yelling at his guards to protect his life. They were afraid the public would become outraged and revolt to bring the former Sultan back on the throne. Within a few days, on 4 June 1876, they arranged for Sultan Abdülaziz to kill himself with beard scissors, and cut his wrists.
https://en.wikipedia.org/wiki?curid=2681
Abdera, Thrace Abdera is a municipality in the Xanthi Prefecture of Thrace, Greece. In classical antiquity, it was a major Greek "polis" on the Thracian coast. The ancient polis is to be distinguished from the municipality, which was named in its honor. The polis lay 17 km east-northeast of the mouth of the Nestos River, almost directly opposite the island of Thasos. It was a colony placed in previously unsettled Thracian territory, not then a part of Hellas, during the age of Greek colonization. The city that developed from it became of major importance in ancient Greece. After the 4th century it declined, contracted to its acropolis, and was abandoned, never to be reoccupied except by archaeologists. Meanwhile, life went on as the changing population settled in other communities in the region. One named Polystylus changed its name to Abdera. In 2011 the municipality of Abdera was synoecized from three previous municipalities comprising a number of modern settlements. The ancient site remains in it as a ruin. The municipality of Abdera has 19,005 inhabitants (2011). The seat of the municipality is the town Genisea. The name "Abdera" is of Phoenician origin and was shared in antiquity by Abdera, Spain and a town near Carthage in North Africa. It was variously Hellenized as ("Ábdēra"), ("Aúdēra"), ("Ábdara"), ("Ábdēron"), and ("Ábdēros"), before being Latinized as "Abdera". Greek legend attributed the name to an eponymous Abderus who fell nearby and was memorialized by Hercules's founding of a city at the location. The present-day town is written Avdira () and pronounced in modern Greek. The Phoenicians apparently began the settlement of Abdera at some point before the mid-7th century and the town long maintained Phoenician standards in its coinage. The Greek settlement was begun as a failed colony from Klazomenai, traditionally dated to 654. (Evidence in 7th-century- Greek pottery tends to support the traditional date but the exact timing remains uncertain.) Herodotus reports that the leader of the colony had been Timesios but, within his generation, the Thracians had expelled the colonists. Timesios was subsequently honored as a local protective spirit by the later Abderans from Teos. Others recount various legends about this colony. Plutarch and Aelian relate that Timesios grew insufferable to his colonists because of his desire to do everything by himself; when one of their children let him know how they all really felt, he quit the settlement in disgust; modern scholars have tried to split the difference between the two accounts of early Abdera's failure by giving the latter as the reason for Timesios's having left Klazomenai. The successful foundation occurred in 544BC, when the majority of the people of Teos (including the poet Anacreon) migrated to Abdera to escape the Persian yoke. The chief coin type, a "griffon", is identical with that of Teos; the rich silver coinage is noted for the beauty and variety of its reverse types.
https://en.wikipedia.org/wiki?curid=2685
Abdul Hamid I Abdülhamid I, Abdul Hamid I or Abd Al-Hamid I (, "`Abdü’l-Ḥamīd-i evvel"; ; 20 March 1725 – 7 April 1789) was the 27th Sultan of the Ottoman Empire, reigning over the Ottoman Empire from 1774 to 1789. He was born on 20 March 1725 in Constantinople, a younger son of Sultan Ahmed III (reigned 1703–1730) and his consort Şermi Kadın. Ahmed III abdicated in favor of his nephew Mahmud I, who was succeeded by his brother Osman III, and Osman by Ahmed's elder son Mustafa III. As a potential heir to the throne, Abdul Hamid was imprisoned in comfort by his cousins and older brother, as was customary. This lasted until 1767. During this period, he received his early education from his mother Rabia Şermi, who taught him history and calligraphy. On the day of Mustafa's death on 21 January 1774 , Abdul Hamid ascended to the throne with a ceremony held in the palace. The next day Mustafa III's funeral procession was held. The new sultan sent a letter to the Grand Vizier and Serdar-ı Ekrem Muhsinzade Mehmed Pasha on the front and informed him to continue his duty. On 27 January 1774, the sword was armed in Eyup Sultan. At the time, the Ottoman-Russian front wars continued, the army was at once, and there was a shortage of food in Istanbul. Abdul Hamid's long imprisonment had left him indifferent to state affairs and malleable to the designs of his advisors. Yet he was also very religious and a pacifist by nature. At his accession the financial straits of the treasury were such that the usual donative could not be given to the Janissary Corps. The new Sultan told the Janissaries "There are no longer gratuities in our treasury, as all of our soldier sons should learn." Abdul Hamid now sought to reform the Empire's armed forces. He enumerated the Janissary corps and tried to renovate it, and also the navy. He established a new artillery corps. He was also credited with the creation of the Imperial Naval Engineering School. Abdul Hamid tried to strengthen Ottoman rule over Syria, Egypt, and Iraq. However, slight successes against rebellions in Syria and the Morea could not compensate for the loss of the Crimean Peninsula, which had become nominally independent in 1774, but was in practice now controlled by Russia. Russia repeatedly exploited its position as protector of Eastern Christians to interfere in the Ottoman Empire, and explicitly. Finally the Ottomans declared war against Russia in 1787. Austria soon joined Russia. Turkey held its own in the conflict, at first, but on 6 December 1788, Ochakov fell to Russia (all of its inhabitants being massacred). It is said that this sad defeat broke Abdul Hamid's spirit, as he died four months later. In spite of his failures, Abdul Hamid was regarded as the most gracious Ottoman Sultan. He personally directed the fire brigade during the Constantinople fire of 1782. He was admired by the people for his religious devotion, and was even called a "Veli" ("saint"). He also outlined a reform policy, supervised the government closely, and worked with statesmen. Abdul Hamid I turned to internal affairs after the war with Russia ended in this way. He tried to suppress internal revolts through Algerian Gazi Hasan Pasha, and to regulate the reform works through Silâhdar Seyyid Mehmed Pasha (Karavezir) and Halil Hamid Pasha. Especially in Syria, the rebellion of Zahir al-Omar, who cooperated with the admirals of the Russian navy in the Mediterranean, benefiting from the confusion caused by the Russian expedition of the 1768 Russian campaign, and suppressed the rebellion in Egypt in 1775, as well as the Kölemen who were in rebellion in Egypt. was brought to the road. On the other hand, the confusion in Peloponnese was ended, and calm was achieved. Kaptanıderyâ Gazi Hasan Pasha and Cezzâr Ahmed Pasha played an important role in suppressing all these events. Despite his pacific inclinations, the Ottoman Empire was forced to renew the ongoing war with Russia almost immediately. This led to complete Turkish defeat at Kozludzha and the humiliating Treaty of Küçük Kaynarca, signed on 21 July 1774. The Ottomans ceded territory to Russia, and also the right to intervene on behalf of the Orthodox Christians in the Empire. With the Treaty of Küçük Kaynarca, the territory left, as well as Russia's ambassador at Istanbul level and an authorized representative, this ambassador's participation in other ceremonies at the state ceremonies, the right to pass through the Straits to Russia, as the envoys of the Russian envoy were given immunity. Marketing opportunities for all kinds of commodities in Istanbul and other ports, as well as the full commercial rights of England and France were given. It was also in the treaty that the Russian state had a church built in Ga lata. Under the circumstances, this church would be open to the public, referred to as the Russo-Greek Church, and forever under the protection of Russian ambassadors in Istanbul. In 1789, Tipu Sultan, ruler of the Sultanate of Mysore sent an embassy to Abdul Hamid, urgently requesting assistance against the British East India Company, and proposed an offensive and defensive alliance. Abdul Hamid informed the Mysori ambassadors that the Ottomans were still entangled and exhausted from the ongoing war with Russia and Austria. Abdul Hamid I left many architectural works, mostly in Istanbul. The most important of these is the current in Sirkeci in 1777. It is the building built by the Vakıf Inn. He built a fountain, an elementary school, a madrasah and a library next to this building. The books in the library are kept in the Süleymaniye Library today and the madrasah is used as a stock exchange building. During the construction of the Vakıf Inn, the fountain was removed by construction and the fountain was transferred to the corner of Zeynep Sultan Mosque opposite Gülhane Park. In addition to these works, in 1778, he built a mosque with a mosque on the beach of Beylerbeyi on behalf of Râbia Şermi Kadın, and built fountains in Çamlıca Kısıklı Square, other than places such as İskele Square, Çınarönü, Havuzbaşı and Car Square. In addition, he had the Beylerbeyi Mosque substantially repaired, built a mosque, a fountain and a bath and shops around Emirgi in Emirgân in 1783, and another one for Hümâşah Sultan and his son Mehmed. The fountain has been built more. In addition to these, there is a fountain next to Neslişah Mosque in Istinye, and another fountain on the embankment between Dolmabahçe and Kabataş. He wrote down the troubles he saw before, to the grand vizier or to the governor of his empire. He accepted the invitations of the and his grand vizier and went to his mansions, followed by the reading of Quran. He was humble and a religious Sultan. It is known that Abdul Hamid I was fond of his children, was interested in family life, spent the summer months in Karaağaç, Beşiktaş with his consorts, daughters and sons. His daughter Esma Sultan's dressing styles, her passion for entertainment, her journey to the objects with her journeymen and concubines have set an example for Istanbul ladies. Abdul Hamid had nine wives: His sons were: His daughters were: Abdul Hamid had adopted a daughter when he was a prince: Abdul Hamid died on 7 April 1789, at the age of sixty-four, in Constantinople. He was buried in Bahcekapi, a tomb he had built for himself. He bred Arabian horses with great passion. One breed of Küheylan Arabians was named "Küheylan Abdülhamid" after him. [aged 64]
https://en.wikipedia.org/wiki?curid=2692
Abdur Rahman Khan Abdur Rahman Khan () (between 1840 and 1844October 1, 1901) was Emir of Afghanistan from 1880 to 1901. He is known for uniting the country after years of internal fighting and negotiation of the Durand Line Agreement with British India. Abdur Rahman Khan was the first child and only son of Mohammad Afzal Khan, and grandson of Dost Mohammad Khan. Abdur Rahman Khan re-established the writ of the Afghan government after the disarray that followed the second Anglo-Afghan war. He became known as "The Iron Amir" because his government was a military despotism resting upon a well-appointed army administered through officials absolutely subservient to an inflexible will and controlled by a widespread system of espionage; and for his crushing of a number of rebellions by various tribes who were led by his relatives. Before his death in Herat, on June 9, 1863, Abdur Rahman's grandfather, Dost Mohammad Khan, nominated his third son, Sher Ali Khan, as his successor, passing over the two elder brothers, Afzal Khan and Azam Khan. At first, the new Amir was quietly recognized. But after a few months, Afzal Khan raised an insurrection in the north of the country, where he had been governing when his father died. This began a fierce internecine conflict for power between Dost Mohammad's sons, which lasted for nearly five years. The Musahiban are descendants of Dost Mohammad Khan's older brother, Sultan Mohammad Khan. Described by the American scholar and explorer Eugene Schuyler as "a tall well-built man, with a large head, and a marked Afghan, almost Jewish, face", Abdur Rahman distinguished himself for his ability and energetic daring. Although his father, Afzal Khan came to terms with the Amir Sher Ali, Abdur's behaviour in the northern province soon excited the Amir's suspicion and, when he was summoned to Kabul, fled across the Oxus into Bukhara. Sher Ali threw Afzal Khan into prison, and a revolt followed in southern Afghanistan. The Amir had scarcely suppressed it by winning a desperate battle when Abdur Rahman's reappearance in the north was a signal for a mutiny by troops stationed in those parts and a gathering of armed bands to his standard. After some delay and desultory fighting, he and his uncle, Azam Khan, occupied Kabul in March 1866. The Amir Sher Ali marched up against them from Kandahar; but in the battle that ensued at Sheikhabad on May 10, he was deserted by a large body of his troops, and after his signal defeat Abdur Rahman released his father, Afzal Khan, from prison in Ghazni, and installed him upon the throne as Amir of Afghanistan. Notwithstanding the new Amir's incapacity, and some jealousy between the real leaders, Abdur Rahman and his uncle, they again routed Sher Ali's forces, and occupied Kandahar in 1867. When Afzal Khan died at the end of the year, Azam Khan became the new ruler, with Abdur Rahman installed as governor in the northern province. But towards the end of 1868, Sher Ali's return and a general rising in his favour resulted in Abdur Rahman and Azam Khan's defeat at "Tinah Khan" on January 3, 1869. Both sought refuge to the east in Central Asia, where Abdur Rahman placed himself under Russian protection at Samarkand. Azam died in Kabul in October 1869. Abdur Rahman lived in exile in Tashkent. He was being told to cross the Oxus and claim throne for Amir. In March 1880, a report reached India that Abdur Rahman was in northern Afghanistan; and the Governor-General, Lord Lytton, opened communications with him to the effect that the British government were prepared to withdraw their troops, and to recognize Abdur Rahman as Amir of Afghanistan, with the exception of Kandahar and some districts adjacent to it. After some negotiations, and an interview with Lepel Griffin, the diplomatic representative at Kabul of the Indian government. Griffin described Abdur Rahman as a man of middle height, with an exceedingly intelligent face and frank and courteous manners, shrewd and able in conversation on the business in hand. At the "durbar" on July 22, 1880, Abdur Rahman was officially recognized as Amir, granted assistance in arms and money, and promised, in case of unprovoked foreign aggression, such further aid as might be necessary to repel it, provided that he align his foreign policy with the British. The British evacuation of Afghanistan was settled on the terms proposed, and in 1881, the British troops also handed over Kandahar to the new "Amir". However, Ayub Khan, one of Sher Ali Khan's sons, marched upon that city from Herat, defeated Abdur Rahman's troops, and occupied the place in July 1880. This serious reverse roused the "Amir", who had not displayed much activity. Instead, Ayub Khan was defeated in Kandahar by the British General Frederick Roberts on 1 September 1880. Ayub Khan was forced to flee into Persia. From that time Abdur Rahman was fairly seated firm on the throne at Kabul, thanks to the unwavering British protections in terms of giving large supplies of arms and money. In the course of the next few years, Abdul Rahman consolidated his grip over all Afghanistan, suppressing insurrection by a relentless and brutal use of his despotic authority. The powerful Ghilzai revolted against the severity of his measures several times. In that same year, Ayub Khan made a fruitless inroad from Persia. In 1885, at the moment when the "Amir" was in conference with the British viceroy, Lord Dufferin, in India, the news came of a skirmish between Russian and Afghan troops at Panjdeh, over a disputed point in the demarcation of the northwestern frontier of Afghanistan. Abdur Rahman's attitude at this critical juncture is a good example of his political sagacity. To one who had been a man of war from his youth, who had won and lost many fights, the rout of a detachment and the forcible seizure of some debatable frontier lands was an untoward incident; but it was not a sufficient reason for calling upon the British, although they had guaranteed his territory's integrity, to vindicate his rights by hostilities which would certainly bring upon him a Russian invasion from the north, and would compel his British allies to throw an army into Afghanistan from the southeast. He also published his autobiography in 1885, which served more as an advice guide for princes than anything else. His interest lay in keeping powerful neighbours, whether friends or foes, outside his kingdom. He knew this to be the only policy that would be supported by the Afghan nation; and although for some time a rupture with Russia seemed imminent, while the Government of India made ready for that contingency, the Amir's reserved and circumspect tone in the consultations with him helped to turn the balance between peace and war, and substantially conduced towards a pacific solution. Abdur Rahman left on those who met him in India the impression of a clear-headed man of action, with great self-reliance and hardihood, not without indications of the implacable severity that too often marked his administration. His investment with the insignia of the highest grade of the Order of the Star of India appeared to give him much pleasure. His adventurous life, his forcible character, the position of his state as a barrier between the Indian and the Russian empires, and the skill with which he held the balance in dealing with them, combined to make him a prominent figure in contemporary Asian politics and will mark his reign as an epoch in the history of Afghanistan. The Amir received an annual subsidy from the British government of 1,850,000 rupees. He was allowed to import munitions of war. He succeeded in imposing an organized government upon the fiercest and most unruly population in Asia; he availed himself of European inventions for strengthening his armament, while he sternly set his face against all innovations which, like railways and telegraphs, might give Europeans a foothold within his country. In 1893 he built himself a summer house called the Bagh-e Bala Palace. The Amir found himself unable, by reason of ill-health, to accept an invitation from Queen Victoria to visit England; but his second son Nasrullah Khan went instead. In 1893, Mortimer Durand was deputed to Kabul by the government of British India for this purpose of settling an exchange of territory required by the demarcation of the boundary between northeastern Afghanistan and the Russian possessions, and in order to discuss with Amir Abdur Rahman Khan other pending questions. Abdur Rahman Khan showed his usual ability in diplomatic argument, his tenacity where his own views or claims were in debate, with a sure underlying insight into the real situation. In the agreement that followed relations between the British Indian and Afghan governments, as previously arranged, were confirmed; and an understanding was reached upon the important and difficult subject of the border line of Afghanistan on the east, towards India. A Royal Commission was set up to determine the boundary between Afghanistan and British Governed India was tasked to negotiate terms for agreeing to the Durand Line, between the two parties camped at Parachinar, now part of FATA Pakistan, which is near Khost, Afghanistan. From the British side the camp was attended by Mortimer Durand and Sahibzada Abdul Qayyum, Political Agent Khyber. Afghanistan was represented by Sahibzada Abdul Latif and the Governor Sardar Shireendil Khan representing Amir Abdur Rahman Khan. In 1893, Mortimer Durand negotiated with Abdur Rahman Khan the "Durand Line Treaty" for the demarcation of the frontier between Afghanistan, the FATA, North-West Frontier Province and Baluchistan Provinces of Pakistan the successor state of British India. In 1905, Amir Habibullah Khan signed a new agreement with the United Kingdom which confirmed the legality of Durand Line. Similarly, the legality durand line was once again confirmed by King Amanullah Khan through the treaty of Rawalpindi in 1919. Durand Line was once again recognised as international border between Pakistan and Afghanistan by Sardar Mohammed Daoud Khan (former prime minister and later president of Afghanistan) during his visit to Pakistan in August 1976. Abdur Rahman Khan's government was a military despotism resting upon a well-appointed army; it was administered through officials absolutely subservient to an inflexible will and controlled by a widespread system of espionage; while the exercise of his personal authority was too often stained by acts of unnecessary cruelty. He held open courts for the receipt of petitioners and the dispensation of justice; and in the disposal of business he was indefatigable. In the 1880s, the "Iron Emir" decided to strategically displace some members of different ethnic groups in order to bring better security. For example, he "uprooted troublesome Durrani and Ghilzai Pashtun tribes and transported them to Uzbek and Tajik populated areas in the north, where they could spy on local Dari-speaking, non-Pashtun ethnic groups and act as a screen against further Russian encroachments on Afghan territory." From the end of 1888, the Amir spent eighteen months in his northern provinces bordering upon the Oxus, where he was engaged in pacifying the country that had been disturbed by revolts, and in punishing with a heavy hand all who were known or suspected to have taken any part in rebellion. In 1895–1896, Abdur Rahman directed the invasion of Kafiristan and the conversion of its indigenous peoples to Islam. The region was subsequently renamed Nuristan. In 1896, he adopted the title of "Zia-ul-Millat-Wa-ud Din" ("Light of the nation and religion"), and his zeal for the cause of Islam induced him to publish treatises on "jihad". Chitral, Yarkand and Ferghana became shelters for refugees in 1887 and 1883 from Badakhshan who fled from the campaigns of Abdul Rahman. In the early 1890s some Hazara tribes revolted against Abdur Rahman. In 1888, the Amir's cousin, Ishak Khan, rebelled against him in the north; and also the Ghilzais (Hotakis, Tokhis and Andaris) revolted against him in 1887; but these enterprises came to nothing. As the Kabul Newsletters written by the British agents indicate, Abdul Rahman was an extremely ruthless man. He has been called 'The Dracula Amir' by some writers According to Syed Askar Mousavi, "thousands of Hazara men, women, and children were sold as slaves in the markets of Kabul and Qandahar, while numerous towers of human heads were made from the defeated rebels as a warning to others who might challenge the rule of the Amir." During this period some Hazaras migrated to Quetta in Balochistan, while smaller number moved to Mashhad in northeastern Iran. The crackdown resulted in mass killing, enslavement and displacement of the Hazaras. Abdur Rahman died on October 1, 1901 inside his summer palace, being succeeded by his son Habibullah Khan. Today, his descendants can be found in many places outside Afghanistan, such as in America, France, Germany, and even in Scandinavian countries such as Denmark. His two eldest sons, Habibullah Khan and Nasrullah Khan, were born at Samarkand. His youngest son, Mahomed Omar Jan, was born in 1889 of an Afghan mother, connected by descent with the Barakzai family. Afghan society has mixed feelings about his rule. Some remember him as a ruler who initiated many programs for modernization and effectively prevented the country from being occupied by Russia and Britain during the Great Game. On the other hand, some sectors of Afghanistan remember him as a domestically violent and geopolitically weak ruler in that he was brought to power by the British and declared war on Afghan minorities instead of fighting the British who were deciding Afghanistan's foreign policy for him.
https://en.wikipedia.org/wiki?curid=2695
Abencerrages The Abencerrages or Abencerrajes (from the Arabic for "Saddler's Son") were a family or faction that is said to have held a prominent position in the Kingdom of Granada in the 15th century. The name appears to have been derived from Yussuf ben-Serragh, the head of the tribe in the time of Muhammed VII, Sultan of Granada (1370–1408), who did that sovereign good service in his struggles to retain the crown of which he was three times deprived. Little is known of the family with certainty. The Chambers Biographical Dictionary records that they arrived in Spain in the 8th century but the name is familiar from the romance by Ginés Pérez de Hita, "Guerras civiles de Granada", which celebrates the feuds of the Abencerrages and the rival family of the Benedin (Arabic banu Edin), and the cruel treatment to which the former were subjected. J. P. de Florian's "Gonsalve de Cordoue" and Chateaubriand's "Le dernier des Abencerrages" are adaptations of Pérez de Hita's story. The story is told that one of the Abencerrages, having fallen in love with a lady of the royal family, was caught in the act of climbing up to her window. The assassinations were ordered by Ibrahim Benedin who had a feud with the family. He was enraged and shut up the whole family in one of the halls of the Alhambra, and gave orders to kill them all. The apartment where this is said to have taken place is one of the most beautiful courts of the Alhambra, and is still called the Hall of the Abencerrages. Washington Irving in "Tales of the Alhambra" (1832) disagrees, saying the massacre was a fiction, but that a number of Abencerrages were killed in one of the battles at the time. Nonetheless, many poems and plays, the novella "The Abencerraje" and two operas ("Les Abencérages", by Luigi Cherubini, and "L'esule di Granata", by Giacomo Meyerbeer) mention the legend.
https://en.wikipedia.org/wiki?curid=2696
Aberdare Aberdare ( ; ) is a town in the Cynon Valley area of Rhondda Cynon Taf, Wales, at the confluence of the Rivers Dare (Dâr) and Cynon. Aberdare has a population of 39550(mid-2017 estimate). Aberdare is south-west of Merthyr Tydfil, north-west of Cardiff and east-north-east of Swansea. During the 19th century it became a thriving industrial settlement, which was also notable for the vitality of its cultural life and as an important publishing centre. The settlement of Aberdare dates from at least the Middle Ages, with the first known reference being in a monastic chapter of 1203 concerning grazing right on Hirwaun Common. It was originally a small village in an agricultural district, centred around the Church of St John the Baptist, said to date from 1189. By the middle of the 15th century, Aberdare contained a water mill in addition to a number of thatched cottages, of which no evidence remains. In the early 19th century the population of Aberdare grew rapidly, owing to the abundance of coal and iron ore,: the population of the whole parish, 1,486 in 1801, increased tenfold during the first half of the 19th century. Two major industries supported the growth of the community: first iron, then coal. A branch of the Glamorganshire Canal (1811) was used to transport these products; then the railway became the main means of transport to the South Wales coast. From the 1870s onwards, the economy of the town was dominated by the coal mining industry, with only a small tinplate works. There were also several brickworks and breweries. During the latter half of the 19th century, considerable improvements were made to the town, which became a pleasant place to live, despite the nearby collieries. A postgraduate theological college opened in connection with the Church of England in 1892, but in 1907 it moved to Llandaff. With the ecclesiastical parishes of St Fagan's (Trecynon) and Aberaman carved out of the ancient parish, Aberdare had 12 Anglican churches and one Catholic church, built in 1866 in Monk Street near the site of a cell attached to Penrhys monastery; and at one time there were over 50 Nonconformist chapels (including those in surrounding settlements such as Cwmaman and Llwydcoed). The services in the majority of the chapels were in Welsh. Most of these chapels have now closed, with many converted to other uses. The urban district includes what were once the separate villages of Aberaman, Abernant, Cwmaman, Cwmbach, Cwmdare, Llwydcoed, Penywaun and Trecynon. There are several cairns and the remains of a circular British encampment on the mountain between Aberdare and Merthyr. Hirwaun moor, 4 miles to the north west of Aberdare, was according to tradition the scene of a battle at which Rhys ap Tewdwr, prince of Dyfed, was defeated by the allied forces of the Norman Robert Fitzhamon and Iestyn ap Gwrgant, the last Welsh prince of Glamorgan. The parish population was 1,486 in 1801, but expanded fast, especially during the 1840s and 1850s: the population of the Aberdare District, centred on the town, was 9,322 in 1841; 18,774 in 1851; and 37,487 in 1861. This population growth, a result of the growth of the steam coal trade (see below) was increasingly concentrated in the previously agricultural areas of Blaengwawr and Cefnpennar to the south of the town. Many of the migrants came from the rural parts of west Wales which had been affected by an agricultural depression. Population levels continued to increase over the next forty years, albeit with a small decline in the 1870s. The first decade of the 20th century saw a further sharp increase, largely as a result of the steam coal trade, reaching 53,779 in 1911. The population has since declined owing to the loss of most of the heavy industry. The Aberdare population at the 2001 census was 31,705 (ranked 13th largest in Wales). By 2011 it was 29,748, though the figure includes the surrounding populations of Aberaman, Abercwmboi, Cwmbach and Llwydcoed. On 1 December 2016, following "The Rhondda Cynon Taf (Communities) Order 2016", the community of Aberdare was split into two new communities, Aberdare East and Aberdare West. These are coterminous with the electoral wards of the same names. Aberdare East includes Aberdare town centre and the village of Abernant. Aberdare West includes Cwmdare, Cwm Sian and Trecynon. Welsh was the prominent language until the mid 20th century and Aberdare was an important centre of Welsh language publishing. A large proportion of the early migrant population were Welsh speaking, and in 1851 only ten per cent of the population had been born outside of Wales. In his controversial evidence to the 1847 Education Reports (known in Wales as the "Brad y Llyfrau Gleision" or "Treachery of the Blue Books"), the Anglican vicar of Aberdare, John Griffith, stated that the English language was "generally understood" and referred to the arrival of people from anglicised areas such as Radnorshire and south Pembrokeshire. Griffith also made allegations about the Welsh speaking population and what he considered to be the degraded character of the women of Aberdare, alleging sexual promiscuity was an accepted social convention, that drunkenness and improvidence amongst the miners was common and attacking what he saw as exaggerated emotion in the religious practices of the Nonconformists. This evidence helped inform the findings of the report which would go on to stigmatise Welsh people as "ignorant", "lazy" and "immoral" and found the reason for this was the continued use of the Welsh language, which it described as "evil". The controversial reports allowed the local non-conformist minister Thomas Price of Calfaria to arrange public meetings, from which he would emerge as a leading critic of the vicar's evidence and, by implication, a defender of both the Welsh language and the morality of the local population, It is still contended that Griffiths was made vicar of Merthyr in the neighbouring valley to escape local anger, even though it was over ten years before he left Aberdare. The reports and subsequent defence would maintain the perceptions of Aberdare, the Cynon Valley and even the wider area as proudly nonconformist and defiantly Welsh speaking throughout its industrialised history. By 1901, the census recorded that 71.5% of the population of Aberdare Urban District spoke Welsh, but this fell to 65.2% in 1911. The 1911 data shows that Welsh was more widely spoken among the older generation compared to the young, and amongst women compared to men. A shift in language was expedited with the loss of men during the First World War and the resulting economic turmoil. English gradually began to replace Welsh as the community language, as shown by the decline of the Welsh language press in the town. This pattern continued after the Second World War despite the advent of Welsh medium education. Ysgol Gymraeg Aberdâr, the Welsh-medium primary school, was established in the 1950s with Idwal Rees as head teacher. According to the 2011 Census, 97.7% of Aberdare residents aged three years and over can speak Welsh, with 24.8% of 3–15 year olds stating that they can speak it. Ironworks were established at Llwydcoed and Abernant in 1799 and 1800 respectively, followed by others at Gadlys and Aberaman in 1827 and 1847. The iron industry began to expand in a significant way around 1818 when the Crawshay family of Merthyr purchased the Hirwaun ironworks and place them under independent management. In the following year, Rowland Fothergill took over the ironworks at Abernant and a few years later did the same at Llwydcoed. Both concerns later fell into the hands of his nephew Richard Fothergill. The Gadlys Ironworks was established in 1827 by Matthew Wayne, who had previously managed the Cyfarthfa ironworks at Merthyr. The Gadlys works, now considered an important archaeological site, originally comprised four blast furnaces, inner forges, rowing mills and puddling furnaces. The development of these works provided impetus to the growth of Aberdare as a nucleated town. The iron industry was gradually superseded by coal and all the five iron works had closed by 1875, as the local supply of iron ore was inadequate to meet the ever-increasing demand created by the invention of steel, and as a result the importing of ore proved more profitable. The iron industry had a relatively small impact upon the economy of Aberdare and in 1831 only 1.2% of the population was employed in manufacturing, as opposed to 19.8% in neighbouring Merthyr Tydfil. In the early years of Aberdare's development, most of the coal worked in the parish was coking coal, and was consumed locally, chiefly in the ironworks. Although the Gadlys works was small in comparison with the other ironworks it became significant as the Waynes also became involved in the production of sale coal. In 1836, this activity led to the exploitation of the "Four-foot Seam" of high-calorific value steam coal began, and pits were sunk in rapid succession. In 1840, Thomas Powell sank a pit at Cwmbach, and during the next few years he opened another four pits. In the next few years, other local entrepreneurs now became involved in the expansion of the coal trade, including David Williams at Ynysgynon and David Davis at Blaengwawr, as well as the latter's son David Davis, Maesyffynnon. They were joined by newcomers such as Crawshay Bailey at Aberaman and, in due course, George Elliot in the lower part of the valley. This coal was valuable for steam railways and steam ships, and an export trade began, via the Taff Vale Railway and the port of Cardiff. The population of the parish rose from 6,471 in 1841 to 14,999 in 1851 and 32,299 in 1861 and John Davies described it as "the most dynamic place in Wales". In 1851, the Admiralty decided to use Welsh steam coal in ships of the Royal Navy, and this decision boosted the reputation of Aberdare's product and launched a huge international export market. Coal mined in Aberdare parish rose from in 1844 to in 1850, and the coal trade, which after 1875 was the chief support of the town, soon reached huge dimensions. The growth of the coal trade inevitably led to a number of industrial disputes, some of which were local and others which affected the wider coalfield. Trade unionism began to appear in the Aberdare Valley at intervals from the 1830s onwards but the first significant manifestation occurred during the Aberdare Strike of 1857–8. The dispute was initiated by the depression in trade which followed the Crimean War and saw the local coal owners successfully impose a reduction in wages. The dispute did, however, witness an early manifestation of mass trade unionism amongst the miners of the valley and although unsuccessful the dispute saw the emergence of a stronger sense of solidarity amongst the miners. Steam coal was subsequently found in the Rhondda and further west, but many of the great companies of the Welsh coal industry's Gilded Age started operation in Aberdare and the lower Cynon Valley, including those of Samuel Thomas, David Davies and Sons, Nixon's Navigation and Powell Duffryn. During the early years of the twentieth century, the Aberdare valley became the focus of increased militancy among the mining workforce and an unofficial strike by 11,000 miners in the district from 20 October 1910 unyil 2 February 1911 attracted much attention at the time, although it was ultimately overshadowed by the Cambrian dispute in the neighbouring Rhondda valley which became synonymous with the so-called Tonypandy Riots. In common with the rest of the South Wales coalfield, Aberdare's coal industry commenced a long decline after World War I, and the last two deep mines still in operation in the 1960s were the small Aberaman and Fforchaman collieries, which closed in 1962 and 1965 respectively. On 11 May 1919, an extensive fire broke out on Cardiff Street, Aberdare. With the decline of both iron and coal, Aberdare has become reliant on commercial businesses as a major source of employment. Its industries include cable manufacture, smokeless fuels, and tourism. As a small village in the upland valleys of Glamorgan, Aberdare did not play any significant part in political life until its development as an industrial settlement. It was part of the lordship of Miskin, and the ancient office of High Constable continued in ceremonial form until relatively recent times. In 1832, Aberdare was removed from the county of Glamorgan and became part of the parliamentary borough of Merthyr Tydfil. For much of the nineteenth century, the representation was initially controlled by the ironmasters of Merthyr, notably the Guest family. From 1852 until 1868 the seat was held by Henry Austen Bruce whose main industrial interests lay in the Aberdare valley. Bruce was a Liberal but was viewed with suspicion by the more radical faction which became increasingly influential within Welsh Liberalism in the 1860s. The radicals supported such policies as the disestablishment of the Church of England and were closely allied to the Liberation Society. Nonconformist ministers played a prominent role in this new politics and, at Aberdare, they found an effective spokesman in the Rev Thomas Price minister of Calfaria, Aberdare. Following the granting of a second parliamentary seat to the borough of Merthyr Tydfil in 1867, the Liberals of Aberdare sought to ensure that a candidate from their part of the constituency was returned alongside the sitting member, Henry Austen Bruce. Their choice fell upon Richard Fothergill, owner of the ironworks at Abernant, who was enthusiastically supported by the Rev Thomas Price. Shortly before the election, however, Henry Richard intervened as a radical Liberal candidate, invited by the radicals of Merthyr. To many people's surprise, Price was lukewarm about his candidature and continued to support Fothergill. Ultimately, Henry Richard won a celebrated victory with Fothergill in second place and Bruce losing his seat. Richard thus became one of the-first radical MPs from Wales. At the 1874 General Election, both Richard and Fothergill were again returned, although the former was criticised for his apparent lack of sympathy towards the miners during the industrial disputes of the early 1870s. This led to the emergence of Thomas Halliday as the first labour or working-class candidate to contest a Welsh constituency. Although he polled well, Halliday fell short of being elected. For the remainder of the nineteenth century, the constituency was represented by industrialists, most notably David Alfred Thomas. In 1900, however, Thomas was joined by Keir Hardie, the ILP candidate, who became the first labour representative to be returned for a Welsh constituency independent of the Liberal Party. The Aberdare constituency came into being at the 1918 election. The first representative was Charles Butt Stanton who had been elected at a by-election following Hardie's death in 1915. However, in 1922, Stanton was defeated by a Labour candidate and the party has held the seat ever since. The only significant challenge came from Plaid Cymru at the 1970 and February 1974 General Elections but this performance has not since been repeated. Since 1984 the parliamentary seat, now known as Cynon Valley has been held by Ann Clwyd. Until the mid-nineteenth century the local government of Aberdare and its locality remained in the hands of traditional structures such as the parish vestry and the High Constable, who was chosen on an annual basis. However, the rapid industrial development of the parsing resulted in the situation where these traditional bodies could not cope with the realities of an urbanised, industrial community which had developed without any planning or facilities. During the early decades of the century the iron masters gradually imposed their influence over local affairs and this remained the case following the formation of the Merthyr Board of Guardians in 1836. During the 1850s and early 1860s, however, as coal displaced iron as the main industry in the valley, the ironmasters were displaced as the dominant group in local government and administration by an alliance between mostly indigenous coal owners, shopkeepers and tradesmen, professional men and dissenting ministers. A central figure in this development was the Rev Thomas Price. The growth of this alliance was rooted in the reaction to the 1847 Education Reports and the subsequent efforts to establish a British School at Aberdare. In the 1840s there were no adequate sanitary facilities or water supply and life expectancy was low. Outbreaks of cholera and typhus were commonplace. Against this background, Thomas Webster Rammell prepared a report for the General Board of Health on the sanitary condition of the parish, which concluded that a Local Board of Health be established. This happened in 1854. Its first chairman was Richard Fothergill and the members included David Davis, Blaengwawr, David Williams ("Alaw Goch"), Rees Hopkin Rhys and the Rev. Thomas Price. It was followed by the Aberdare School Board in 1871. By 1889, the Local Board of Health had initiated a number of developments which included the purchase of local reservoirs from the Aberdare Waterworks Company for £97,000, a sewerage scheme costing £35,000, as well as the opening of Aberdare Public Park and a local fever hospital. The lack of a Free Library, however, remained a concern. Later, the formation of the Glamorgan County Council (upon which Aberdare had five elected members) in 1889, followed by the Aberdare Urban District Council, which replaced the Local Board in 1889, transformed the local politics of the Aberdare valley. At the 1889 Glamorgan County Council Elections most of the elected representatives were coalowners and industrialists and the only exception in the earlier period was the miners' agent David Morgan (Dai o'r Nant), elected in 1892 as a labour representative. From the early 1900s, however, Labour candidates began to gain ground and dominated local government from the 1920s onwards. The same pattern was seen on the Aberdare UDC. In 1974, following local government re-organization, Aberdare became part of the county of Mid Glamorgan and the Cynon Valley Borough Council. Labour members held a majority of seats on both authorities until their abolition in 1996. Since the latest re-organization, Aberdare has been part of the Rhondda Cynon Taff unitary authority. Once again, Labour has been the majority party although Plaid Cymru controlled the authority from 1999 until 2003. Since 1995 Aberdare has elected county councillors to Rhondda Cynon Taf County Borough Council. The town lies mainly in the Aberdare East ward, represented by two county councillors. Nearby Cwmdare, Llwydcoed and Trecynon are represented by the Aberdare West/Llwydcoed ward. Both wards have been represented by the Labour Party since 2012. Aberdare, during its boom years, was considered a centre of Welsh culture: it hosted the first National Eisteddfod in 1861, with which David Williams (Alaw Goch) was closely associated. A number of local eisteddfodau had long been held in the locality, associated with figures such as William Williams (Carw Coch) The Eisteddfod was again held in Aberdare in 1885, and also in 1956 at Aberdare Park where the Gorsedd standing stones still exist. At the last National Eisteddfod held in Aberdare in 1956 Mathonwy Hughes won the chair. From the mid nineteenth century, Aberdare was an important publishing centre where a large number of books and journals were produced, the majority of which were in the Welsh language. A newspaper entitled Y Gwladgarwr (the Patriot) was published at Aberdare from 1856 until 1882 and was circulated widely throughout the South Wales valleys. From 1875 a more successful newspaper, Tarian y Gweithiwr (the Workman's Shield) was published at Aberdare by John Mills. Y Darian, as it was known, strongly supported the trade union movements among the miners and ironworkers of the valleys. The miners' leader, William Abraham, derived support from the newspaper, which was also aligned with radical nonconformist liberalism. The rise of the political labour movement and the subsequent decline of the Welsh language in the valleys, ultimately led to its decline and closure in 1934. The Coliseum Theatre is Aberdare's main arts venue, containing a 600-seat auditorium and cinema. It is situated in nearby Trecynon and was built in 1938 using miners' subscriptions. The Second World War poet Alun Lewis, was born near Aberdare in the village Cwmaman and there is a plaque commemorating him, including a quotation from his poem "The Mountain over Aberdare". The founding members of the rock band Stereophonics originated from the nearby village of Cwmaman. It is also the hometown of guitarist Mark Parry of Vancouver rock band The Manvils. Famed anarchist-punk band Crass played their last live show for striking miners in Aberdare during the UK miners' strike. Griffith Rhys Jones − or Caradog as he was commonly known − was the Conductor of the famous 'Côr Mawr' of some 460 voices (the South Wales Choral Union), which twice won first prize at Crystal Palace choral competitions in London in the 1870s. He is depicted in the town's most prominent statue by sculptor Goscombe John, unveiled on Victoria Square in 1920. Aberdare was culturally twinned with the German town of Ravensburg. The original parish church of St John the Baptist was originally built in 1189. Some of its original architecture is still intact. With the development of Aberdare as an industrial centre in the nineteenth century it became increasingly apparent that the ancient church was far too small to service the perceived spiritual needs of an urban community, particularly in view of the rapid growth of nonconformity from the 1830s onwards. Eventually, John Griffith, the rector of Aberdare undertook to raise funds to build a new church, leading to the rapid construction of St Elvan's Church in the town centre between 1851 and 1852. This Church in Wales church still stands the heart of the parish of Aberdare and has had extensive work since its erection. The church has a modern electrical, two-manual and pedal board pipe organ, that is still used in services. John Griffith, vicar of Aberdare, who built St Elvan's, transformed the role of the Anglican church in the valley by building a number of other churches, including St Fagan's, Trecynon. Other churches in the parish are St Luke's (Cwmdare), St James's (Llwydcoed) and St Matthew's Church (1891) (Abernant). In the parish of Aberaman and Cwmaman is St Margaret's Church, with an old, but beautiful, pipe organ with two manuals and a pedal board. Also in this parish is St Joseph's Church, Cwmaman. St Joseph's has recently undergone much recreational work, almost converting the church into a community centre. However, regular church services still take place. Here, there is a two-manual and pedal board electric organ, with speakers at the front and sides of the church. In 1910 there were 34 Anglican churches in the Urban District of Aberdare. A survey of the attendance at places of worship on a particular Sunday in that year recorded that 17.8% of worshippers attended church services, with the remainder attending nonconformist chapels. The Aberdare Valley was a stronghold of Nonconformity from the mid-nineteenth century until the inter-war years. In the aftermath of the 1847 Education Reports nonconformists became increasingly active in the political and educational life of Wales and in few places was this as prevalent as at Aberdare. The leading figure was Thomas Price, minister of Calfaria, Aberdare. Aberdare was a major centre of the 1904–05 Religious Revival, which had begun at Loughor near Swansea. The revival aroused alarm among ministers for the revolutionary, even anarchistic, impact it had upon chapel congregations and denominational organisation. In particular, it was seen as drawing attention away from pulpit preaching and the role of the minister. The local newspaper, the "Aberdare Leader", regarded the revival with suspicion from the outset, objecting to the 'abnormal heat' which it engendered. Trecynon was particularly affected by the revival, and the meetings held there were sais to have aroused more emotion and excitement than the more restrained meetings in Aberdare itself. The impact of the revival was significant in the short term, but in the longer term was fairly transient. Once the immediate impact of the revival had faded, it was clear from the early twentieth century that there was a gradual decline in the influence of the chapels. This can be explained by several factors, including the rise of socialism and the process of linguistic change which saw the younger generation increasingly turn to the English language. There were also theological controversies such as that over the New Theology propounded by R.J. Campbell. Of the many chapels, few are still used for their original purpose and a number of closed since the turn of the millennium. Many have been converted for housing or other purposes (including one at Robertstown which has become a mosque), and others demolished. Among the notable chapels were Calfaria, Aberdare and Seion, Cwmaman (Baptist); Saron, Aberaman and Siloa, Aberdare (Independent); and Bethania, Aberdare (Calvinistic Methodist). The earliest Welsh Independent, or Congregationalist chapel in the Aberdare area was Ebenezer, trecynon, although meetings had been held from the latter years of the eighteenth century in dwelling houses in the locality, for example at Hirwaun. During the nineteenth century, the Independents showed the biggest increases in terms of places of worship: from two in 1837 to twenty-five (four of them being English causes), in 1897. By 1910 there were 35 Independent chapels, with a total membership of 8,612. Siloa Chapel was the largest of the Independent chapels in Aberdare and is one of the few that remain open today, having been 're-established' as a Welsh language chapel. The Independent ministers of nineteenth-century Aberdare included some powerful personalities but none had the kind of wider social authority which Thomas Price enjoyed amongst the Baptists. Of the other Independent chapels in the valley Saron, in Davis Street, Aberaman, was used for regular services by a small group of members until 2011. For many years, these were held in a small side-room, and not the chapel itself. The chapel has a large vestry comprising rows of two-way-facing wooden benches and a stage, with a side entrance onto Beddoe Street and back entrance to Lewis Street. Although the building is not in good repair, the interior, including pulpit and balcony seating area (back & sides), was in good order but the chapel eventually closed due to the very small number of members remaining. In February 1999, Saron was made a Grade II Listed Building. The Baptists were the most influential of the nonconformist denominations in Aberdare and their development was led by the Rev. Thomas Price who came to Aberdare in the early 1840s as minister of Calfaria Chapel. In 1837 the Baptists had three chapels, but in 1897 there were twenty, seventeen of them being Welsh. By 1910 the number of chapels had increased to 30, with a total membership of 7,422. Most of these Baptist chapels were established under the influence of Thomas Price who encouraged members to establish branch chapels to attract migrants who flocked to the town and locality from rural Wales. The chapels came together for regular gatherings, including baptismal services which were held in the River Cynon As a result, Price exerted an influence in the religious life of the locality which was far greater than that of any other minister. By 1910 there were 24 Calvinistic Methodist chapels in the Aberdare Urban District with a total membership of 4,879. The most prominent of these was Bethania, Aberdare, once the largest chapel in Aberdare. Derelict for many years, it was demolished in 2015. The Methodists were numerically powerful and while some of their ministers such as William James of Bethania served on the Aberdare School Board and other public bodies, their constitution militated against the sort of active political action which came more naturally to the Baptists and Independents. The other denominations were weaker, including the Wesleyan Methodists who had 14 places of worship by 1910. There was also a significant Unitarian tradition in the valley and three places of worship by 1910. Highland Place Unitarian Church celebrated its 150th anniversary in 2010, with a number of lectures on its history and the history of Unitarianism in Wales taking place there. The church has a two-manual pipe organ with pedal board that is used to accompany all services. The current organist is Grace Jones, the sister of the former organist Jacob Jones. The connected schoolroom is used for post-service meetings and socialising. Seymour Street was once home to a synagogue which opened its doors in the late 1800s and which closed in 1957. The site now has a blue plaque. The state of education in the parish was a cause for concern during the early industrial period as is illustrated by the reaction to the 1847 Education Reports. Initially, there was an outcry, led by the Rev Thomas Price against the comments made by the vicar of Aberdare in his submission to the commissioners. However, on closer reflection, the reports related the deficiencies of educational provision, not only in Aberdare itself but also in the communities of the valleys generally. In so doing they not only criticised the ironmasters for their failure to provide schools for workers' children but also the nonconformists for not establishing British Schools. At the ten schools in Aberdare there was accommodation for only 1,317 children, a small proportion of the population. Largely as a result of these criticisms, the main nonconformist denominations worked together to establish a British School, known locally as Ysgol y Comin, which was opened in 1848, accommodating 200 pupils. Funds were raised which largely cleared the debts and the opening of the school was marked by a public meeting addressed by Price and David Williams ("Alaw Goch"). Much energy was expended during this period on conflicts between Anglicans and nonconformists over education. The establishment of the Aberdare School Board in 1871 brought about an extension of educational provision but also intensified religious rivalries. School Board elections were invariably fought on religious grounds. Despite these tensions the Board took over a number of existing schools and established new ones. By 1889, fourteen schools were operated by the Board but truancy and lack of attendance remained a problem, as in many industrial districts. In common with other public bodies at the time (see 'Local Government' above), membership of the School Board was dominated by coal owners and colliery officials, nonconformist ministers, professional men and tradesmen. Only occasionally was an Anglican clergyman elected and, with the exception of David Morgan ("Dai o'r Nant"), no working class candidates were elected for more than one term. The town is served by Aberdare railway station and Aberdare bus station, opposite each other in the town centre. The town has also been subject to an extensive redevelopment scheme during 2012–13. Aberdare Athletic F.C. were members of the Football League between 1921 and 1927 before being replaced by Torquay United after finishing bottom. The senior club folded a year later. They played their football league games at the Aberdare Athletic Ground and were known as the Darians. The reserve team carried on as Aberaman and Aberdare Athletic for one more season and were known as Aberaman Athletic F.C. Now renamed as Aberdare Town They play in the Welsh League Division Two at Aberaman Park Aberdare Rugby Football Club are a rugby union team formed in 1890 which still play in Aberdare today at the Ynys Stadium. The Aberdare Athletic Ground was the venue of the first rugby league international, played between Wales and the New Zealand All Golds on New Year's Day 1908, which was won by the Welsh 9-8. Notable current and former residents and natives of Aberdare include:
https://en.wikipedia.org/wiki?curid=2701
Aberration (astronomy) In astronomy, aberration (also referred to as astronomical aberration, stellar aberration, or velocity aberration) is a phenomenon which produces an apparent motion of celestial objects about their true positions, dependent on the velocity of the observer. It causes objects to appear to be displaced towards the direction of motion of the observer compared to when the observer is stationary. The change in angle is of the order of "v/c" where "c" is the speed of light and "v" the velocity of the observer. In the case of "stellar" or "annual" aberration, the apparent position of a star to an observer on Earth varies periodically over the course of a year as the Earth's velocity changes as it revolves around the Sun, by a maximum angle of approximately 20 arcseconds in right ascension or declination. The term "aberration" has historically been used to refer to a number of related phenomena concerning the propagation of light in moving bodies. Aberration is distinct from parallax, which is a change in the apparent position of a relatively nearby object, as measured by a moving observer, relative to more distant objects that define a reference frame. The amount of parallax depends on the distance of the object from the observer, whereas aberration does not. Aberration is also related to light-time correction and relativistic beaming, although it is often considered separately from these effects. Aberration is historically significant because of its role in the development of the theories of light, electromagnetism and, ultimately, the theory of special relativity. It was first observed in the late 1600s by astronomers searching for stellar parallax in order to confirm the heliocentric model of the Solar System. However, it was not understood at the time to be a different phenomenon. In 1727, James Bradley provided a classical explanation for it in terms of the finite speed of light relative to the motion of the Earth in its orbit around the Sun, which he used to make one of the earliest measurements of the speed of light. However, Bradley's theory was incompatible with 19th century theories of light, and aberration became a major motivation for the aether drag theories of Augustin Fresnel (in 1818) and G. G. Stokes (in 1845), and for Hendrik Lorentz's aether theory of electromagnetism in 1892. The aberration of light, together with Lorentz's elaboration of Maxwell's electrodynamics, the moving magnet and conductor problem, the negative aether drift experiments, as well as the Fizeau experiment, led Albert Einstein to develop the theory of special relativity in 1905, which presents a general form of the equation for aberration in terms of such theory. Aberration may be explained as the difference in angle of a beam of light in different inertial frames of reference. A common analogy is to consider the apparent direction of falling rain. If rain is falling vertically in the frame of reference of a person standing still, then to a person moving forwards the rain will appear to arrive at an angle, requiring the moving observer to tilt their umbrella forwards. The faster the observer moves, the more tilt is needed. The net effect is that light rays striking the moving observer from the sides in a stationary frame will come angled from ahead in the moving observer's frame. This effect is sometimes called the "searchlight" or "headlight" effect. In the case of annual aberration of starlight, the direction of incoming starlight as seen in the Earth's moving frame is tilted relative to the angle observed in the Sun's frame. Since the direction of motion of the Earth changes during its orbit, the direction of this tilting changes during the course of the year, and causes the apparent position of the star to differ from its true position as measured in the inertial frame of the Sun. While classical reasoning gives intuition for aberration, it leads to a number of physical paradoxes observable even at the classical level (see history). The theory of special relativity is required to correctly account for aberration. The relativistic explanation is very similar to the classical one however, and in both theories aberration may be understood as a case of addition of velocities. In the Sun's frame, consider a beam of light with velocity equal to the speed of light c, with x and y velocity components formula_1 and formula_2, at an angle formula_3. If the Earth is moving at velocity formula_4 in the x direction relative to the Sun, then by velocity addition the x component of the beam's velocity in the Earth's frame of reference is formula_5, and the y velocity is unchanged, formula_6. Thus the angle of the light in the Earth's frame in terms of the angle in the Sun's frame is In the case of formula_8, this result reduces to formula_9, which in the limit formula_10 may be approximated by formula_11. The reasoning in the relativistic case is the same except that the relativistic velocity addition formulas must be used, which can be derived from Lorentz transformations between different frames of reference. These formulas are where formula_14, giving the components of the light beam in the Earth's frame in terms of the components in the Sun's frame. The angle of the beam in the Earth's frame is thus In the case of formula_8, this result reduces to formula_17, and in the limit formula_10 this may be approximated by formula_11. This relativistic derivation keeps the speed of light formula_20 constant in all frames of reference, unlike the classical derivation above. Aberration is related to two other phenomena, light-time correction, which is due to the motion of an observed object during the time taken by its light to reach an observer, and relativistic beaming, which is an angling of the light emitted by a moving light source. It can be considered equivalent to them but in a different inertial frame of reference. In aberration, the observer is considered to be moving relative to a (for the sake of simplicity) stationary light source, while in light-time correction and relativistic beaming the light source is considered to be moving relative to a stationary observer. Consider the case of an observer and a light source moving relative to each other at constant velocity, with a light beam moving from the source to the observer. At the moment of emission, the beam in the observer's rest frame is tilted compared to the one in the source's rest frame, as understood through relativistic beaming. During the time it takes the light beam to reach the observer the light source moves in the observer's frame, and the 'true position' of the light source is displaced relative to the apparent position the observer sees, as explained by light-time correction. Finally, the beam in the observer's frame at the moment of observation is tilted compared to the beam in source's frame, which can be understood as an aberrational effect. Thus, a person in the light source's frame would describe the apparent tilting of the beam in terms of aberration, while a person in the observer's frame would describe it as a light-time effect. The relationship between these phenomena is only valid if the observer and source's frames are inertial frames. In practice, because the Earth is not an inertial rest frame but experiences centripetal acceleration towards the Sun, many aberrational effects such as annual aberration on Earth cannot be considered light-time corrections. However, if the time between emission and detection of the light is short compared to the orbital period of the Earth, the Earth may be approximated as an inertial frame and aberrational effects are equivalent to light-time corrections. There are a number of types of aberration, caused by the differing components of the Earth's and observed object's motion: Annual aberration is caused by the motion of an observer on Earth as the planet revolves around the Sun. Due to orbital eccentricity, the orbital velocity formula_4 of Earth (in the Sun's rest frame) varies periodically during the year as the planet traverses its elliptic orbit and consequently the aberration also varies periodically, typically causing stars to appear to move in small ellipses. Approximating Earth's orbit as circular, the maximum displacement of a star due to annual aberration is known as the "constant of aberration", conventionally represented by formula_22. It may be calculated using the relation formula_23 substituting the Earth's average speed in the Sun's frame for formula_4 and the speed of light formula_25. Its accepted value is 20.49552" or arcseconds or 0.00099365 rad or radian (at J2000). Assuming a circular orbit, annual aberration causes stars exactly on the ecliptic (the plane of Earth's orbit) to appear to move back and forth along a straight line, varying by formula_22 on either side of their position in the Sun's frame. A star that is precisely at one of the ecliptic poles (at 90° from the ecliptic plane) will appear to move in a circle of radius formula_22 about its true position, and stars at intermediate ecliptic latitudes will appear to move along a small ellipse. For illustration, consider a star at the northern ecliptic pole viewed by an observer at a point on the Arctic Circle. Such an observer will see the star transit at the zenith, once every day (strictly speaking sidereal day). At the time of the March equinox, Earth's orbit carries the observer in a southwards direction, and the star's apparent declination is therefore displaced to the south by an angle of formula_22. On the September equinox, the star's position is displaced to the north by an equal and opposite amount. On either solstice, the displacement in declination is 0. Conversely, the amount of displacement in right ascension is 0 on either equinox and at maximum on either solstice. In actuality, Earth's orbit is slightly elliptic rather than circular, and its speed varies somewhat over the course of its orbit, which means the description above is only approximate. Aberration is more accurately calculated using Earth's instantaneous velocity relative to the barycenter of the Solar System. Note that the displacement due to aberration is orthogonal to any displacement due to parallax. If parallax were detectable, the maximum displacement to the south would occur in December, and the maximum displacement to the north in June. It is this apparently anomalous motion that so mystified early astronomers. A special case of annual aberration is the nearly constant deflection of the Sun from its position in the Sun's rest frame by formula_22 towards the "west" (as viewed from Earth), opposite to the apparent motion of the Sun along the ecliptic (which is from west to east, as seen from Earth). The deflection thus makes the Sun appear to be behind (or retarded) from its rest-frame position on the ecliptic by a position or angle formula_22. This deflection may equivalently be described as a light-time effect due to motion of the Earth during the 8.3 minutes that it takes light to travel from the Sun to Earth. The relation with "κ" is : [0.000099365 rad / 2 π rad] x [ 365.25 d x 24 h/d x 60 min/h] = 8.3 min = 8 min 19 s. This is possible since the transit time of sunlight is short relative to the orbital period of the Earth, so the Earth's frame may be approximated as inertial. In the Earth's frame, the Sun moves by a distance formula_31 in the time it takes light to reach Earth, formula_32 for the orbit of radius formula_33. This gives an angular correction formula_34 which can be solved to give formula_35, the same as the aberrational correction (here "κ" is in radian and not in arcsecond). Planetary aberration is the combination of the aberration of light (due to Earth's velocity) and light-time correction (due to the object's motion and distance), as calculated in the rest frame of the Solar System. Both are determined at the instant when the moving object's light reaches the moving observer on Earth. It is so called because it is usually applied to planets and other objects in the Solar System whose motion and distance are accurately known. Diurnal aberration is caused by the velocity of the observer on the surface of the rotating Earth. It is therefore dependent not only on the time of the observation, but also the latitude and longitude of the observer. Its effect is much smaller than that of annual aberration, and is only 0.32 arcseconds in the case of an observer at the Equator, where the rotational velocity is greatest. The Sun and Solar System are revolving around the center of the Galaxy. Aberration due to this motion is known as secular aberration and affects the apparent positions of distant stars and extragalactic objects. However, since the galactic year is about 230 million years, the aberration varies very slowly and this change is extremely difficult to observe. Therefore, secular aberration is usually ignored when considering the positions of stars. In other words, star maps show the observed apparent positions of the stars, not their calculated true positions after accounting for secular aberration. For stars significantly less than 230 million light years away, the Solar System may be approximated as an inertial frame and so the effect of secular aberration is equivalent to a light-time correction. This includes stars in the Milky Way, since the Milky Way is about 100,000 light years in diameter. For these stars the true position of the star is then easily computed from its proper motion and its distance. Secular aberration is typically a small number of arcminutes, for example the stationary star Groombridge 1830 is displaced by approximately 3 arcminutes, due to secular aberration. This is roughly 8 times the effect of annual aberration, as one would expect since the velocity of the Solar System relative to the center of the Galaxy is about 8 times the velocity of the Earth relative to the Sun. The discovery of the aberration of light was totally unexpected, and it was only by considerable perseverance and perspicacity that Bradley was able to explain it in 1727. It originated from attempts to discover whether stars possessed appreciable parallaxes. The Copernican heliocentric theory of the Solar System had received confirmation by the observations of Galileo and Tycho Brahe and the mathematical investigations of Kepler and Newton. As early as 1573, Thomas Digges had suggested that parallactic shifting of the stars should occur according to the heliocentric model, and consequently if stellar parallax could be observed it would help confirm this theory. Many observers claimed to have determined such parallaxes, but Tycho Brahe and Giovanni Battista Riccioli concluded that they existed only in the minds of the observers, and were due to instrumental and personal errors. However, in 1680 Jean Picard, in his "Voyage d’Uranibourg," stated, as a result of ten years' observations, that Polaris, the Pole Star, exhibited variations in its position amounting to 40″ annually. Some astronomers endeavoured to explain this by parallax, but these attempts failed because the motion differed from that which parallax would produce. John Flamsteed, from measurements made in 1689 and succeeding years with his mural quadrant, similarly concluded that the declination of Polaris was 40″ less in July than in September. Robert Hooke, in 1674, published his observations of γ Draconis, a star of magnitude 2m which passes practically overhead at the latitude of London (hence its observations are largely free from the complex corrections due to atmospheric refraction), and concluded that this star was 23″ more northerly in July than in October. Consequently, when Bradley and Samuel Molyneux entered this sphere of research in 1725, there was still considerable uncertainty as to whether stellar parallaxes had been observed or not, and it was with the intention of definitely answering this question that they erected a large telescope at Molyneux's house at Kew. They decided to reinvestigate the motion of γ Draconis with a telescope constructed by George Graham (1675–1751), a celebrated instrument-maker. This was fixed to a vertical chimney stack in such manner as to permit a small oscillation of the eyepiece, the amount of which (i.e. the deviation from the vertical) was regulated and measured by the introduction of a screw and a plumb line. The instrument was set up in November 1725, and observations on γ Draconis were made starting in December. The star was observed to move 40″ southwards between September and March, and then reversed its course from March to September. At the same time, 35 Camelopardalis, a star with a right ascension nearly exactly opposite to that of γ Draconis, was 19" more northerly at the beginning of March than in September. These results were completely unexpected and inexplicable by existing theories. Bradley and Molyneux discussed several hypotheses in the hope of finding the solution. Since the apparent motion was evidently caused neither by parallax nor observational errors, Bradley first hypothesized that it could be due to oscillations in the orientation of the Earth's axis relative to the celestial sphere – a phenomenon known as nutation. 35 Camelopardalis was seen to possess an apparent motion which could be consistent with nutation, but since its declination varied only one half as much as that of γ Draconis, it was obvious that nutation did not supply the answer (however, Bradley later went on to discover that the Earth does indeed nutate). He also investigated the possibility that the motion was due to an irregular distribution of the Earth's atmosphere, thus involving abnormal variations in the refractive index, but again obtained negative results. On August 19, 1727, Bradley embarked upon a further series of observations using a telescope of his own erected at the Rectory, Wanstead. This instrument had the advantage of a larger field of view and he was able to obtain precise positions of a large number of stars over the course of about twenty years. During his first two years at Wanstead, he established the existence of the phenomenon of aberration beyond all doubt, and this also enabled him to formulate a set of rules that would allow the calculation of the effect on any given star at a specified date. Bradley eventually developed his explanation of aberration in about September 1728 and this theory was presented to the Royal Society in mid January the following year. One well-known story was that he saw the change of direction of a wind vane on a boat on the Thames, caused not by an alteration of the wind itself, but by a change of course of the boat relative to the wind direction. However, there is no record of this incident in Bradley's own account of the discovery, and it may therefore be apocryphal. The following table shows the magnitude of deviation from true declination for γ Draconis and the direction, on the planes of the solstitial colure and ecliptic prime meridian, of the tangent of the velocity of the Earth in its orbit for each of the four months where the extremes are found, as well as expected deviation from true ecliptic longitude if Bradley had measured its deviation from right ascension: Bradley proposed that the aberration of light not only affected declination, but right ascension as well, so that a star in the pole of the ecliptic would describe a little ellipse with a diameter of about 40", but for simplicity, he assumed it to be a circle. Since he only observed the deviation in declination, and not in right ascension, his calculations for the maximum deviation of a star in the pole of the ecliptic are for its declination only, which will coincide with the diameter of the little circle described by such star. For eight different stars, his calculations are as follows: Based on these calculations, Bradley was able to estimate the constant of aberration at 20.2", which is equal to 0.00009793 radians, and with this was able to estimate the speed of light at per second. By projecting the little circle for a star in the pole of the ecliptic, he could simplify the calculation of the relationship between the speed of light and the speed of the Earth's annual motion in its orbit as follows: Thus, the speed of light to the speed of the Earth's annual motion in its orbit is 10,210 to one, from whence it would follow, that light moves, or is propagated as far as from the Sun to the Earth in 8 minutes 12 seconds. The discovery and elucidation of aberration is now regarded as a classic case of the application of scientific method, in which observations are made to test a theory, but unexpected results are sometimes obtained that in turn lead to new discoveries. It is also worth noting that part of the original motivation of the search for stellar parallax was to test the Copernican theory that the Earth revolves around the Sun, but of course the existence of aberration also establishes the truth of that theory. The phenomenon of aberration became a driving force for many physical theories during the 200 years between its observation and the conclusive explanation by Albert Einstein. The first classical explanation was provided in 1729, by James Bradley as described above, who attributed it to the finite speed of light and the motion of Earth in its orbit around the Sun. However, this explanation proved inaccurate once the wave nature of light was better understood, and correcting it became a major goal of the 19th century theories of luminiferous aether. Augustin-Jean Fresnel proposed a correction due to the motion of a medium (the aether) through which light propagated, known as "partial aether drag". He proposed that objects partially drag the aether along with them as they move, and this became the accepted explanation for aberration for some time. George Stokes proposed a similar theory, explaining that aberration occurs due to the flow of aether induced by the motion of the Earth. Accumulated evidence against these explanations, combined with new understanding of the electromagnetic nature of light, led Hendrik Lorentz to develop an electron theory which featured an immobile aether, and he explained that objects contract in length as they move through the aether. Motivated by these previous theories, Albert Einstein then developed the theory of special relativity in 1905, which provides the modern account of aberration. Bradley conceived of an explanation in terms of a corpuscular theory of light in which light is made of particles. His classical explanation appeals to the motion of the earth relative to a beam of light-particles moving at a finite velocity, and is developed in the Sun's frame of reference, unlike the classical derivation given above. Consider the case where a distant star is motionless relative to the Sun, and the star is extremely far away, so that parallax may be ignored. In the rest frame of the Sun, this means light from the star travels in parallel paths to the Earth observer, and arrives at the same angle regardless of where the Earth is in its orbit. Suppose the star is observed on Earth with a telescope, idealized as a narrow tube. The light enters the tube from the star at angle formula_37 and travels at speed formula_25 taking a time formula_39 to reach the bottom of the tube, where it is detected. Suppose observations are made from Earth, which is moving with a speed formula_4. During the transit of the light, the tube moves a distance formula_41. Consequently, for the particles of light to reach the bottom of the tube, the tube must be inclined at an angle formula_42 different from formula_37, resulting in an "apparent" position of the star at angle formula_42. As the Earth proceeds in its orbit it changes direction, so formula_42 changes with the time of year the observation is made. The apparent angle and true angle are related using trigonometry as: In the case of formula_8, this gives formula_9. While this is different from the more accurate relativistic result described above, in the limit of small angle and low velocity they are approximately the same, within the error of the measurements of Bradley's day. These results allowed Bradley to make one of the earliest measurements of the speed of light. In the early nineteenth century the wave theory of light was being rediscovered, and in 1804 Thomas Young adapted Bradley's explanation for corpuscular light to wavelike light traveling through a medium known as the luminiferous aether. His reasoning was the same as Bradley's, but it required that this medium be immobile in the Sun's reference frame and must pass through the earth unaffected, otherwise the medium (and therefore the light) would move along with the earth and no aberration would be observed. He wrote: However, it soon became clear Young's theory could not account for aberration when materials with a non-vacuum index of refraction were present. An important example is of a telescope filled with water. The velocity of the light in such a telescope will be slower than in vacuum, and is given by formula_49 rather than formula_25 where formula_51 is the index of refraction of the water. Thus, by Bradley and Young's reasoning the aberration angle is given by which predicts a medium-dependent angle of aberration. When refraction at the telescope's objective is taken into account this result deviates even more from the vacuum result. In 1810 François Arago performed a similar experiment and found that the aberration was unaffected by the medium in the telescope, providing solid evidence against Young's theory. This experiment was subsequently verified by many others in the following decades, most accurately by Airy in 1871, with the same result. In 1818, Augustin Fresnel developed a modified explanation to account for the water telescope and for other aberration phenomena. He explained that the aether is generally at rest in the Sun's frame of reference, but objects partially drag the aether along with them as they move. That is, the aether in an object of index of refraction formula_51 moving at velocity formula_4 is partially dragged with a velocity formula_55 bringing the light along with it. This factor is known as "Fresnel's dragging coefficient". This dragging effect, along with refraction at the telescope's objective, compensates for the slower speed of light in the water telescope in Bradley's explanation. With this modification Fresnel obtained Bradley's vacuum result even for non-vacuum telescopes, and was also able to predict many other phenomena related to the propagation of light in moving bodies. Fresnel's dragging coefficient became the dominant explanation of aberration for the next decades. However, the fact that light is polarized (discovered by Fresnel himself) led scientists such as Cauchy and Green to believe that the aether was a totally immobile elastic solid as opposed to Fresnel's fluid aether. There was thus renewed need for an explanation of aberration consistent both with Fresnel's predictions (and Arago's observations) as well as polarization. In 1845, Stokes proposed a 'putty-like' aether which acts as a liquid on large scales but as a solid on small scales, thus supporting both the transverse vibrations required for polarized light and the aether flow required to explain aberration. Making only the assumptions that the fluid is irrotational and that the boundary conditions of the flow are such that the aether has zero velocity far from the Earth, but moves at the Earth's velocity at its surface and within it, he was able to completely account for aberration. The velocity of the aether outside of the Earth would decrease as a function of distance from the Earth so light rays from stars would be progressively dragged as they approached the surface of the Earth. The Earth's motion would be unaffected by the aether due to D'Alembert's paradox. Both Fresnel and Stokes' theories were popular. However, the question of aberration was put aside during much of the second half of the 19th century as focus of inquiry turned to the electromagnetic properties of aether. In the 1880s once electromagnetism was better understood, interest turned again to the problem of aberration. By this time flaws were known to both Fresnel's and Stokes' theories. Fresnel's theory required that the relative velocity of aether and matter to be different for light of different colors, and it was shown that the boundary conditions Stokes had assumed in his theory were inconsistent with his assumption of irrotational flow. At the same time, the modern theories of electromagnetic aether could not account for aberration at all. Many scientists such as Maxwell, Heaviside and Hertz unsuccessfully attempted to solve these problems by incorporating either Fresnel or Stokes' theories into Maxwell's new electromagnetic laws. Hendrik Lorentz spent considerable effort along these lines. After working on this problem for a decade, the issues with Stokes' theory caused him to abandon it and to follow Fresnel's suggestion of a (mostly) stationary aether (1892, 1895). However, in Lorentz's model the aether was "completely" immobile, like the electromagnetic aethers of Cauchy, Green and Maxwell and unlike Fresnel's aether. He obtained Fresnel's dragging coefficient from modifications of Maxwell's electromagnetic theory, including a modification of the time coordinates in moving frames ("local time"). In order to explain the Michelson–Morley experiment (1887), which apparently contradicted both Fresnel's and Lorentz's immobile aether theories, and apparently confirmed Stokes' complete aether drag, Lorentz theorized (1892) that objects undergo "length contraction" by a factor of formula_56 in the direction of their motion through the aether. In this way, aberration (and all related optical phenomena) can be accounted for in the context of an immobile aether. Lorentz' theory became the basis for much research in the next decade, and beyond. Its predictions for aberration are identical to those of the relativistic theory. Lorentz' theory matched experiment well, but it was complicated and made many unsubstantiated physical assumptions about the microscopic nature of electromagnetic media. In his 1905 theory of special relativity, Albert Einstein reinterpreted the results of Lorentz' theory in a much simpler and more natural conceptual framework which disposed of the idea of an aether. His derivation is given above, and is now the accepted explanation. Robert S. Shankland reported some conversations with Einstein, in which Einstein emphasized the importance of aberration: Other important motivations for Einstein's development of relativity were the moving magnet and conductor problem and (indirectly) the negative aether drift experiments, already mentioned by him in the introduction of his first relativity paper. Einstein wrote in a note in 1952: While Einstein's result is the same as Bradley's original equation except for an extra factor of formula_57, Bradley's result does not merely give the classical limit of the relativistic case, in the sense that it gives incorrect predictions even at low relative velocities. Bradley's explanation cannot account for situations such as the water telescope, nor for many other optical effects (such as interference) that might occur within the telescope. This is because in the Earth's frame it predicts that the direction of propagation of the light beam in the telescope is not normal to the wavefronts of the beam, in contradiction with Maxwell's theory of electromagnetism. It also does not preserve the speed of light c between frames. However, Bradley did correctly infer that the effect was due to relative velocities.
https://en.wikipedia.org/wiki?curid=2703
Optical aberration In optics, aberration is a property of optical systems such as lenses that causes light to be spread out over some region of space rather than focused to a point. Aberrations cause the image formed by a lens to be blurred or distorted, with the nature of the distortion depending on the type of aberration. Aberration can be defined as a departure of the performance of an optical system from the predictions of paraxial optics. In an imaging system, it occurs when light from one point of an object does not converge into (or does not diverge from) a single point after transmission through the system. Aberrations occur because the simple paraxial theory is not a completely accurate model of the effect of an optical system on light, rather than due to flaws in the optical elements. An image-forming optical system with aberration will produce an image which is not sharp. Makers of optical instruments need to correct optical systems to compensate for aberration. Aberration can be analyzed with the techniques of geometrical optics. The articles on reflection, refraction and caustics discuss the general features of reflected and refracted rays. With an ideal lens, light from any given point on an object would pass through the lens and come together at a single point in the "image plane" (or, more generally, the "image surface"). Real lenses do not focus light exactly to a single point, however, even when they are perfectly made. These deviations from the idealized lens performance are called "aberrations" of the lens. Aberrations fall into two classes: "monochromatic" and "chromatic". Monochromatic aberrations are caused by the geometry of the lens or mirror and occur both when light is reflected and when it is refracted. They appear even when using monochromatic light, hence the name. Chromatic aberrations are caused by dispersion, the variation of a lens's refractive index with wavelength. Because of dispersion, different wavelengths of light come to focus at different points. Chromatic aberration does not appear when monochromatic light is used. The most common monochromatic aberrations are: Although defocus is technically the lowest-order of the optical aberrations, it is usually not considered as a lens aberration, since it can be corrected by moving the lens (or the image plane) to bring the image plane to the optical focus of the lens. In addition to these aberrations, piston and tilt are effects which shift the position of the focal point. Piston and tilt are not true optical aberrations, since when an otherwise perfect wavefront is altered by piston and tilt, it will still form a perfect, aberration-free image, only shifted to a different position. Chromatic aberration occurs when different wavelengths are not focussed to the same point. Types of chromatic aberration are: A perfect optical system would follow the theorem: Rays of light proceeding from any "object point" unite in an "image point"; and therefore the "object space" is reproduced in an "image space." The introduction of simple auxiliary terms, due to Gauss, named the focal lengths and focal planes, permits the determination of the image of any object for any system (see lens). The Gaussian theory, however, is only true so long as the angles made by all rays with the optical axis (the symmetrical axis of the system) are infinitely small, "i.e". with infinitesimal objects, images and lenses; in practice these conditions may not be realized, and the images projected by uncorrected systems are, in general, ill-defined and often completely blurred, if the aperture or field of view exceeds certain limits. The investigations of James Clerk Maxwell and Ernst Abbe showed that the properties of these reproductions, "i.e". the relative position and magnitude of the images, are not special properties of optical systems, but necessary consequences of the supposition (per Abbe) of the reproduction of all points of a space in image points, and are independent of the manner in which the reproduction is effected. These authors showed, however, that no optical system can justify these suppositions, since they are contradictory to the fundamental laws of reflection and refraction. Consequently, the Gaussian theory only supplies a convenient method of approximating to reality; and actual optical systems can only attempt to realize this unattainable ideal. At present, all that can be attempted is to reproduce a single plane in another plane; but even this has not been altogether satisfactorily accomplished: aberrations always occur, and it is improbable that these will ever be entirely corrected. The classical theory of optics and related systems has been analyzed by numerous authors. Let S (fig. 1) be any optical system, rays proceeding from an axis point O under an angle u1 will unite in the axis point O'1; and those under an angle u2 in the axis point O'2. If there is refraction at a collective spherical surface, or through a thin positive lens, O'2 will lie in front of O'1 so long as the angle u2 is greater than u1 ("under correction"); and conversely with a dispersive surface or lenses ("over correction"). The caustic, in the first case, resembles the sign > (greater than); in the second < (less than). If the angle u1 is very small, O'1 is the Gaussian image; and O'1 O'2 is termed the "longitudinal aberration," and O'1R the "lateral aberration" of the pencils with aperture u2. If the pencil with the angle u2 is that of the maximum aberration of all the pencils transmitted, then in a plane perpendicular to the axis at O'1 there is a circular "disk of confusion" of radius O'1R, and in a parallel plane at O'2 another one of radius O'2R2; between these two is situated the "disk of least confusion." The largest opening of the pencils, which take part in the reproduction of O, "i.e". the angle u, is generally determined by the margin of one of the lenses or by a hole in a thin plate placed between, before, or behind the lenses of the system. This hole is termed the "stop" or "diaphragm"; Abbe used the term "aperture stop" for both the hole and the limiting margin of the lens. The component S1 of the system, situated between the aperture stop and the object O, projects an image of the diaphragm, termed by Abbe the "entrance pupil"; the "exit pupil" is the image formed by the component S2, which is placed behind the aperture stop. All rays which issue from O and pass through the aperture stop also pass through the entrance and exit pupils, since these are images of the aperture stop. Since the maximum aperture of the pencils issuing from O is the angle u subtended by the entrance pupil at this point, the magnitude of the aberration will be determined by the position and diameter of the entrance pupil. If the system be entirely behind the aperture stop, then this is itself the entrance pupil ("front stop"); if entirely in front, it is the exit pupil ("back stop"). If the object point be infinitely distant, all rays received by the first member of the system are parallel, and their intersections, after traversing the system, vary according to their "perpendicular height of incidence," i.e. their distance from the axis. This distance replaces the angle u in the preceding considerations; and the aperture, "i.e". the radius of the entrance pupil, is its maximum value. If rays issuing from O (fig. 1) are concurrent, it does not follow that points in a portion of a plane perpendicular at O to the axis will be also concurrent, even if the part of the plane be very small. As the diameter of the lens increases ("i.e.", with increasing aperture_, the neighboring point N will be reproduced, but attended by aberrations comparable in magnitude to ON. These aberrations are avoided if, according to Abbe, the "sine condition," sin u'1/sin u1=sin u'2/sin u2, holds for all rays reproducing the point O. If the object point O is infinitely distant, u1 and u2 are to be replaced by h1 and h2, the perpendicular heights of incidence; the "sine condition" then becomes sin u'1/h1=sin u'2/h2. A system fulfilling this condition and free from spherical aberration is called "aplanatic" (Greek a-, privative, plann, a wandering). This word was first used by Robert Blair to characterize a superior achromatism, and, subsequently, by many writers to denote freedom from spherical aberration as well. Since the aberration increases with the distance of the ray from the center of the lens, the aberration increases as the lens diameter increases (or, correspondingly, with the diameter of the aperture), and hence can be minimized by reducing the aperture, at the cost of also reducing the amount of light reaching the image plane. A point O (fig. 2) at a finite distance from the axis (or with an infinitely distant object, a point which subtends a finite angle at the system) is, in general, even then not sharply reproduced if the pencil of rays issuing from it and traversing the system is made infinitely narrow by reducing the aperture stop; such a pencil consists of the rays which can pass from the object point through the now infinitely small entrance pupil. It is seen (ignoring exceptional cases) that the pencil does not meet the refracting or reflecting surface at right angles; therefore it is astigmatic (Gr. a-, privative, stigmia, a point). Naming the central ray passing through the entrance pupil the "axis of the pencil" or "principal ray," it can be said: the rays of the pencil intersect, not in one point, but in two focal lines, which can be assumed to be at right angles to the principal ray; of these, one lies in the plane containing the principal ray and the axis of the system, i.e. in the "first principal section" or "meridional section", and the other at right angles to it, i.e. in the second principal section or sagittal section. We receive, therefore, in no single intercepting plane behind the system, as, for example, a focusing screen, an image of the object point; on the other hand, in each of two planes lines O' and O" are separately formed (in neighboring planes ellipses are formed), and in a plane between O' and O" a circle of least confusion. The interval O'O", termed the astigmatic difference, increases, in general, with the angle W made by the principal ray OP with the axis of the system, i.e. with the field of view. Two "astigmatic image surfaces" correspond to one object plane; and these are in contact at the axis point; on the one lie the focal lines of the first kind, on the other those of the second. Systems in which the two astigmatic surfaces coincide are termed anastigmatic or stigmatic. Sir Isaac Newton was probably the discoverer of astigmation; the position of the astigmatic image lines was determined by Thomas Young; and the theory was developed by Allvar Gullstrand. A bibliography by P. Culmann is given in Moritz von Rohr's "Die Bilderzeugung in optischen Instrumenten". By opening the stop wider, similar deviations arise for lateral points as have been already discussed for axial points; but in this case they are much more complicated. The course of the rays in the meridional section is no longer symmetrical to the principal ray of the pencil; and on an intercepting plane there appears, instead of a luminous point, a patch of light, not symmetrical about a point, and often exhibiting a resemblance to a comet having its tail directed towards or away from the axis. From this appearance it takes its name. The unsymmetrical form of the meridional pencil—formerly the only one considered—is coma in the narrower sense only; other errors of coma have been treated by Arthur König and Moritz von Rohr, and later by Allvar Gullstrand. If the above errors be eliminated, the two astigmatic surfaces united, and a sharp image obtained with a wide aperture—there remains the necessity to correct the curvature of the image surface, especially when the image is to be received upon a plane surface, e.g. in photography. In most cases the surface is concave towards the system. Even if the image is sharp, it may be distorted compared to ideal pinhole projection. In pinhole projection, the magnification of an object is inversely proportional to its distance to the camera along the optical axis so that a camera pointing directly at a flat surface reproduces that flat surface. Distortion can be thought of as stretching the image non-uniformly, or, equivalently, as a variation in magnification across the field. While "distortion" can include arbitrary deformation of an image, the most pronounced modes of distortion produced by conventional imaging optics is "barrel distortion", in which the center of the image is magnified more than the perimeter (figure 3a). The reverse, in which the perimeter is magnified more than the center, is known as "pincushion distortion" (figure 3b). This effect is called lens distortion or image distortion, and there are algorithms to correct it. Systems free of distortion are called "orthoscopic" (orthos, right, skopein to look) or "rectilinear" (straight lines). This aberration is quite distinct from that of the sharpness of reproduction; in unsharp, reproduction, the question of distortion arises if only parts of the object can be recognized in the figure. If, in an unsharp image, a patch of light corresponds to an object point, the "center of gravity" of the patch may be regarded as the image point, this being the point where the plane receiving the image, e.g., a focusing screen, intersects the ray passing through the middle of the stop. This assumption is justified if a poor image on the focusing screen remains stationary when the aperture is diminished; in practice, this generally occurs. This ray, named by Abbe a "principal ray" (not to be confused with the "principal rays" of the Gaussian theory), passes through the center of the entrance pupil before the first refraction, and the center of the exit pupil after the last refraction. From this it follows that correctness of drawing depends solely upon the principal rays; and is independent of the sharpness or curvature of the image field. Referring to fig. 4, we have O'Q'/OQ = a' tan w'/a tan w = 1/N, where N is the "scale" or magnification of the image. For N to be constant for all values of w, a' tan w'/a tan w must also be constant. If the ratio a'/a be sufficiently constant, as is often the case, the above relation reduces to the "condition of Airy," i.e. tan w'/ tan w= a constant. This simple relation (see Camb. Phil. Trans., 1830, 3, p. 1) is fulfilled in all systems which are symmetrical with respect to their diaphragm (briefly named "symmetrical or holosymmetrical objectives"), or which consist of two like, but different-sized, components, placed from the diaphragm in the ratio of their size, and presenting the same curvature to it (hemisymmetrical objectives); in these systems tan w' / tan w = 1. The constancy of a'/a necessary for this relation to hold was pointed out by R. H. Bow (Brit. Journ. Photog., 1861), and Thomas Sutton (Photographic Notes, 1862); it has been treated by O. Lummer and by M. von Rohr (Zeit. f. Instrumentenk., 1897, 17, and 1898, 18, p. 4). It requires the middle of the aperture stop to be reproduced in the centers of the entrance and exit pupils without spherical aberration. M. von Rohr showed that for systems fulfilling neither the Airy nor the Bow-Sutton condition, the ratio a' cos w'/a tan w will be constant for one distance of the object. This combined condition is exactly fulfilled by holosymmetrical objectives reproducing with the scale 1, and by hemisymmetrical, if the scale of reproduction be equal to the ratio of the sizes of the two components. Circular wavefront profiles associated with aberrations may be mathematically modeled using Zernike polynomials. Developed by Frits Zernike in the 1930s, Zernike's polynomials are orthogonal over a circle of unit radius. A complex, aberrated wavefront profile may be curve-fitted with Zernike polynomials to yield a set of fitting coefficients that individually represent different types of aberrations. These Zernike coefficients are linearly independent, thus individual aberration contributions to an overall wavefront may be isolated and quantified separately. There are even and odd Zernike polynomials. The even Zernike polynomials are defined as and the odd Zernike polynomials as where "m" and "n" are nonnegative integers with formula_3, Φ is the azimuthal angle in radians, and ρ is the normalized radial distance. The radial polynomials formula_4 have no azimuthal dependence, and are defined as and formula_6 if formula_7 is odd. The first few Zernike polynomials are: where formula_8 is the normalized pupil radius with formula_9, formula_10 is the azimuthal angle around the pupil with formula_11, and the fitting coefficients formula_12 are the wavefront errors in wavelengths. As in Fourier synthesis using sines and cosines, a wavefront may be perfectly represented by a sufficiently large number of higher-order Zernike polynomials. However, wavefronts with very steep gradients or very high spatial frequency structure, such as produced by propagation through atmospheric turbulence or aerodynamic flowfields, are not well modeled by Zernike polynomials, which tend to low-pass filter fine spatial definition in the wavefront. In this case, other fitting methods such as fractals or singular value decomposition may yield improved fitting results. The circle polynomials were introduced by Frits Zernike to evaluate the point image of an aberrated optical system taking into account the effects of diffraction. The perfect point image in the presence of diffraction had already been described by Airy, as early as 1835. It took almost hundred years to arrive at a comprehensive theory and modeling of the point image of aberrated systems (Zernike and Nijboer). The analysis by Nijboer and Zernike describes the intensity distribution close to the optimum focal plane. An extended theory that allows the calculation of the point image amplitude and intensity over a much larger volume in the focal region was recently developed (Extended Nijboer-Zernike theory). This Extended Nijboer-Zernike theory of point image or ‘point-spread function’ formation has found applications in general research on image formation, especially for systems with a high numerical aperture, and in characterizing optical systems with respect to their aberrations. The preceding review of the several errors of reproduction belongs to the "Abbe theory of aberrations," in which definite aberrations are discussed separately; it is well suited to practical needs, for in the construction of an optical instrument certain errors are sought to be eliminated, the selection of which is justified by experience. In the mathematical sense, however, this selection is arbitrary; the reproduction of a finite object with a finite aperture entails, in all probability, an infinite number of aberrations. This number is only finite if the object and aperture are assumed to be "infinitely small of a certain order"; and with each order of infinite smallness, i.e. with each degree of approximation to reality (to finite objects and apertures), a certain number of aberrations is associated. This connection is only supplied by theories which treat aberrations generally and analytically by means of indefinite series. A ray proceeding from an object point O (fig. 5) can be defined by the coordinates (ξ, η). Of this point O in an object plane I, at right angles to the axis, and two other coordinates (x, y), the point in which the ray intersects the entrance pupil, i.e. the plane II. Similarly the corresponding image ray may be defined by the points (ξ', η'), and (x', y'), in the planes I' and II'. The origins of these four plane coordinate systems may be collinear with the axis of the optical system; and the corresponding axes may be parallel. Each of the four coordinates ξ', η', x', y' are functions of ξ, η, x, y; and if it be assumed that the field of view and the aperture be infinitely small, then ξ, η, x, y are of the same order of infinitesimals; consequently by expanding ξ', η', x', y' in ascending powers of ξ, η, x, y, series are obtained in which it is only necessary to consider the lowest powers. It is readily seen that if the optical system be symmetrical, the origins of the coordinate systems collinear with the optical axis and the corresponding axes parallel, then by changing the signs of ξ, η, x, y, the values ξ', η', x', y' must likewise change their sign, but retain their arithmetical values; this means that the series are restricted to odd powers of the unmarked variables. The nature of the reproduction consists in the rays proceeding from a point O being united in another point O'; in general, this will not be the case, for ξ', η' vary if ξ, η be constant, but x, y variable. It may be assumed that the planes I' and II' are drawn where the images of the planes I and II are formed by rays near the axis by the ordinary Gaussian rules; and by an extension of these rules, not, however, corresponding to reality, the Gauss image point O'0, with coordinates ξ'0, η'0, of the point O at some distance from the axis could be constructed. Writing Dξ'=ξ'-ξ'0 and Dη'=η'-η'0, then Dξ' and Dη' are the aberrations belonging to ξ, η and x, y, and are functions of these magnitudes which, when expanded in series, contain only odd powers, for the same reasons as given above. On account of the aberrations of all rays which pass through O, a patch of light, depending in size on the lowest powers of ξ, η, x, y which the aberrations contain, will be formed in the plane I'. These degrees, named by J. Petzval ("Bericht uber die Ergebnisse einiger dioptrischer Untersuchungen", Buda Pesth, 1843; "Akad. Sitzber., Wien," 1857, vols. xxiv. xxvi.) "the numerical orders of the image," are consequently only odd powers; the condition for the formation of an image of the mth order is that in the series for Dξ' and Dη' the coefficients of the powers of the 3rd, 5th...(m-2)th degrees must vanish. The images of the Gauss theory being of the third order, the next problem is to obtain an image of 5th order, or to make the coefficients of the powers of 3rd degree zero. This necessitates the satisfying of five equations; in other words, there are five alterations of the 3rd order, the vanishing of which produces an image of the 5th order. The expression for these coefficients in terms of the constants of the optical system, i.e. the radii, thicknesses, refractive indices and distances between the lenses, was solved by L. Seidel (Astr. Nach., 1856, p. 289); in 1840, J. Petzval constructed his portrait objective, from similar calculations which have never been published (see M. von Rohr, "Theorie und Geschichte des photographischen Objectivs", Berlin, 1899, p. 248). The theory was elaborated by S. Finterswalder (Munchen. Acad. Abhandl., 1891, 17, p. 519), who also published a posthumous paper of Seidel containing a short view of his work ("München. Akad. Sitzber.," 1898, 28, p. 395); a simpler form was given by A. Kerber ("Beiträge zur Dioptrik", Leipzig, 1895-6-7-8-9). A. Konig and M. von Rohr (see M. von Rohr, "Die Bilderzeugung in optischen Instrumenten", pp. 317–323) have represented Kerber's method, and have deduced the Seidel formulae from geometrical considerations based on the Abbe method, and have interpreted the analytical results geometrically (pp. 212–316). The aberrations can also be expressed by means of the "characteristic function" of the system and its differential coefficients, instead of by the radii, &c., of the lenses; these formulae are not immediately applicable, but give, however, the relation between the number of aberrations and the order. Sir William Rowan Hamilton (British Assoc. Report, 1833, p. 360) thus derived the aberrations of the third order; and in later times the method was pursued by Clerk Maxwell ("Proc. London Math. Soc.," 1874–1875; (see also the treatises of R. S. Heath and L. A. Herman), M. Thiesen ("Berlin. Akad. Sitzber.," 1890, 35, p. 804), H. Bruns ("Leipzig. Math. Phys. Ber.," 1895, 21, p. 410), and particularly successfully by K. Schwarzschild ("Göttingen. Akad. Abhandl.," 1905, 4, No. 1), who thus discovered the aberrations of the 5th order (of which there are nine), and possibly the shortest proof of the practical (Seidel) formulae. A. Gullstrand (vide supra, and "Ann. d. Phys.," 1905, 18, p. 941) founded his theory of aberrations on the differential geometry of surfaces. The aberrations of the third order are: (1) aberration of the axis point; (2) aberration of points whose distance from the axis is very small, less than of the third order — the deviation from the sine condition and coma here fall together in one class; (3) astigmatism; (4) curvature of the field; (5) distortion. The classical imaging problem is to reproduce perfectly a finite plane (the object) onto another plane (the image) through a finite aperture. It is impossible to do so perfectly for "more than one" such pair of planes (this was proven with increasing generality by Maxwell in 1858, by Bruns in 1895, and by Carathéodory in 1926, see summary in Walther, A., J. Opt. Soc. Am. A 6, 415–422 (1989)). For a single pair of planes (e.g. for a single focus setting of an objective), however, the problem can in principle be solved perfectly. Examples of such a theoretically perfect system include the Luneburg lens and the Maxwell fish-eye. Practical methods solve this problem with an accuracy which mostly suffices for the special purpose of each species of instrument. The problem of finding a system which reproduces a given object upon a given plane with given magnification (insofar as aberrations must be taken into account) could be dealt with by means of the approximation theory; in most cases, however, the analytical difficulties were too great for older calculation methods but may be ameliorated by application of modern computer systems. Solutions, however, have been obtained in special cases (see A. Konig in M. von Rohr's "Die Bilderzeugung", p. 373; K. Schwarzschild, Göttingen. Akad. Abhandl., 1905, 4, Nos. 2 and 3). At the present time constructors almost always employ the inverse method: they compose a system from certain, often quite personal experiences, and test, by the trigonometrical calculation of the paths of several rays, whether the system gives the desired reproduction (examples are given in A. Gleichen, "Lehrbuch der geometrischen Optik", Leipzig and Berlin, 1902). The radii, thicknesses and distances are continually altered until the errors of the image become sufficiently small. By this method only certain errors of reproduction are investigated, especially individual members, or all, of those named above. The analytical approximation theory is often employed provisionally, since its accuracy does not generally suffice. In order to render spherical aberration and the deviation from the sine condition small throughout the whole aperture, there is given to a ray with a finite angle of aperture u* (width infinitely distant objects: with a finite height of incidence h*) the same distance of intersection, and the same sine ratio as to one neighboring the axis (u* or h* may not be much smaller than the largest aperture U or H to be used in the system). The rays with an angle of aperture smaller than u* would not have the same distance of intersection and the same sine ratio; these deviations are called "zones," and the constructor endeavors to reduce these to a minimum. The same holds for the errors depending upon the angle of the field of view, w: astigmatism, curvature of field and distortion are eliminated for a definite value, w*, "zones of astigmatism, curvature of field and distortion," attend smaller values of w. The practical optician names such systems: "corrected for the angle of aperture u* (the height of incidence h*) or the angle of field of view w*." Spherical aberration and changes of the sine ratios are often represented graphically as functions of the aperture, in the same way as the deviations of two astigmatic image surfaces of the image plane of the axis point are represented as functions of the angles of the field of view. The final form of a practical system consequently rests on compromise; enlargement of the aperture results in a diminution of the available field of view, and vice versa. But the larger aperture will give the larger resolution. The following may be regarded as typical: In optical systems composed of lenses, the position, magnitude and errors of the image depend upon the refractive indices of the glass employed (see Lens (optics) and Monochromatic aberration, above). Since the index of refraction varies with the color or wavelength of the light (see dispersion), it follows that a system of lenses (uncorrected) projects images of different colors in somewhat different places and sizes and with different aberrations; i.e. there are "chromatic differences" of the distances of intersection, of magnifications, and of monochromatic aberrations. If mixed light be employed (e.g. white light) all these images are formed and they cause a confusion, named chromatic aberration; for instance, instead of a white margin on a dark background, there is perceived a colored margin, or narrow spectrum. The absence of this error is termed achromatism, and an optical system so corrected is termed achromatic. A system is said to be "chromatically under-corrected" when it shows the same kind of chromatic error as a thin positive lens, otherwise it is said to be "overcorrected." If, in the first place, monochromatic aberrations be neglected — in other words, the Gaussian theory be accepted — then every reproduction is determined by the positions of the focal planes, and the magnitude of the focal lengths, or if the focal lengths, as ordinarily happens, be equal, by three constants of reproduction. These constants are determined by the data of the system (radii, thicknesses, distances, indices, etc., of the lenses); therefore their dependence on the refractive index, and consequently on the color, are calculable. The refractive indices for different wavelengths must be known for each kind of glass made use of. In this manner the conditions are maintained that any one constant of reproduction is equal for two different colors, i.e. this constant is achromatized. For example, it is possible, with one thick lens in air, to achromatize the position of a focal plane of the magnitude of the focal length. If all three constants of reproduction be achromatized, then the Gaussian image for all distances of objects is the same for the two colors, and the system is said to be in "stable achromatism." In practice it is more advantageous (after Abbe) to determine the chromatic aberration (for instance, that of the distance of intersection) for a fixed position of the object, and express it by a sum in which each component conlins the amount due to each refracting surface. In a plane containing the image point of one color, another colour produces a disk of confusion; this is similar to the confusion caused by two "zones" in spherical aberration. For infinitely distant objects the radius Of the chromatic disk of confusion is proportional to the linear aperture, and independent of the focal length ("vide supra", "Monochromatic Aberration of the Axis Point"); and since this disk becomes the less harmful with an increasing image of a given object, or with increasing focal length, it follows that the deterioration of the image is proportional to the ratio of the aperture to the focal length, i.e. the "relative aperture." (This explains the gigantic focal lengths in vogue before the discovery of achromatism.) Examples: Newton failed to perceive the existence of media of different dispersive powers required by achromatism; consequently he constructed large reflectors instead of refractors. James Gregory and Leonhard Euler arrived at the correct view from a false conception of the achromatism of the eye; this was determined by Chester More Hall in 1728, Klingenstierna in 1754 and by Dollond in 1757, who constructed the celebrated achromatic telescopes. (See telescope.) Glass with weaker dispersive power (greater formula_43) is named "crown glass"; that with greater dispersive power, "flint glass". For the construction of an achromatic collective lens (formula_15 positive) it follows, by means of equation (4), that a collective lens I. of crown glass and a dispersive lens II. of flint glass must be chosen; the latter, although the weaker, corrects the other chromatically by its greater dispersive power. For an achromatic dispersive lens the converse must be adopted. This is, at the present day, the ordinary type, e.g., of telescope objective; the values of the four radii must satisfy the equations (2) and (4). Two other conditions may also be postulated: one is always the elimination of the aberration on the axis; the second either the "Herschel" or "Fraunhofer Condition," the latter being the best vide supra, "Monochromatic Aberration"). In practice, however, it is often more useful to avoid the second condition by making the lenses have contact, i.e. equal radii. According to P. Rudolph (Eder's Jahrb. f. Photog., 1891, 5, p. 225; 1893, 7, p. 221), cemented objectives of thin lenses permit the elimination of spherical aberration on the axis, if, as above, the collective lens has a smaller refractive index; on the other hand, they permit the elimination of astigmatism and curvature of the field, if the collective lens has a greater refractive index (this follows from the Petzval equation; see L. Seidel, Astr. Nachr., 1856, p. 289). Should the cemented system be positive, then the more powerful lens must be positive; and, according to (4), to the greater power belongs the weaker dispersive power (greater formula_43), that is to say, crown glass; consequently the crown glass must have the greater refractive index for astigmatic and plane images. In all earlier kinds of glass, however, the dispersive power increased with the refractive index; that is, formula_43 decreased as formula_13 increased; but some of the Jena glasses by E. Abbe and O. Schott were crown glasses of high refractive index, and achromatic systems from such crown glasses, with flint glasses of lower refractive index, are called the "new achromats," and were employed by P. Rudolph in the first "anastigmats" (photographic objectives). Instead of making formula_29 vanish, a certain value can be assigned to it which will produce, by the addition of the two lenses, any desired chromatic deviation, e.g. sufficient to eliminate one present in other parts of the system. If the lenses I. and II. be cemented and have the same refractive index for one color, then its effect for that one color is that of a lens of one piece; by such decomposition of a lens it can be made chromatic or achromatic at will, without altering its spherical effect. If its chromatic effect (formula_49) be greater than that of the same lens, this being made of the more dispersive of the two glasses employed, it is termed "hyper-chromatic." For two thin lenses separated by a distance formula_50 the condition for achromatism is formula_51; if formula_52 (e.g. if the lenses be made of the same glass), this reduces to formula_53, known as the "condition for oculars." If a constant of reproduction, for instance the focal length, be made equal for two colors, then it is not the same for other colors, if two different glasses are employed. For example, the condition for achromatism (4) for two thin lenses in contact is fulfilled in only one part of the spectrum, since formula_54 varies within the spectrum. This fact was first ascertained by J. Fraunhofer, who defined the colors by means of the dark lines in the solar spectrum; and showed that the ratio of the dispersion of two glasses varied about 20% from the red to the violet (the variation for glass and water is about 50%). If, therefore, for two colors, a and b, formula_55, then for a third color, c, the focal length is different; that is, if c lies between a and b, then formula_56, and vice versa; these algebraic results follow from the fact that towards the red the dispersion of the positive crown glass preponderates, towards the violet that of the negative flint. These chromatic errors of systems, which are achromatic for two colors, are called the "secondary spectrum," and depend upon the aperture and focal length in the same manner as the primary chromatic errors do. In fig. 6, taken from M. von Rohr's "Theorie und Geschichte des photographischen Objectivs", the abscissae are focal lengths, and the ordinates wavelengths. The Fraunhofer lines used are shown in adjacent table. The focal lengths are made equal for the lines C and F. In the neighborhood of 550 nm the tangent to the curve is parallel to the axis of wavelengths; and the focal length varies least over a fairly large range of color, therefore in this neighborhood the color union is at its best. Moreover, this region of the spectrum is that which appears brightest to the human eye, and consequently this curve of the secondary on spectrum, obtained by making formula_57, is, according to the experiments of Sir G. G. Stokes (Proc. Roy. Soc., 1878), the most suitable for visual instruments ("optical achromatism,"). In a similar manner, for systems used in photography, the vertex of the color curve must be placed in the position of the maximum sensibility of the plates; this is generally supposed to be at G'; and to accomplish this the F and violet mercury lines are united. This artifice is specially adopted in objectives for astronomical photography ("pure actinic achromatism"). For ordinary photography, however, there is this disadvantage: the image on the focusing-screen and the correct adjustment of the photographic sensitive plate are not in register; in astronomical photography this difference is constant, but in other kinds it depends on the distance of the objects. On this account the lines D and G' are united for ordinary photographic objectives; the optical as well as the actinic image is chromatically inferior, but both lie in the same place; and consequently the best correction lies in F (this is known as the "actinic correction" or "freedom from chemical focus"). Should there be in two lenses in contact the same focal lengths for three colours a, b, and c, i.e. formula_58, then the relative partial dispersion formula_59 must be equal for the two kinds of glass employed. This follows by considering equation (4) for the two pairs of colors ac and bc. Until recently no glasses were known with a proportional degree of absorption; but R. Blair (Trans. Edin. Soc., 1791, 3, p. 3), P. Barlow, and F. S. Archer overcame the difficulty by constructing fluid lenses between glass walls. Fraunhofer prepared glasses which reduced the secondary spectrum; but permanent success was only assured on the introduction of the Jena glasses by E. Abbe and O. Schott. In using glasses not having proportional dispersion, the deviation of a third colour can be eliminated by two lenses, if an interval be allowed between them; or by three lenses in contact, which may not all consist of the old glasses. In uniting three colors an "achromatism of a higher order" is derived; there is yet a residual "tertiary spectrum," but it can always be neglected. The Gaussian theory is only an approximation; monochromatic or spherical aberrations still occur, which will be different for different colors; and should they be compensated for one color, the image of another color would prove disturbing. The most important is the chromatic difference of aberration of the axis point, which is still present to disturb the image, after par-axial rays of different colors are united by an appropriate combination of glasses. If a collective system be corrected for the axis point for a definite wavelength, then, on account of the greater dispersion in the negative components — the flint glasses, — overcorrection will arise for the shorter wavelengths (this being the error of the negative components), and under-correction for the longer wavelengths (the error of crown glass lenses preponderating in the red). This error was treated by Jean le Rond d'Alembert, and, in special detail, by C. F. Gauss. It increases rapidly with the aperture, and is more important with medium apertures than the secondary spectrum of par-axial rays; consequently, spherical aberration must be eliminated for two colors, and if this be impossible, then it must be eliminated for those particular wavelengths which are most effectual for the instrument in question (a graphical representation of this error is given in M. von Rohr, "Theorie und Geschichte des photographischen Objectivs"). The condition for the reproduction of a surface element in the place of a sharply reproduced point — the constant of the sine relationship must also be fulfilled with large apertures for several colors. E. Abbe succeeded in computing microscope objectives free from error of the axis point and satisfying the sine condition for several colors, which therefore, according to his definition, were "aplanatic for several colors"; such systems he termed "apochromatic". While, however, the magnification of the individual zones is the same, it is not the same for red as for blue; and there is a chromatic difference of magnification. This is produced in the same amount, but in the opposite sense, by the oculars, which Abbe used with these objectives ("compensating oculars"), so that it is eliminated in the image of the whole microscope. The best telescope objectives, and photographic objectives intended for three-color work, are also apochromatic, even if they do not possess quite the same quality of correction as microscope objectives do. The chromatic differences of other errors of reproduction have seldom practical importances.
https://en.wikipedia.org/wiki?curid=2704
Amy Grant Amy Lee Grant (born November 25, 1960) is an American singer, songwriter, musician, author and media personality. She is known for performing contemporary Christian music (CCM) and for a successful crossover to pop music in the 1980s and 1990s. She has been referred to as "The Queen of Christian Pop". , she had sold more than 30 million albums worldwide, won six Grammy Awards and 22 Gospel Music Association Dove Awards, and had the first Christian album to go Platinum. She was honored with a star on Hollywood Walk of Fame in 2006 for her contributions to the entertainment industry. She made her debut as a teenager, and gained fame in Christian music during the 1980s with such hits as "Father's Eyes", "El Shaddai", and "Angels". In the mid-1980s, she began broadening her audience and soon became one of the first CCM artists to cross over into mainstream pop on the heels of her successful albums "Unguarded" and "Lead Me On". In 1986, she scored her first "Billboard" Hot 100 No. 1 song in a duet with Peter Cetera, "The Next Time I Fall". In 1991, she released the blockbuster album "Heart in Motion" which became her best-selling album to date, topping the "Billboard" Christian album chart for 32 weeks, selling five million copies in the U.S., and producing her second No. 1 pop single "Baby Baby". She is the author of several books, including a memoir, "", and a book based on the popular Christmas song "Breath of Heaven (Mary's Song)" which she co-wrote. Born in Augusta, Georgia, Grant is the youngest of four sisters. Her family settled in Nashville, Tennessee, in 1967. She is a great-granddaughter of Nashville philanthropist A. M. Burton (founder of Life and Casualty Insurance Company, eponym of Nashville's Life & Casualty Tower, WLAC Radio, and WLAC-TV) and Lillie Burton. She has acknowledged the influence of the Burtons on her development as a musician, starting with their common membership in Nashville's Ashwood Church of Christ. In 1976, Grant wrote her first song ("Mountain Top"), performed in public for the first time at Harpeth Hall School, the all-girls school she attended in Nashville. She recorded a demo tape for her parents with church youth-leader Brown Bannister. When Bannister was dubbing a copy of the tape, Chris Christian, the owner of the recording studio, heard the demo and called Word Records. He played it over the phone, and she was offered a recording contract, five weeks before her 16th birthday. In 1977, she recorded her first album, "Amy Grant", produced by Brown Bannister, who would also produce her next 11 albums. It was released in early 1978, one month before her high-school graduation. Toward the end of 1978 she performed her first ticketed concert after beginning her first year at Furman University. In May 1979, while at the album-release party for her second album, "My Father's Eyes", Grant met Gary Chapman, who had written the title track and would become her first husband. Grant and Chapman toured together in mid-1979. In late 1980, she transferred to Vanderbilt University, where she was a member of the sorority Kappa Alpha Theta. Grant then made a few more albums before dropping out of college to pursue a career in music—"Never Alone", followed by a pair of live albums in 1981 ("In Concert" and "In Concert Volume Two"), both backed by an augmented edition of the DeGarmo & Key band. It was during these early shows that Grant also established one of her concert trademarks: performing barefoot. To date, Grant continues to take off her shoes midway through performances, as she has said, "it is just more comfortable." 1982 saw the release of her breakthrough album "Age to Age". The album contains the signature track, "El Shaddai" (written by Michael Card) and the Grant-Chapman penned song, "In a Little While". "El Shaddai" was later awarded one of the "Songs of the Century" by the RIAA in 2001. Grant received her first Grammy Award for Best Contemporary Gospel Performance, as well as two GMA Dove Awards for Gospel Artist of the Year and Pop/Contemporary Album of the Year. "Age to Age" became the first Christian album by a solo artist to be certified gold (1983) and the first Christian album to be certified platinum (1985). In the mid-1980s, Grant began touring and recording with young up-and-coming songwriter Michael W. Smith. Grant and Smith continue to have a strong friendship and creative relationship, often writing songs for or contributing vocals to each other's albums, and as of 2019, often touring together annually during November and December putting on Christmas concerts. During the 1980s, Grant was also a backup singer for Bill Gaither. Grant followed this album with the first of her Christmas albums, which would later be the basis for her holiday shows. In 1984, she released another pop-oriented Christian hit, "Straight Ahead", earning Grant her first appearance at the Grammy Awards show in 1985. The head of NBC took notice of Grant's performance and called her manager to book her for her own Christmas special. Hardly had Grant established herself as the "Queen of Christian Pop" when she changed directions to widen her fan base (and hence her musical message). Her goal was to become the first Christian singer-songwriter who was also successful as a contemporary pop singer. "Unguarded" (1985) surprised some fans for its very mainstream sound (and Grant's leopard-print jacket, in four poses for four different covers). "Find a Way", from "Unguarded", became the first non-Christmas Christian song to hit "Billboard" Top 40 list, also reaching No. 7 on the Adult Contemporary chart. She also scored No. 18 on "Billboard" AC in 1986 with "Stay for Awhile". Grant scored her first "Billboard" No. 1 song in 1986 with "The Next Time I Fall", a duet with former Chicago singer/bassist Peter Cetera. That year, she also recorded a duet with singer Randy Stonehill for his "Love Beyond Reason" album, titled "I Could Never Say Goodbye", and recorded "The Animals' Christmas" with Art Garfunkel. "Lead Me On" (1988) contained many songs that were about Christianity and love relationships, but some interpreted it as not being an obviously "Christian" record. Years later, "Lead Me On" would be chosen as the greatest Contemporary Christian album of all time by "CCM Magazine". The mainstream song "Saved by Love" was a minor hit, receiving airplay on radio stations featuring the newly emerging Adult Contemporary format. The album's title song received some pop radio airplay and crossed over to No. 96 on the "Billboard" Hot 100, and "1974 (We Were Young)" and "Saved By Love" also charted as Adult Contemporary songs. In 1989, she appeared in a Target ad campaign, performing songs off the album. When "Heart in Motion" was released in 1991, many fans were surprised that the album was so clearly one of contemporary pop music. Grant's desire to widen her audience was frowned upon by the confines of the popular definitions of ministry at the time. The track "Baby Baby" (written for Grant's newborn daughter Millie, of whom Grant wrote, her "six-week-old face was my inspiration") became a pop hit (hitting No. 1 on the "Billboard" Hot 100), and Grant was established as a name in the mainstream music world. "Baby Baby" received Grammy nominations for Best Female Pop Vocal Performance, and Record and Song of the Year (although it failed to win in any of those categories). Four other hits from the album made the Pop top 20: "Every Heartbeat" (No. 2), "That's What Love Is For" (No. 7), "Good for Me" (No. 8), and "I Will Remember You" (No. 20). On the Adult Contemporary chart, all five songs were top 10 hits, with two of the five ("Baby Baby" and "That's What Love Is For") reaching No. 1. Many Christian fans remained loyal, putting the album atop "Billboard" Contemporary Christian Chart for 32 weeks. "Heart in Motion" is Grant's best-selling album, having sold over five million copies according to the RIAA. Grant followed the album with her second Christmas album, "Home For Christmas" in 1992, which included the song "Breath of Heaven (Mary's Song)", written by Chris Eaton and Grant, and would later be covered by many artists, including Donna Summer, Jessica Simpson (who acknowledged Grant as one of her favorite artists), Vince Gill, Sara Groves, Point of Grace, Gladys Knight, and Broadway star Barbara Cook. "House of Love" in 1994 continued in the same vein, boasting catchy pop songs mingled with spiritual lyrics. The album was a multi-platinum success and produced the pop hit "Lucky One" (No. 18 pop and No. 2 AC; No. 1 on Radio & Records) as well as the title track (a duet with country music star and future husband Vince Gill) (No. 37 pop) and a cover of Joni Mitchell's frequently covered "Big Yellow Taxi" (No. 67 pop) (in which she changed the line "And they charged the people "a dollar and a half" just to see'em" to "And then they charged the people "25 bucks" just to see'em"). After she covered the 10cc song "The Things We Do for Love" for the "Mr. Wrong" soundtrack, "Behind the Eyes" was released in September 1997. The album struck a much darker note, leaning more towards downtempo, acoustic soft-rock songs, with more mature (yet still optimistic) lyrics. She called it her "razor blades and Prozac" album. Although "Takes a Little Time" was a moderate hit single, the album failed to sell like the previous two albums, which had both gone multi-platinum. "Behind The Eyes" was eventually certified Gold by the RIAA. The video for "Takes a Little Time" was a new direction for Grant; with a blue light filter, acoustic guitar, the streets and characters of New York City, and a plot, Grant was re-cast as an adult light rocker. She followed up "Behind The Eyes" with "A Christmas To Remember", her third Christmas album, in 1999. The album was certified Gold in 2000. In 2001, Grant sang "God Bless America" in front of a sellout crowd at the Owen County Fair Grounds in Spencer, Indiana. She dedicated her performance to the victims of 9/11, and officially started the Demolition Derby. Following the 9/11 attacks, Grant's "I Will Remember You" saw a resurgence in popularity as many radio DJs mixed a special tribute version of the song. That same year, Grant won $125,000 for charity on the "Rock Star Edition" of "Who Wants to Be a Millionaire?" Grant returned to her roots with the 2002 release of an album of hymns titled "Legacy... Hymns and Faith". The album featured a Vince Gill-influenced mix of bluegrass and pop and marked Grant's 25th anniversary in the music industry. Grant followed this up with "Simple Things" in 2003. The album did not have the success of her previous pop or gospel efforts. Soon after "Simple Things", Grant and Interscope/A&M parted ways. The same year, Grant was inducted into the Gospel Music Hall of Fame by the Gospel Music Association, an industry trade organization of which she is a longstanding member, in her first year of eligibility. Grant released a sequel in 2005 titled "Rock of Ages...Hymns and Faith". Grant joined the reality television phenomenon by hosting "Three Wishes", a show in which she and a team of helpers make wishes come true for small-town residents. The show debuted on NBC in the fall of 2005 but was canceled at the end of its first season because of high production costs. After "Three Wishes" was canceled, Grant won her 6th Grammy Award for "Rock of Ages... Hymns & Faith". In a February 2006 webchat, Grant stated she believes her "best music is still ahead". In April 2006, a live CD/DVD titled "Time Again... Amy Grant Live" was recorded in Fort Worth, Texas, at Bass Performance Hall. (Grant's first paid public performance was at the Will Rogers Auditorium in Fort Worth.) The concert was released on September 26, 2006. In addition to receiving a star on the Hollywood Walk of Fame, media appearances included write-ups in "CCM Magazine", and a performance on "The View". In a February 2007 web chat on her web site, Grant discussed a book she was working on titled "", saying, "It's not an autobiography, but more a collection of memories, song lyrics, poetry and a few pictures." The book was released on October 16, 2007. In November, it debuted at No. 35 on the New York Times Best Seller list. In the same web chat, Grant noted that she is "anxious to get back in the studio after the book is finished, and reinvent myself as an almost-50 performing woman". 2007 was Grant's 30th year in music. She left Word/Warner, and contracted with EMI CMG who re-released her regular studio albums as remastered versions on August 14, 2007. Marking the start of Grant's new contract is a career-spanning greatest hits album, with all the songs digitally remastered. The album was released as both a single-disc CD edition, and a two-disc CD/DVD Special Edition, the DVD featuring music videos and interviews. Grant appeared with Gill on "The Oprah Winfrey Show" for a holiday special in December 2007. Grant has plans to appear on CMT, a Food Network special, the Gospel Music Channel, and The Hour of Power. In February 2008, Grant joined the writing team from Compassionart as a guest vocalist at the Abbey Road studios, London, to record a song called "Highly Favoured", which was included on the album "CompassionArt". On June 24, 2008, Grant re-released her 1988 album, "Lead Me On", in honor of its 20th anniversary. The two-disc release includes the original album and a second disc with new acoustic recordings, live performances from 1989, and interviews with Amy. Grant recreated the "Lead Me On" tour in the fall of 2008. On June 27, 2008, at Creation Festival Northeast she performed "Lead Me On" and a few other songs backed by Hawk Nelson. At the end of the concert, Grant returned to the stage and sang "Thy Word". She appeared on the 2008 album "" singing "Could I Have This Dance". On May 5, 2009, Grant released an EP containing two new songs, "She Colors My Day", and "Unafraid", as well as the previously released songs "Baby Baby" and "Oh How the Years Go By". The EP, exclusively through iTunes, benefited the Entertainment Industry Foundation's (EIF) Women's Cancer Research Fund. In 2010, Grant released "Somewhere Down the Road", featuring the hit single "Better Than a Hallelujah", which peaked at No. 8 on "Billboard" Top Christian Songs chart. When asked about the new album during an interview with CBN.com, Grant says, "... my hope is just for those songs to provide companionship, remind myself and whoever else is listening what's important. I feel like songs have the ability to connect us to ourselves and to each other, and to our faith, to the love of Jesus, in a way that conversation doesn't do. Songs kind of slip in and move you before you realize it." In September 2012, Grant took part in a campaign called "30 Songs / 30 Days" to support "", a multi-platform media project inspired by Nicholas Kristof and Sheryl WuDunn's book. Grant's next album, "How Mercy Looks from Here", was released on May 14, 2013 and was produced by Marshall Altman. The album reached No. 12 on the "Billboard" 200 chart, making it her highest charting album since 1997's "Behind the Eyes". Two singles were released from the album: "Don't Try So Hard" and "If I Could See", both of which charted on the US "Billboard" Hot Christian Songs chart. On August 19, 2014, she released an album of hits remixed by well known engineers and DJs. The album was titled "In Motion: The Remixes". It charted at 110 on the US "Billboard" 200 chart and at No. 5 on the US Dance chart. To promote the album, several new remix EPs were released on iTunes the following month including "Find a Way, "Stay for Awhile", "Baby Baby, "Every Heartbeat" and "That's What Love Is For". Due to club play of the remixes of "Baby Baby" and "Every Heartbeat", they charted at No. 3 and 13, respectively on the U.S. Dance Chart. This marked her first appearance on that chart in 23 years. On September 30, 2014, Grant released a new single titled "Welcome Yourself". In honor of Breast Cancer Awareness Month, proceeds of the single go to breast cancer research. On February 12, 2015, she announced a new compilation album titled "Be Still and Know... Hymns & Faith", to be released. The album was released on April 14, 2015 and charted at No. 7 in the U.S. on the "Billboard" Christian Albums chart. . Grant released a Christmas album on October 21, 2016, "Tennessee Christmas", which is a combination of classic Christmas songs and original material. It charted in the U.S. at No. 31 on the "Billboard" 200 and at No. 3 on the "Billboard" Top Holiday Albums chart. The single from the album, "To Be Together", reached No. 32 on the Hot Christian Songs chart and No. 19 on the Holiday Digital Song Sales chart. She supported the album with a series of Christmas concerts with Vince Gill at the Ryman Auditorium. She also toured the U.S. and Canada with Christmas concerts accompanied by Michael W. Smith and season 9 winner of "The Voice", Jordan Smith. In February 2017, she released a new song, "Say It With a Kiss", with accompanying video. During November and December 2017, Grant performed another series of Christmas concerts with Vince Gill at the Ryman and embarked on another U.S. and Canada Christmas tour with Michael W. Smith and Jordan Smith. On June 25, 2019, "The New York Times Magazine" listed Amy Grant among hundreds of artists whose material was reportedly destroyed in the 2008 Universal fire. On June 19, 1982, Grant married fellow Christian musician Gary Chapman. Their marriage produced three children. In March 1999, she filed for divorce from Chapman, citing "irreconcilable differences", and the divorce was finalized three months later. On March 10, 2000, Grant married country singer-songwriter Vince Gill, who had been previously married to country singer Janis Oliver of Sweethearts of the Rodeo. Grant and Gill have one daughter together, Corrina Grant Gill, born March 12, 2001. In the November 1999 "CCM Magazine", Grant explained why she left Chapman and married Gill: In June 2020, Grant had heart surgery to repair partial anomalous pulmonary venous return (PAPVR), a congenital heart condition. Among praise for her contributions to the Contemporary Christian genre, Grant has also generated controversy within the Christian community, from "complaints that she was too worldly and too sexy" to a "barrage of condemnation" following her divorce and remarriage. In an interview early in her career, Grant stated, "I have a healthy sense of right and wrong, but sometimes, for example, using foul, exclamation-point words among friends can be good for a laugh." The article which was based on that interview was constructed in such a manner so as to make it appear as though Grant condoned premarital sex. Later Grant reflected on how the article misrepresented her views, stating: "We probably talked for two hours about sexual purity, but when the interview finally came out he worded it in such a way that it sounded like I condoned premarital sex. So I picked up that article and thought, 'You've made me say something I've never said, and you've totally disregarded two hours of Bible put in one flippant comment that I made about a moan.'"
https://en.wikipedia.org/wiki?curid=2705
Aberdeen, South Dakota Aberdeen (Lakota: "Ablíla") is a city in and the county seat of Brown County, South Dakota, United States, located approximately northeast of Pierre. The city population was 26,091 at the 2010 census, making it the third most populous city in the state after Sioux Falls and Rapid City. Aberdeen is the principal city of the Aberdeen Micropolitan Statistical Area, which includes all of Brown and Edmunds counties and has a population of 40,602 in 2010. Aberdeen is considered a college town, being the home of both Northern State University and Presentation College. Before Aberdeen or Brown County was inhabited by European settlers, it was inhabited by the Sioux Indians from approximately 1700 to 1879. Europeans entered the region for business, founding fur trading posts during the 1820s; these trading posts operated until the mid-1830s. The first "settlers" of this region were the Arikara Indians, but they would later be joined by others. The first group of Euro-American settlers to reach the area that is now Brown County was a party of four people, three horses, two mules, fifteen cattle, and two wagons. This group of settlers was later joined by another group the following spring, and, eventually, more settlers migrated toward this general area, currently known as Columbia, South Dakota. This town was established on June 15, 1879. The town was settled in 1880, and incorporated in 1882. Aberdeen, like many towns of the Midwest, was built around the newly developing railroad systems. Aberdeen was first officially plotted as a town site on January 3, 1881, by Charles Prior, the superintendent of the Minneapolis office of the Chicago, Milwaukee, and St. Paul Railroad, or the Milwaukee Road for short, which was presided over by Alexander Mitchell. Mitchell, Charles Prior's boss, was responsible for the choice of town names, was born in Aberdeen, Scotland, after which the town of Aberdeen, South Dakota, was named. Aberdeen was officially founded on July 6, 1881, the date of the first arrival of a Milwaukee Railroad train. Aberdeen then operated under a city charter granted by the Territorial Legislature in March 1883. As Aberdeen grew, many businesses and buildings were constructed along Aberdeen's Main Street. However, this soon became a problem due to Aberdeen's periodic flooding, which led to it being referred to as "The Town in the Frog Pond". At first, this unique condition presented no problem to the newly constructed buildings because it had not rained very much but, when heavy rains fell, the Pond reappeared and flooded the basements of every building on Main Street, causing many business owners and home owners much turmoil. When this flooding happened, the city had one steam-powered pump that had to be used to dry out the entire area that had been flooded, which would take days, if not weeks – and more often than not, it would have rained again in this time period and caused even more flooding, even in the basements that had already been emptied of the water. When the water was gone from the basements, the city still had to deal with the mud that also resulted from the heavy rains. The city decided in 1882 to build an artesian ditch to control the "Frog Pond" effects; the plan was later upgraded and developed into an artesian well in 1884 to combat the heavy rains and keep the basements from flooding. The artesian well was designed by the city engineers to prevent flooding and develop a water system. However, during the digging of the well, the water stream that was found underground was too powerful to be contained. The water came blasting out with violent force and had the entire Main Street submerged in up to four feet of water. The engineers realized the previous flaws of the artesian well plan and soon added a gate valve to the well to control the flow of water, giving Aberdeen its first working water supply. Aberdeen had four different railroad companies with depots built in the newly developing town. With these four railroads intersecting here, Aberdeen soon became known as the "Hub City of the Dakotas". When looking down on Aberdeen from above, the railroad tracks converging in Aberdeen resembled the spokes of a wheel converging at a hub, hence the name "Hub City of the Dakotas". These four railroad companies are the reason why Aberdeen was able to grow and flourish as it did. The only railroad still running through Aberdeen is the Burlington Northern Santa Fe. On October 25, 1999, a Learjet 35 carrying golfing star Payne Stewart and five others crashed in a field near Mina, west of Aberdeen. All on board died. Aberdeen is located in northeastern South Dakota, in the James River valley, approximately west of the river. The James River enters northeastern South Dakota in Brown County, where it is dammed to form two reservoirs northeast of Aberdeen. The city is bisected by "Moccasin Creek", a slow-moving waterway which flows south and then northeast to the James River. According to the United States Census Bureau, the city has a total area of , of which is land and is water. Aberdeen has been assigned the ZIP code range 57401−57402. Aberdeen experiences a humid continental climate (Köppen "Dfb/Dfa") influenced by its position far from moderating bodies of water. This brings four distinct seasons, a phenomenon that is characterized by hot, relatively humid summers and cold, dry winters, and it lies in USDA Hardiness Zone 4. The monthly daily average temperature ranges from in January to in July, while there are 13 days of + highs and 37 days with sub- lows annually. Snowfall occurs mostly in light to moderate amounts during the winter, totaling . Precipitation, at annually, is concentrated in the warmer months. Extreme temperatures have ranged from on January 12, 1912, and February 8, 1895, to on July 6 and 15, 1936, although a reading occurred as recently as January 15, 2009. The National Oceanographic and Atmospheric Administration maintains a National Weather Service office in Aberdeen. Their area of responsibility includes northern and eastern South Dakota and two counties in west-central Minnesota. Aberdeen is the county seat of Brown County. The original county seat was, however, Columbia. During the days of the railroad construction, plans were laid to bring the railroad through Columbia, then the county seat. When word of this spread, land in and around Columbia soared in price due to speculation. When the time came for the railroads to purchase land, the increase in land prices led them to change their decision and instead to route the rail lines through Aberdeen. However, once Aberdeen became a town in 1881, there was a long-running controversy concerning which town would be the county seat, which continued until 1890, when it was declared by the newly formed South Dakota state constitution in 1889 that a majority vote could move the county seat if the county seat in question had originally been established by less than a majority vote. The result of the vote declared that Aberdeen would be the county seat once and for all, so all of the records were once again transferred to Aberdeen's courthouse; during the battle for county seat, the records had been moved from Columbia's courthouse to Aberdeen's courthouse (which was built from 1886 to 1887), and back again to Columbia's in what seemed to be a never-ending cycle of the transferring of records. This was typically done in the form of nighttime raids from the two towns. As of the census of 2010, there were 26,091 people, 11,418 households, and 6,354 families residing in the city. The population density was . There were 12,158 housing units at an average density of . The racial makeup of the city was 91.8% White, 0.7% African American, 3.6% Native American, 1.3% Asian, 0.2% Pacific Islander, 0.5% from other races, and 2.0% from two or more races. Hispanic or Latino of any race were 1.6% of the population. There were 11,418 households of which 27.1% had children under the age of 18 living with them, 42.1% were married couples living together, 9.5% had a female householder with no husband present, 4.0% had a male householder with no wife present, and 44.4% were non-families. 36.9% of all households were made up of individuals and 13.1% had someone living alone who was 65 years of age or older. The average household size was 2.18 and the average family size was 2.86. The median age in the city was 36.4 years. 22.2% of residents were under the age of 18; 12.8% were between the ages of 18 and 24; 24.1% were from 25 to 44; 24.4% were from 45 to 64; 16.4% were 65 years of age or older. The gender makeup of the city was 47.6% male and 52.4% female. As of the census of 2000, there were 24,658 people, 10,553 households and 6,184 families residing in the city. The population density was 1,902.1 per square mile (734.4/km²). There were 11,259 housing units at an average density of 868.5 per square mile (335.3/km²). The racial makeup of the city was 94.61% White, 0.37% Black or African American, 3.17% Native American, 0.54% Asian, 0.13% Pacific Islander, 0.19% from other races, and 0.99% from two or more races. 0.79% of the population were Hispanic or Latino of any race. 53.7% were of German, 15% Norwegian and 8.5% Irish ancestry. There were 10,553 households out of which 27.3% had children under the age of 18 living with them, 47.0% were married couples living together, 8.9% had a female householder with no husband present, and 41.4% were non-families. 34.9% of all households were made up of individuals and 13.6% had someone living alone who was 65 years of age or older. The average household size was 2.21 and the average family size was 2.86. Age spread: 21.8% under the age of 18, 14.1% from 18 to 24, 26.4% from 25 to 44, 20.4% from 45 to 64, and 17.2% who were 65 years of age or older. The median age was 36 years. For every 100 females, there were 89.2 males. For every 100 females age 18 and over, there were 85.3 males. As of 2000, the median income for a household in the city was $33,276, and the median income for a family was $43,882. Males had a median income of $30,355 versus $20,092 for females. The per capita income for the city was $17,923. About 7.6% of families and 10.5% of the population were below the poverty line, including 10.6% of those under age 18 and 10.1% of those age 65 or over. There are several Roman Catholic, Baptist, Lutheran, Methodist, Pentecostal, and the Church of Jesus Christ of Latter-day Saints churches in the area, as well as one synagogue. Super 8 Motels was founded in 1972 by Dennis Brown and Ron Rivett as a motel referral system, which was replaced with a franchise operation in 1973. The first Super 8, with 60 rooms, was opened in 1974 in Aberdeen and still operates today as the Super 8 Aberdeen East. The Aberdeen area has several cultural organizations. The Aberdeen Area Arts Council publishes a small monthly newspaper, "ARTiFACTS", with information on area events. The Aberdeen Community Theatre was created in 1979 and performs at the Capitol Theatre in downtown Aberdeen. The Capitol Theatre opened in 1927 and donated to the Aberdeen Community Theatre in 1991; since then more than $963,000 has been spent on renovating and preserving the historical aspect of the Capitol Theatre. Today, the Aberdeen Community Theatre performs five mainstage productions and three youth productions per year. The South Dakota Film Festival established in 2007 is held annually in the fall. The festival has been host to Kevin Costner, Graham Greene, Adam Greenberg, CSA and many more stars of film and television. The festival's first feature film screened was Into The Wild, shot partially in SD. The festival is held at the historic Capitol Theatre. The Northern State University Theater Department puts on plays during the school year. The ArtWorks Cooperative is a partnership of artists who work to market their artwork in a gallery setting. The ArtWorks Cooperative sells artists' work and provides an environment that will benefit the artist in terms of artist-to-artist communication, and public interest. There are four galleries in Aberdeen: Presentation College's Wein Gallery, Northern State University's Lincoln Gallery, the Aberdeen Recreation & Cultural Center (ARCC) Gallery and the ArtWorks Cooperative Gallery located in The Aberdeen Mall. The Village Bowl in Aberdeen is a modern bowling center with multiple lanes. Located at 1314 8th Ave. NW. Aberdeen has been home to three minor league baseball teams since 1920. The Aberdeen Boosters, a class D league team, played in 1920, the Aberdeen Grays, also a class D team, played from 1921 to 1923. The class C Aberdeen Pheasants from 1946 to 1971, and 1995 to 1997. The Pheasants were the affiliate of the former St. Louis Browns (current Baltimore Orioles) and the Baltimore Orioles. Aberdeen was a stop to the majors for such notable players as Don Larsen (perfect game in the World Series), Lou Piniella (AL rookie of the year with Kansas City Royals in 1969), and Jim Palmer, Baseball Hall of Fame pitcher for the Baltimore Orioles. Aberdeen is presently home to 19 public tennis courts throughout the city – Melgaard Park (4), Northern State University (6), and Holgate Middle School (8). Aberdeen has three golf courses: Lee Park Municipal Golf Course, Moccasin Creek Country Club and Rolling Hills Country Club. Lee Park and Moccasin Creek are both 18-hole courses. Rolling Hills is a combined nine-hole course and housing development which opened in 2005. Aberdeen has multiple outdoor skating rinks and hockey rinks open to the public during winter months. Aberdeen is also home to the NAHL team, Aberdeen Wings. Aberdeen has a skate park located between East Melgaard Road and 17th Ave SE at Melgaard Park. The equipment installed includes a quarter pipe, penalty box with half pyramid, bank ramp, spine, kinked rail and a ground rail. Aberdeen has two disc golf courses, Melgaard Park, and the Richmond Lake Disc Golf Course. Aberdeen has an All-women's Roller Derby league "A-Town Roller Girlz" established 2011, also bringing Junior Roller Derby to the area. Men's league to follow in the midst of interest in the dynamic of the sport. Completed in the summer of 2007, this complex includes a zero entry pool, competition lap pool, lazy river, numerous water slides, play sand area, and a concession area. Wylie Park Recreation Area features go-kart racing, sand volleyball courts, access to Wylie Lake, camping area, picnic areas, and is connected to Storybook Land. Wylie Lake is a small man-made lake, open in the summer months for swimming, lying on the beach, and paddle boating. Storybook Land is a park with attractions from several different children's storybooks. The park contains a castle, as well as a train that takes visitors through the park. There are two barns which contain petting zoos. Humpty Dumpty's Great Fall Roller Coaster was added to the park, summer 2015. Newly added is the Land of Oz, that features characters and attractions from L. Frank Baum's "The Wonderful Wizard of Oz". Baum was a resident of Aberdeen in the 1880s, but left after the failure of the newspaper "The Aberdeen Saturday Pioneer", where he wrote a column, "Our Landlady". The Kuhnert Arboretum provides many new learning experiences for the residents of the Aberdeen area, including school-aged children. The Arboretum offers environmental education, a children's area, rose garden collection, recreational trails and much more. The Richmond Lake Recreation Area is used by all types of outdoors enthusiasts. Three separate areas in this park cater to the needs of campers, swimmers, naturalists, boaters and anglers. Campers stay in the South Unit, while the Forest Drive Unit is a great place for wildlife viewing. The Boat Ramp Unit provides access to the more than lake. Richmond Lake Recreation Area's small campground offers a quiet camping experience. The park also features a wheelchair accessible camping cabin. The park's extensive trail system features over of trails, including both accessible and interpretive trails. Hikers, bikers, and horseback riders can observe the abundance of prairie plants and wildlife of the area up-close. The park has multiple private and public boat ramps as well as an accessible fishing dock. Richmond Lake has a population of walleye, northern pike, bass, perch, crappie, bluegill, catfish, and bullheads within its waters. An entrance fee is required to gain access to the water and park itself. Aberdeen is the center of government for Brown County. City government is overseen by a mayor/city manager and eight council members. The city council is composed of Mayor Travis Schaunaman, City Manager Lynn Lander, and council members Dave Lunzman, Josh Rife, Mark Remily, Rob Ronayne, Alan Johnson, David Bunsness, Clint Rux and Dennis 'Mike' Olson. Each council member serves a five-year term. County government is overseen by five commissioners. Each county commissioner serves a five-year term. The county commissioners include Duane Sutton, Tom Fischbach, Mike Wiese, Rachel Kippley, and Doug Fjeldheim. Aberdeen is home to Brown County offices including clerk-magistrate, county auditor, landfill office, register of deeds, county treasurer, coroner, emergency management, highway superintendent, public welfare, state's attorney, and a few others. The state senators from Brown County include Brock Greenfield and Al Novstrup, and the state representatives included Lana Greefield, Kaleb Wies, Carl Perry and Drew Dennert. They are all in office until December 2020 In 2008, Governor Mike Rounds named Aberdeen as the South Dakota Community of the Year. Aberdeen Public Schools are part of the Aberdeen School District. The school district has six elementary schools, two middle schools, and one high school. The elementary schools are C.C. Lee Elementary School, Lincoln Elementary School, May Overby Elementary School, O.M. Tiffany Elementary School, Simmons Elementary School and Mike Miller Elementary School. The two middle schools are Holgate Middle School, which serves the north side of Aberdeen, and Simmons Middle School, which serves the south side of the city. Students in the district attend Central High School. The Hub Area Technical School is located in the district. Aberdeen also has an alternative middle and high school. The Aberdeen School District's enrollment for the year 2011–2012 was approximately 3,945 students, and the average class size was in the low to mid-twenties. Due to a projected increase in enrollment and the modernization of facilities, Simmons Middle School was completely remodeled with the demolition of the original 1929 building and the addition of a new classroom and cafeteria building which was completed in August 2008. The public school in Aberdeen is AA under the SDHSAA. Aberdeen has several parochial schools, including the Catholic-affiliated Roncalli High School, the nondenominational Aberdeen Christian School, and Trinity Lutheran School of the WELS. The South Dakota School for the Blind and Visually Impaired is a state special school under the direction of the South Dakota Board of Regents. Northern State University (NSU) is a public university that was founded in 1901 and today occupies a campus. 2,528 students, ranging from first-year to graduate students, attended NSU for the 2006–2007 school year. The student to teacher ratio is 19:1. NSU was originally called the Institute of South Dakota before changing its name to Northern Normal and Industrial School in 1901. It changed its name again in 1939 when it became the Northern State Teachers College, and again in 1964, becoming Northern State College before finalizing at Northern State University in 1989. NSU offers thirty-eight majors and forty-two minors as well as other degrees, and also has nine graduate degree areas for students wishing to further their education after achieving their first degree. The mascot of NSU is the wolf named Thunder. Presentation College is a Catholic college on a campus, and was founded in 1951. Enrollment in fall 2014 was reported to be 735. PC offers 26 programs between the main Aberdeen campus and the other campuses located throughout the state. Most of the degrees offered are in the health-care field. The student to teacher ratio is 12:1. Presentation's mascot is the Saint, giving it the nickname the Presentation College Saints. "The American News" was founded as a weekly newspaper in 1885, by C.W. Starling and Paul Ware. It is now a daily newspaper. The Aberdeen Regional Airport is currently served by Delta Connection. It offers flights to Minneapolis-St. Paul International Airport using the Bombardier CRJ200 aircraft. There are two major US highways that serve Aberdeen. One is US Highway 281 that runs north–south from the North Dakota border to the border with Nebraska. The second highway is US Highway 12 that runs east–west across northern South Dakota from the Minnesota border before curving northwest into the southwestern corner of North Dakota. US Highway 12 is the major thoroughfare in Aberdeen. US Highway 12 is signed in the city of Aberdeen as 6th Avenue South. US Highway 281 was recently realigned onto a new bypass that was constructed around the western area of the city. Taxi Aberdeen Taxi service provides general taxi service in Aberdeen. Aberdeen Shuttle provides shuttle service to and from the airport along with general taxi services. Bus Jefferson Lines is a bus service from Aberdeen that connects to Sioux Falls, South Dakota, Fargo, North Dakota, and Minneapolis, Minnesota. Car rental There are five car rental services in Aberdeen: Hertz, Avis, Dollar-Thrifty, Toyota Rent-a-Car and Nissan Rental Car. Hertz and Avis Car rental are located in the airport terminal. Dollar-Thrifty is located in Aberdeen Flying Service. Toyota Rent-a-Car and Nissan Rental Car are located at Harr Motors across from the airport. Train The BNSF Railway conveys freight and grain through Aberdeen. Until 1969 the Milwaukee Road ran trains between Aberdeen and Minneapolis. Earlier, until 1961, trains from Chicago to Seattle ran through the railroad's Aberdeen Station. Aberdeen is currently home to two hospitals, Avera St. Luke's Hospital, and Sanford Aberdeen Medical Center. There are several nursing homes in the area, including Avera Mother Joseph Manor, Manor Care, Bethesda Home of Aberdeen, Aberdeen Health and Rehab, Angelhaus and Gellhaus Carehaus.
https://en.wikipedia.org/wiki?curid=2709
Aberdour Aberdour (; , ) is a scenic and historic village on the south coast of Fife, Scotland. It is on the north shore of the Firth of Forth, looking south to the island of Inchcolm and its Abbey, and to Leith and Edinburgh beyond. According to the 2011 census, the village has a population of 1,633. The village's winding High Street lies a little inland from the coast. Narrow lanes run off it, providing access to the more hidden parts of the village and the shoreline itself. The village nestles between the bigger coastal towns of Burntisland to the east and Dalgety Bay to the west. The parish of Aberdour takes its name from this village, and had a population of 1,972 at the 2011 Census. The origins of the village lie with its harbour, where the Dour Burn enters the River Forth. The place-name itself is Pictish, implying an origin in the Dark Ages: "aber" 'confluence'. The -dour element, referring to the Burn, means simply 'water' (archaic "dobur"), and is unconnected to the Scots/English 'dour'. For much of its history Aberdour was two villages, Wester Aberdour and Easter Aberdour, on either side of the Dour Burn; however this distinction was blurred by the 19th century arrival of the railway. In the 18th century Aberdour's harbour was improved by the addition of a stone pier to help handle the coal traffic from nearby collieries. However, in the 1850s the traffic changed dramatically, and Aberdour Harbour became a popular destination for pleasure steamers from Leith. This in turn led to the building of a deeper water pier a little around the bay at Hawkcraig, and to the development of hotels and many of the other services still on view today in the village. The railway came to Aberdour in 1890, with the building of the line east from the newly opened Forth Bridge. The station has won many "best kept station" awards. The half an hour journey to the centre of Edinburgh helped build on the existing popularity of the village, though it put the steamers out of business. The main result was a growth in the building of large and attractive houses, especially down the hill from Wester Aberdour to the West Sands. Ticket inspectors on the train line through Aberdour were known for their sing song refrain: ""Half an hour, Half an hour, Half an hour to Aberdour — tickets please."" Virtually between the two former settlements, though actually part of Easter Aberdour, lies Aberdour Castle. This started life as a modest hall house on a site overlooking the Dour Burn in the 13th century. The oldest part of the present semi-ruin constitutes one of the earliest surviving stone castles in mainland Scotland. Over the next four hundred years the Castle was successively developed according to contemporary architectural ideas. Notable are the parts, still largely roofed, built by the Earls of Morton, with refined Renaissance detail, in the second half of the 16th century. A fire in the late 17th century was followed by some repairs, but in 1725 the family purchased 17th century Aberdour House, on the west side of the burn and in Wester Aberdour, and the medieval Castle was allowed to fall into relative decay. Aberdour Castle is now in the care of Historic Environment Scotland and open to the public (entrance charge). After a period of dereliction Aberdour House was developed for residential use in the early 1990s. Neighbouring St Fillan's Church is one of the best-preserved medieval parish churches in Scotland, dating largely to the 12th century. A south arcade was added to the nave in the early 16th century (open in summer). The A921, the main road along the south coast of Fife, leads down the High Street of Wester Aberdour, before kinking sharply left to cross the railway line, then right again to progress through Easter Aberdour's Main Street. Wester Aberdour has the more 'olde worlde' feel of the two, with the narrower through road more closely hemmed in by shops and hotels. A number of vernacular buildings of the 17th-early 19th centuries add to the historic scene. Close to the railway bridge, three lanes continue eastwards, presumably once the route of the original High Street before the arrival of the railway. One now leads to Aberdour railway station, a beautifully kept and cared for example of a traditional station, in keeping with its role of transporting at least a quarter of the village's working population to their work each day. A second lane leads alongside the railway line to Aberdour Castle, while a third leads to the restored Aberdour House. A little further west, a narrow road closely lined with high walls, Shore Road, leads down to the West Sands and the Harbour. For many this area is the highlight of any visit to Aberdour; parking at the foot of Shore Road is usually at a premium. Another road leads coastwards from Easter Aberdour. Hawkcraig Road leads past St Fillan's Church and through Silversand Park, home to Aberdour Shinty Club, en route to the much better parking area on Hawkcraig. This was formerly a sandstone quarry and then used as the council refuse tip before becoming a carpark, part of the overgrown and rocky bluff separating Aberdour's two bays. From here is it a short walk to the Silver Sands, Aberdour's busiest and most popular beach. On the west side of Hawkcraig Point there is a short concrete jetty that was used as part of the development of radio controlled torpedoes during World War One. The foundations of the Radio Hut can still be seen in the lea of the hill. The Aberdour obelisk was built by Lord Morton on his departure from the village to relocate to a large home in Edinburgh, it was built so he could see his former hometown from his new house when he looked through binoculars - it stands in a cowfield between the castle and the beach. Aberdour hosts an annual festival, which runs from late July to early August and features musical events, shows, sporting events and children's events. Aberdour was a 2005 finalist in the prestigious "Beautiful Scotland in Bloom" awards. It was nominated for ""Best Coastal Resort"" in Scotland along with St Andrews in Fife, North Berwick in East Lothian, and Rothesay in Argyll and Bute. In 2014 Aberdour was voted Best Coastal Village in Fife and Best Small Coastal Village in Scotland. It also received a Gold Award in Beautiful Fife and Beautiful Scotland. In 2018 Aberdour was voted the second-best place to live in Scotland by The Sunday Times list of Best Places to Live Aberdour is home to Fife's only senior shinty club. Aberdour Shinty Club field teams in both the men's and women's senior national leagues. Aberdour has two beaches - the Silver Sands, and the Black Sands. The Silver Sands are located on the East side of the village, and has previously held a "Blue flag" beach award, which denotes an exemplary standard of cleanliness, facilities, safety, environmental education and management. New facilities are currently under construction by Fife Council, which will much improve the beach throughout the year. The Black Sands (also known as the West Beach), as the contrasting name would suggest, have a rockier and darker sand, and are also popular with visitors exploring the rock caves and fascinating sea life. During the summer months (April–September), dogs are banned from the Silver Sands but they are allowed all year round at the West Beach. The two beaches are linked by part of the Fife Coastal Path which also takes you past the harbour and the Hawkcraig - a popular rock climbing location. Silver Sands is becoming more popular with open water swimmers, who swim daily in the sea, both as a leisure pursuit, and as training for open water competition. The bay provides safety from the currents, although only the adventurous swim round to the harbour. Several scenes of Richard Jobson's 2003 movie "16 Years of Alcohol" were filmed at the Black Sands in Aberdour. The island of Inchcolm, or Island (Gaelic "innis") of Columba, a quarter of a mile from the shore, forms part of the parish of Aberdour. Its name implies associations dating back to the time of Columba and, although undocumented before the 12th century, it may have served the monks of the Columban family as an 'Iona of the east' from early times. During the First and Second World Wars, Inchcolm was occupied by the army as part of the defences of the Firth of Forth. There are extensive remains of gun emplacements, barracks, etc. from these periods. The island is notable for its wildlife, especially seabirds and seals. These draw many visitors in summer, along with the remains of the historic Abbey, and is a popular setting for weddings. Notable past and present residents of the town include:
https://en.wikipedia.org/wiki?curid=2712
Aberfoyle, Stirling Aberfoyle () is a village in the historic county and registration county of Perthshire and the council area of Stirling, Scotland. The settlement lies northwest of Glasgow. The parish of Aberfoyle takes its name from this village, and had a population of 1,065 at the 2011 Census. The town is situated on the River Forth at the foot of Craigmore ( high). Since 1885, when the Duke of Montrose constructed a road over the eastern shoulder of Craigmore to join the older road at the entrance of the Trossachs pass, Aberfoyle has become the alternative route to the Trossachs and Loch Katrine; this road, known as the Duke's Road or Duke's Pass, was opened to the public in 1931 when the Forestry Commission acquired the land. Loch Ard, about west of Aberfoyle, lies above the sea. It is long (including the narrows at the east end) and 1 mile (1.6 km) broad. Towards the west end is Eilean Gorm ("the green isle"), and near the north-western shore are the falls of Ledard. The loch's northern shores are dominated by the mountain ridge of Beinn an Fhogharaidh (). Two miles northwest of Loch Ard is Loch Chon, at above the sea, long and about broad. It drains by the Avon Dhu to Loch Ard, which is drained in turn by the Forth. Aberfoyle supposedly originates from the Brittonic Celtic, "aber poll" or "aber phuill" (Scottish Gaelic, ), meaning (place at the) mouth of the Phuill Burn (the Pow Burn enters the River Forth at Aberfoyle). Historically, alternative spellings such as "Abirfull", "Aberfule", "Aberfoill" and "Aberfoil" have been recorded before the current spelling became accepted by the 20th century. The slate quarries on Craigmore which operated from the 1820s to the 1950s are now defunct; at its peak this was a major industry. Other industries included an ironworks, established in the 1720s, as well as wool spinning and a lint mill. From 1882 the village was served by Aberfoyle railway station, the terminus of the Strathendrick and Aberfoyle Railway which connected to Glasgow via Dumbarton or Kirkintilloch. The station closed to passenger traffic in 1951, and the remaining freight services ceased in 1959. The above industries have since died out, and Aberfoyle is supported mainly by the forestry, industry and tourism. Visitors were first attracted to Aberfoyle and the surrounding area after the publication of "The Lady of the Lake" by Sir Walter Scott in 1810. The poem described the beauty of Loch Katrine. Aberfoyle describes itself as "The Gateway to the Trossachs", and is well situated for visitors to access attractions such as Loch Lomond and Inchmahome Priory at the Lake of Menteith. A tourist information office run by VisitScotland sits in the centre of town, offering free information, selling souvenirs and acting as a booking office for many of the local B&B's and hotels. Aberfoyle Golf Club was built in 1860 and is located just south of town near the Rob Roy restaurant. Aberfoyle is also part of the Loch Lomond and The Trossachs National Park. Aberfoyle is also home to the largest Go Ape adventure course in the UK, featuring the longest death slide, or 'zip-line', in the UK. Aberfoyle has connections to many historical figures such as Rob Roy and Mary, Queen of Scots. Robert Roy MacGregor was born at the head of nearby Loch Katrine, and his well-known cattle stealing exploits took him all around the area surrounding Aberfoyle. It is recorded, for example, that in 1691, the MacGregors raided every barn in the village of Kippen and stole all the villagers' livestock. There currently stands a tree in the village that MacGregor was reputed to have climbed and hid in to escape the clutches of the law. Also, Mary, Queen of Scots, visited nearby Inchmahome Priory often as a child, and during her short reign. She also used the priory during her short reign, particularly in 1547, where she felt safe from the English Army. However, the most local historical figure is the Reverend Robert Kirk, born in 1644. It was the Rev. Kirk who provided the first translation into Scottish Gaelic of the Book of Psalms, however, he is better remembered for the publication of his book ""The Secret Commonwealth of Elves, Fauns, and Fairies"" in 1691. Kirk had long been researching fairies, and the book collected several personal accounts and stories of folk who claimed to have encountered them. It was after this, while Kirk was minister of Aberfoyle parish, that he died in unusual circumstances. Kirk had long believed that the local Doon Hill was the gateway to the "Secret Commonwealth", or the land of the Fairies. It was a place that Kirk visited often, taking daily walks there from his manse. The story goes that the Fairies of Doon Hill were angry with the Rev. Kirk for going into the domain of the Unseelie court, where he had been warned not to go, and decided to imprison him in Doon Hill — for one night in May 1692, the Rev. Kirk went out for a walk to the hill, in his nightshirt. Some accounts claim that he simply vanished, however he suddenly collapsed. He was found and brought home, but died soon afterwards. He was buried in his own kirkyard, although local legends claim that the fairies took his body away, and the coffin contains only stones. The huge pine tree that still stands at the top of Doon Hill is said to contain Kirk's imprisoned spirit. Kirk's cousin, Graham of Duchray, was then to claim that the spectre of Kirk had visited him in the night, and told him that he had been carried off by the Fairies. Having left his widow expecting a child, the spectre of Kirk told Graham that he would appear at the baptism, whereupon Graham was to throw an iron knife at the apparition, thus freeing Kirk from the Fairies' clutches. However, when Kirk's spectre appeared, Graham was apparently too shocked by the vision to throw the knife, and Kirk's ghost faded away forever. Today, visitors to Doon Hill write their wishes on pieces of white silk, or other white cloth, and tie them to the branches of the trees for the Fairies to grant. Unfortunately some people tie plastic confectionery wrappers instead, which slightly spoils the magic of the location and may harm the ecology of the forest. It is also said that if one runs around the great 'Minister's Pine' tree at the summit seven times, then the Fairies will appear. Some people have tried this and afterwards claim to have seen apparitions. Others merely get a bit dizzy and fall over.
https://en.wikipedia.org/wiki?curid=2714
Abergavenny Abergavenny (; , archaically "Abergafenni" meaning "mouth of the River Gavenny") is a market town and community in Monmouthshire, Wales. Abergavenny is promoted as a "Gateway to Wales". It is located on the A40 trunk road and the A465 Heads of the Valleys road and is approximately from the border with England. Originally the site of a Roman fort, Gobannium, it became a medieval walled town within the Welsh Marches. The town contains the remains of a medieval stone castle built soon after the Norman conquest of Wales. Abergavenny is situated at the confluence of the River Usk and a tributary stream, the Gavenny. It is almost entirely surrounded by mountains and hills: the Blorenge (), the Sugar Loaf,(), Ysgyryd Fawr (Great Skirrid), Ysgyryd Fach (Little Skirrid), Deri, Rholben and Mynydd Llanwenarth, known locally as "Llanwenarth Breast". Abergavenny provides access to the nearby Black Mountains and the Brecon Beacons National Park. The Marches Way and Beacons Way pass through Abergavenny whilst the Offa's Dyke Path passes through Pandy five miles to the north and the Usk Valley Walk passes through nearby Llanfoist. In the UK 2011 census, the six relevant wards (Lansdown, Grofield, Castle, Croesonen, Cantref and Priory) collectively listed Abergavenny's population as 12,515. The town hosted the 2016 National Eisteddfod of Wales. The town derives its name from a Brythonic word "Gobannia" meaning "river of the blacksmiths", and relates to the town's pre-Roman importance in iron smelting. The name is related to the modern Welsh word "gof" (blacksmith), and so is also associated with the Welsh smith Gofannon from folklore. The river later became, in Welsh, "Gafenni", and the town's name became "Abergafenni", meaning "mouth of (Welsh: "Aber") the Gavenny ("Gafenni")". In Welsh, the shortened form "Y Fenni" may have come into use after about the 15th century, and is now used as the Welsh name. Abergavenny, the English spelling, is in general use. The town originally developed on the high ground to the north of the floodplain of the River Usk and to the west of the valley of the much smaller Gavenny River though has since extended to the east of the latter. It has merged with the originally separate settlement of Mardy to the north but remains separate from that of Llanfoist to the south due to the presence of the river and its floodplain; nevertheless Llanfoist is in many ways a suburb of the town. The ground rises gradually in the north of the town before steepening to form the Deri and Rholben spurs of Sugar Loaf. The A4143 crossing of the Usk by means of the historic Usk Bridge is sited at the narrowest point of the floodplain, a site also chosen for the former crossing of a tramroad and the later mainline railway. The high ground at either side is formed by a legacy of the last ice age, the recessional Llanfoist moraine which underlies both the village which gives it its name, the town centre and the Nevill Hall area. The older parts of the town north of its centre are built upon a relatively flat-lying alluvial fan extending west from the area of St Mary's Priory to Cantref and of similar age to the moraine. Gobannium was a Roman fort guarding the road along the valley of the River Usk, which linked the legionary fortress of Burrium (Usk) and later Isca Augusta or Isca Silurum (Caerleon) in the south with Y Gaer, Brecon and Mid Wales. It was also built to keep the peace among the local British Iron Age tribe, the Silures. Remains of the walls of this fort were discovered west of the castle when excavating the foundations for a new post office and telephone exchange building in the late 1960s. Abergavenny grew as a town in early Norman times under the protection of the Baron Bergavenny (or Abergavenny). The first Baron was Hamelin de Balun, from Ballon, a small town with a castle in Maine-Anjou near Le Mans. Today it is in the Sarthe département of France. He founded the Benedictine priory, now the Priory Church of St Mary, in the late 11th century. The Priory belonged originally to the Benedictine foundation of St. Vincent Abbaye at Le Mans. It was subsequently endowed by William de Braose, with a tithe of the profits of the castle and town. The church contains some unique alabaster effigies, church monuments and unique medieval wood carving, such as the Tree of Jesse. Owing to its geographical location, the town was frequently embroiled in the border warfare and power play of the 12th and 13th centuries in the Welsh Marches. In 1175, Abergavenny Castle was the site of a massacre of Seisyll ap Dyfnwal and his associates by William de Braose, 4th Lord of Bramber. Reference to a market at Abergavenny is found in a charter granted to the Prior by William de Braose. Owain Glyndŵr attacked Abergavenny in 1404. According to popular legend, his raiders gained access to the walled town with the aid of a local woman who sympathised with the rebellion, letting a small party in via the Market Street gate at midnight. They were able to open the gate and allow a much larger party who set fire to the town and plundered its churches and homes leaving Abergavenny Castle intact. Market Street has been referred to as "Traitors' Lane" thereafter. In 1404 Abergavenny was declared its own nation by Ieuan ab Owain Glyndŵr, illegitimate son of Owain Glyndŵr. The arrangement lasted approximately two weeks. At the Dissolution of the Monasteries in 1541, the priory's endowment went towards the foundation of a free grammar school, King Henry VIII Grammar School, the site itself passing to the Gunter family. During the Civil War, prior to the siege of Raglan Castle in 1645, King Charles I visited Abergavenny and presided in person over the trial of Sir Trefor Williams, 1st Baronet of Llangibby, a Royalist who changed sides, and other Parliamentarians. In 1639, Abergavenny received a charter of incorporation under the title of bailiff and burgesses. A charter with extended privileges was drafted in 1657, but appears never to have been enrolled or to have come into effect. Owing to the refusal of the chief officers of the corporation to take the oath of allegiance to William III in 1688, the charter was annulled, and the town subsequently declined in prosperity. Chapter 28 of the 1535 Act of Henry VIII, which provided that Monmouth, as county town, should return one burgess to Parliament, further stated that other ancient Monmouthshire boroughs were to contribute towards the payment of the member. In consequence of this clause Abergavenny on various occasions shared in the election, the last instance being in 1685. The right to hold two weekly markets and three yearly fairs, beginning in the 13th century, was held ever since as confirmed in 1657. Abergavenny was celebrated for the production of Welsh flannel, and also for the manufacture, whilst the fashion prevailed, of goats' hair periwigs. Abergavenny railway station, situated southeast of the town centre, opened on 2 January 1854 as part of the Newport, Abergavenny and Hereford Railway. The London North Western Railway sponsored the construction of the railway linking Newport station to Hereford station. The line was taken over by the West Midland Railway in 1860 before becoming part of the Great Western Railway in 1863. The station is on the Welsh Marches Line and is mostly served by Transport for Wales services. A railway line also ran up the valley towards Brynmawr and to Merthyr Tydfil. This line was closed following the Beeching cuts and the line to Clydach Gorge is now a cycle track and footpath. The Baker Street drill hall was completed in 1896. Adolf Hitler's deputy Rudolf Hess was kept under escort at Maindiff Court Hospital during the Second World War, after his flight to Britain. The title of Baron Abergavenny, in the Nevill family, dates from the 15th century with Edward Nevill, 3rd Baron Bergavenny, who was the youngest son of Ralph de Neville, 1st Earl of Westmorland by his second wife Joan Beaufort, daughter of John of Gaunt, first Duke of Lancaster. He married the heiress of Richard de Beauchamp, 1st Earl of Worcester, whose father had inherited the castle and estate of Abergavenny, and was summoned in 1392 to parliament as Lord Bergavenny. Edward Nevill was summoned to parliament with this title in 1450. His direct male descendants ended in 1587 in Henry Nevill, 6th Baron Bergavenny, but a cousin, Edward Nevill, 8th Baron Bergavenny, was confirmed in the barony in 1604. From him it has descended continuously, through fifteen individuals, the title being increased to an earldom in 1784; and in 1876 William Nevill 5th Earl, an indefatigable and powerful supporter of the Tory Party, was created 1st Marquess of Abergavenny. Coldbrook Park was a country house in an estate some southeast of the town. The house was originally built in the 14th century and belonged to the Herbert family for many generations until purchased by John Hanbury for his son, the diplomat Sir Charles Hanbury Williams. Sir Charles reconstructed the house in 1746 with the addition of a nine-bay two-storey Georgian façade with a Doric portico. It subsequently passed down in the Hanbury Williams family until it was demolished in 1954. Held during the first week of August every year, the National Eisteddfod is a celebration of the culture and language in Wales. The festival travels from place to place, alternating between north and south Wales, attracting around 150,000 visitors and over 250 tradestands and stalls. In 2016 it was held in Abergavenny for the first time since 1913. The Chair and Crown for 2016 were presented to the festival's Executive Committee at a ceremony held in Monmouth on 14 June 2016. The Abergavenny Food Festival is held in the second week of September each year. The Steam, Veteran and Vintage Rally takes place in May every year. The event expands year on year with the 2016 rally including a rock choir, shire horses, motorcycling stunts, vintage cars and steam engines. The Country and Western Music Festival is attended by enthusiasts of country music. It marked its third year in 2016 and was attended by acts including Ben Thompson, LA Country and many more. The event was last held in 2017. The Abergavenny Writing Festival began in April 2016 and is a celebration of writing and the written word. The Abergavenny Arts Festival, first held in 2018, celebrates arts in their broadest sense and showcases amateur and professional artists from the vibrant local arts scene together with some from further afield. In recent decades the number of Welsh speakers in the town has increased dramatically. The 2001 census recorded 10% of the local population spoke the language, a five-fold increase over ten years from the figure of 2% recorded in 1991. The town has one of the two Welsh-medium primary schools in Monmouthshire, Ysgol Gymraeg y Fenni, which was founded in the early 1990s. It is also home to the Abergavenny Welsh society, Cymreigyddion y Fenni, and the local Abergavenny Eisteddfod. Abergavenny was the home of Abergavenny Thursdays F.C., which was formed in 1927, but wound up in 2013. However, football is now represented in the town by the new Abergavenny Town team which was formerly called Govilon FC. Town play at the Pen-y-pound Stadium and are members of the Welsh League Division 1 for season 2019-20 Abergavenny Cricket Club play at Pen-y-Pound, Avenue Road, and Glamorgan CCC also play some of their games here. Abergavenny Cricket Club is one of the oldest in the country. It was founded in 1834, and celebrated the 175th anniversary of its foundation in 2009. Abergavenny Tennis Club also play at Pen-y-Pound and plays in the South Wales Doubles League and Aegon Team Tennis. The club engages the services of a head tennis professional to run a coaching programme for the town and was crowned Tennis Wales' Club of the Year in 2010. Abergavenny is also the home of Abergavenny RFC, a rugby union club founded in 1875 who play at Bailey Park. In the 2018-19 season they play in the Welsh Rugby Union Division Three East A league. Abergavenny Hockey Club, formed in 1897, currently compete in the Davis Woods hockey league and play at the Old Hereford Road ground. Abergavenny hosted the British National Cycling Championships in 2007, 2009 and 2014, as part of the town's Festival of Cycling. A cattle market was held in Abergavenny from 1863 to December 2013. During the period 1825–1863 a sheep market was held at a site in Castle Street, to stop the sale of sheep on the streets of the town. At the time of its closure the market was leased and operated by Abergavenny Market Auctioneers Ltd, who held regular livestock auctions on the site. Market days were held on Tuesdays for the auction sale of finished sheep, cull ewe/store and fodder (hay and straw), and some Fridays for the auction sale of cattle. Following the closure of Newport's cattle market in 2009 for redevelopment, Newport’s sales were held at Abergavenny every Wednesday. In 2011, doubts about the future of Abergavenny Cattle Market were raised following the granting of planning permission by Monmouthshire County Council for its demolition and replacement with a supermarket, car park, and library. In January 2012, the Welsh Government announced the repeal the Abergavenny Improvement Acts of 1854 to 1871 which obliged the holding of a livestock market within the boundaries of Abergavenny town; that repeal being effective from 26 March 2012. Monmouthshire County Council, which requested that the Abergavenny Improvement Acts be repealed, supported plans for a new cattle market to be established about from Abergavenny in countryside at Bryngwyn, some from Raglan. There was extensive local opposition to this site, which is situated on a notoriously dangerous B road. The new Monmouthshire Livestock Centre, a 27-acre site at Bryngwyn, opened in November 2013. Various markets are held in the Market Hall, for example: Tuesdays, Fridays and Saturdays – retail market; Wednesdays – flea market; fourth Thursday of each month – farmers' market; third Sunday of each month – antique fair; second Saturday of each month – craft fair. Abergavenny Castle is located strategically just south of the town centre overlooking the River Usk. It was built in about 1067 by the Norman baron Hamelin de Ballon to guard against incursions by the Welsh from the hills to the north and west. All that remains is defensive ditches and the ruins of the stone keep, towers, and part of the curtain wall. It is a Grade I listed building. The Church in Wales church of the Holy Trinity is in the Diocese of Monmouth. Holy Trinity Church was consecrated by the Bishop of Llandaff on 6 November 1840. It was originally built as a chapel to serve the adjacent almshouses and the nearby school. It has been Grade II listed since January 1974. Other listed buildings in the town include the parish Priory Church of St Mary, a medieval and Victorian building that was originally the church of the Benedictine priory founded in Abergavenny before 1100; the sixteenth century tithe barn near St Mary's; the Victorian Church of the Holy Trinity; the Grade II* listed St John's Masonic Lodge; the Museum; the Public Library; the Town Hall; and the remains of Abergavenny town walls behind Neville Street. From 1851, the Monmouthshire lunatic asylum, later Pen-y-Fal Hospital, a psychiatric hospital, stood on the outskirts of Abergavenny. Between 1851 and 1950, over 3,000 patients died at the hospital. A memorial plaque for the deceased has now been placed at the site. After closure in the 1990s, its buildings and grounds were redeveloped as a luxury housing development comprising houses as well as apartments. Some psychiatric services are now administered from Maindiff Court Hospital on the outskirts of the town, close to the foot of the Skirrid mountain. The hospital is housed in historic buildings, and is known for Rudolf Hess who was incarcerated here during WW2. One of the eleven Victoria Cross medals won at Rorke's Drift was awarded to John Fielding from Abergavenny. He had enlisted under the false name of Williams. One was also awarded for the same action to Robert Jones, born at Clytha between Abergavenny and Raglan. Another Abergavenny-born soldier, Thomas Monaghan received his VC for defending his colonel during the Indian Rebellion. In 1908 following the formation of the Territorial Force the Abergavenny Cadet Corps was formed and affiliated with the 3rd Battalion, The Monmouthshire Regiment. In 1912 the regiment was affiliated with the new formed 1st Cadet Battalion, The Monmouthshire Regiment. "See also "
https://en.wikipedia.org/wiki?curid=2715
Abersychan Abersychan is a settlement and community north of Pontypool in Torfaen, Wales, and lies within the boundaries of the historic county of Monmouthshire and the preserved county of Gwent. Abersychan lies in the narrow northern section of the Afon Lwyd valley. The town includes two schools; Abersychan Comprehensive School and Victoria Primary School; together with various shops and other amenities including Abersychan Rugby Club. Abersychan was the birthplace of the politicians Roy Jenkins, Don Touhig and Paul Murphy (MP for Torfaen); and of the rugby footballers Wilfred Hodder, Candy Evans and Bryn Meredith. Like many of the 17th century isolated agricultural hamlets in the forested South Wales Valleys, Abersychan became a thriving industrial centre in the 19th and early 20th centuries, particularly for iron production. After the discovery of iron stone locally, the principal ironworks were built by the British Iron Company in 1825, served mainly by the LNWR's Brynmawr and Blaenavon Railway. The ironwork's main office building and quadrangle were designed by architect Decimus Burton, best known for his design of London Zoo. The works passed to the New British Iron Company in 1843 and to the Ebbw Vale Company in 1852, before closing in 1889. On 6 February 1890, an underground explosion at Llanerch Colliery killed 176. The site of the former ironworks today is a core site of , and a total land area of , includes a number of listed buildings: Various proposals have been made over the years to redevelop the site, currently under the ownership of HSBC, but none have so far passed the requirements of Torfaen county council. Abersychan constitutes a community and electoral ward of the county borough of Torfaen. The area was part of the ancient parish of Trevethin, in Monmouthshire. On 3 June 1864 Abersychan was constituted a local government district, governed by a local board. In 1894 Abersychan became an urban district and civil parish. The urban district was abolished in 1935, with most of its area passing to Pontypool urban district, and a small area going to Abercarn UD. In 1974 the area became part of the borough of Torfaen, in the new local government county of Gwent. The community of Abersychan was formed in 1985, but no community council has yet been formed. Abersychan and Cwmavon is now a ward for the Pontypool Community Council. In 1996 Torfaen became a unitary authority. The Abersychan community includes Abersychan, Cwmavon, Garndiffaith, Pentwyn, Talywain, Varteg, and Victoria Village. The nearest railway stations to Abersychan are Pontypool & New Inn (3 miles), Llanhilleth (3.5 miles) and Abergavenny (7 miles). Abersychan was served by the following (disused) stations: "Pentwyn", Torfaen is a small village located in the district of Abersychan. It contains a post office, several houses and a small play park. The village has a cricket team (Pentwyn CC) and is located right next to the old railway line. The cricket club celebrated its 100-year anniversary in 2006 with a successful tour to Cork, Ireland. The village has superb views over the River Severn and Newport to the south. "Victoria Village" is a small hamlet located in the district of Abersychan. It comprises a small village school and a number of houses. A small group of houses on Incline Road mark the beginning of the village and the village boundary is near Cwmavon. Victoria Primary School is also in this area, housed in large grounds. Many homes are built around the school's boundaries. Victoria Village primary school was opened in 1903 and closed by the council in 2018. The last head of the school was Miss Joy Dando. And caretaker Miss Debbie Williams.
https://en.wikipedia.org/wiki?curid=2716
Abertillery Abertillery (; , meaning mouth of the River Tyleri) is a town and a community of the Ebbw Fach valley in the historic county of Monmouthshire, Wales. Following local government reorganisation it became part of the Blaenau Gwent County Borough administrative area. The surrounding landscape borders the Brecon Beacons National Park and the Blaenavon World heritage Site. Formerly a major coal mining centre the Abertillery area was transformed in the 1990s using EU and other funding to return to a greener environment. Situated on the A467 the town is north of the M4 and south of the A465 "Heads of the Valleys" trunk road. It is about by road from Cardiff and from Bristol. According to the 2001 Census and information gathered by The Welsh Language Board (), 1,146 (9.9%) of Abertillery speaks Welsh. In the 2011 Census, this figure dropped to 7.2%, a 2.7 percentage point drop. Abertillery's traditional-style town centre mainly developed in the late 19th century and as such has some interesting Victorian architecture. Spread over 4 main streets the town in its heyday had two department stores and a covered Victorian arcade linking two of the main shopping areas. These were all included in a Blaenau Gwent Borough Council remodelling and modernisation project using European Union funding in a £13 million programme spread over a 5-year period ending in 2015. The project included a new multi-storey car park, a revamp of public areas and the town's Metropole Theatre. This RICS award-winning building provides state of the art production, exhibition, conference and meeting facilities as well as housing Abertillery museum. In March 2014 Prince Edward, the Earl of Wessex, officiated at the launch of Jubilee Square, a public facility in the town centre next to St Michael's Church. Major industry came to the area in 1843 when the locality's first deep coal mine was sunk at Tir Nicholas Farm, Cwmtillery. The town developed rapidly thereafter and played a major part in the South Wales coalfield. Its population rose steeply, being 10,846 in the 1891 census and 21,945 ten years later. The population peaked just short of 40,000 around the beginning of the 1930s. Eventually there were six deep coal mines, numerous small coal levels, a tin works, brick works, iron foundry and light engineering businesses in the area. Just one of the coal mines, Cwmtillery, produced over 32 million tons of coal in its lifetime and at its height employed 2760 men and boys. In 1960 an underground explosion at Six Bells Colliery resulted in the loss of life of 45 local miners. Fifty years later the archbishop of Canterbury Rowan Williams officiated at the launch of the "Guardian" mining memorial. This artistically acclaimed monument standing at 20m tall overlooks Parc Arael Griffin, the now reclaimed and landscaped former colliery site. The adjoining Ty Ebbw Fach visitor centre provides conference facilities, a restaurant and a "mining valley" experience room. Not long after the disaster the renowned artist L. S. Lowry visited the area and recorded the scene. The resultant landscape painting now hangs in the Museum of Wales in Cardiff. The coal mines remained the predominant economic emphasis until the general run down of the industry in the 1980s. Away from the town centre, the often steep sided nature of the landscape, imposes its own demands on development. Whilst this sounds limiting it has helped provide the almost amphitheatre nature of Abertillery Park, often described as one of the most attractive rugby grounds in world rugby. The street plan and housing stock flow uninterrupted from Cwmtillery in the north to Six Bells in the south, forming the town that is Abertillery. Prior to 1974 local government was provided by Abertillery Urban District Council (AUDC). Its area included the small neighbouring villages of Aberbeeg, Llanhilleth and Brynithel. Historical data relating to Abertillery occasionally refers to this AUDC area meaning that it can be difficult to compare like with like. For example, the 2014 population for the wider conurbation area is around 20,000 rather than the 11,000 often quoted for Abertillery itself. Whilst in the main the area has an older housing stock there are several developments of modern, often large homes, generally found on the outskirts of the town with views out over the surrounding area. These apart, terraced council tax band A and B properties predominate, meaning that average house prices are among the most affordable in the UK. There are very few written historical records relating to the area before the town developed in the middle of the 19th century. Nevertheless, there are facts that you can use to outline important events. Before the coming of major industry, Abertillery was little more than an area of scattered farms in the ancient parish of Aberystruth. In 1779 the parish minister Edmund Jones described the area thus: "The valley of Tyleri ... is the most delightful. The trees ... especially the beech trees, abounding about rivers great and small, the hedges and lanes make these places exceeding pleasant and the passing by them delightful and affecting ... in these warm valleys, with the prospect of the grand high mountains about them would make very delightful habitations." In 1799 clergyman and historian Archdeacon William Coxe toured the area and in writing a diary of his travels described it as "... richly wooded, and highly cultivated...we looked down with delight upon numerous valleys ... with romantic scenery". The entire population of Aberystruth parish at the turn of the 19th century was just a little over 800. It is not known what the population of Abertillery was at the time but it was probably in the very low hundreds, all of whom would have spoken Welsh only. The area's first deep coal mine was sunk in 1843. The first reading rooms were set up in Abertillery in 1856. However, when Thomas Powell took over the Tillery Colliery in 1882 he made a commitment to establish educational facilities for his workers. Formed in 1877, Abertillery Urban District Council incorporated the adjoining smaller communities of Six Bells, Cwmtillery, Brynithel, Aberbeeg and Llanhilleth. The population of this conurbation climbed to almost 40,000 in 1931 making it the second largest town in Monmouthshire. The council was abolished in 1974 as part of major UK wide local government reorganisation. The reopening of Abertillery railway station has been identified as a future development of the Ebbw Valley Railway. Abertillery Learning Community provides all-through education for the town and neighbouring areas. Until the 1970s the town had its own local authority-run Grammar school providing education up to the age of eighteen. Tertiary education is now provided by Coleg Gwent at Ebbw Vale – opened in 2013. There are several small and medium-sized business parks in the area offering a range of business premises. In 2014 the largest employer was Tyleri Valley Foods. Many local people commute outside the area to work. Abertillery Town cricket club and Abertillery Blaenau Gwent RFC formed in the 1880s. Both have their playing headquarters at "the Park" one of the most picturesque sporting complexes in the UK. The town supports two local Saturday football teams: Abertillery Bluebirds and Abertillery Excelsiors. There are numerous other sports activities running on an organized basis such as bowls, badminton, squash etc. The surrounding landscape provides hill walking opportunities and walker led groups are thriving in the area. One example is Ebbw Fach Trekkers walking group. The local museum has displays showing what life was like in the area in its heyday. It also has its own "valleys" Italian café complete with original furnishings. The Metropole theatre holds musical and drama events – from Blues to amateur dramatics and dance. The "Guardian" memorial is a destination for visitors to South Wales and amateur photographers in particular as evidenced by trip advisor. The visitor centre Tŷ Ebbw Fach stands nearby and provides cafe and visitor "mining valley" experience facilities. Find local people of note in the fields of civil engineering, sport, science, medicine, religion and art. Abertillery is twinned with:
https://en.wikipedia.org/wiki?curid=2717
Abeyance Abeyance (from the Old French "abeance" meaning "gaping") is a state of expectancy in respect of property, titles or office, when the right to them is not vested in any one person, but awaits the appearance or determination of the true owner. In law, the term abeyance can be applied only to such future estates as have not yet vested or possibly may not vest. For example, an estate is granted to A for life, with remainder to the heir of B. During B's lifetime, the remainder is in abeyance, for until the death of A it is uncertain who is B's heir. Similarly the freehold of a benefice, on the death of the incumbent, is said to be in abeyance until the next incumbent takes possession. The term hold in abeyance is used in lawsuits and court cases when a case is temporarily put on hold. The most common use of the term is in the case of English peerage dignities. Most such peerages pass to heirs-male, but the ancient baronies created by writ, as well as some very old earldoms, pass instead to heirs-general (by cognatic primogeniture). In this system, sons are preferred from eldest to youngest, the heirs of a son over the next son, and any son over daughters, but there is no preference among daughters: they or their heirs inherit equally. If the daughter is an only child or her sisters are deceased and have no living issue, she (or her heir) is vested with the title; otherwise, since a peerage cannot be shared nor divided, the dignity goes into abeyance between the sisters or their heirs, and is held by no one. If through lack of issue, marriage, or both, eventually only one person represents the claims of all the sisters, he or she can claim the dignity as a matter of right, and the abeyance is said to be terminated. On the other hand, the number of prospective heirs can grow quite large, since each share potentially can be divided between daughters, where the owner of a share dies without leaving a son. A co-heir may petition the Crown for a termination of the abeyance. The Crown may choose to grant the petition, but if there is any doubt whatsoever as to the pedigree of the petitioner, the claim is normally referred to the Committee for Privileges. If the claim is unopposed, the Committee will generally award the claim, unless there is evidence of collusion, the peerage has been in abeyance for more than a century, or the petitioner holds less than one-third of the claim. This doctrine is a 17th-century innovation, although it is now applied retrospectively for centuries. It cannot be applied perfectly; for example, the eighth Baron De La Warr had three surviving sons; the first died without children, the second left two daughters, and the third left a son. In modern law, the title would have fallen into abeyance between the two daughters of the second son, and nobody else would have been able to claim it even if the abeyance were settled; however, in 1597, the grandson of the third son (whose father had been re-created Baron De La Warr in 1570) claimed the title and its precedence. In 1604, the Baron le Despencer case was the first peerage abeyance ever settled; the second was at the Restoration in 1660. Most subsequent abeyances (only a few dozen cases) were settled after a few years, in favour of the holder of the family properties; there were two periods in which long-abeyant peerages (in some cases peerages of doubtful reality) were brought back: between 1838 and 1841 and between 1909 and 1921. "The Complete Peerage" reports that only baronies have been called out of abeyance, although the Earldom of Cromartie was called out of a two-year abeyance in 1895. It is entirely possible for a peerage to remain in abeyance for centuries. For example, the Barony of Grey of Codnor was in abeyance for over 490 years between 1496 and 1989, and the Barony of Hastings was similarly in abeyance for over 299 years from 1542 to 1841. Some other baronies became abeyant in the 13th century, and the abeyance has yet to be terminated. The only modern examples of titles other than a barony that have yet gone into abeyance are the earldom of Arlington and the viscountcy of Thetford, which are united, and (as noted above) the earldom of Cromartie. It is no longer straightforward to claim English peerages after long abeyances. In 1927 a parliamentary Select Committee on Peerages in Abeyance recommended that no claim should be considered where the abeyance has lasted more than 100 years, nor where the claimant lays claim to less than one third of the dignity. The Barony of Grey of Codnor was treated as an exception to this principle, as a claim to it had been submitted prior to these recommendations being made to the Sovereign. Titles in the Peerage of Scotland cannot go into abeyance, because in Scottish law the eldest sister is preferred over younger sisters; sisters are not considered equal co-heirs. It is common, but incorrect, to speak of peerage dignities which are dormant (i.e. unclaimed) as being in abeyance. Abeyance can be used in cases where parties are interested in temporarily settling litigation while still holding the right to seek relief later if necessary. This may be considered a desirable outcome in cases where the party to the lawsuit is an organization with a transient membership and political perspective. The use of abeyance in such instances can allow such an organization to 'settle' with the party without officially binding its actions in the future, should a new group of decision makers within the organization choose to pursue taking the dispute to court. For example, abeyance was used as a settlement method in a Canadian lawsuit involving the University of Victoria Students' Society (UVSS), the BCCLA, and a campus pro-life club to whom the UVSS denied funding. The parties agreed to settle the lawsuit by holding the case in abeyance in return for the UVSS temporarily giving resources back to the club. With this arrangement, the pro-life club held on to its right to immediately reopen the case again should the UVSS deny resources to the club in the future, and the UVSS was able to avoid an expensive legal battle it did not have the will to pursue at the time. Thus the use of abeyance provided the security of a settlement for the pro-life campus club, while preserving the student society's voting membership's ability to take the matter back to court should they choose in the future to deny resources to the club. Other court cases may be held in abeyance when the issue may be resolved by another court or another event. This saves time and effort trying to resolve a dispute that may be made moot by the other events. During lawsuits related to the Patient Protection and Affordable Care Act after the Supreme Court of the United States granted certiorari in "King v. Burwell", attorneys in "Halbig v. Burwell" requested abeyance of that case as the matter would be resolved in "King" and it would be a waste of time and effort to try to resolve it in the "Halbig" case.
https://en.wikipedia.org/wiki?curid=2720
Anders Celsius Anders Celsius (, 27 November 170125 April 1744) was a Swedish astronomer, physicist and mathematician. He was professor of astronomy at Uppsala University from 1730 to 1744, but traveled from 1732 to 1735 visiting notable observatories in Germany, Italy and France. He founded the Uppsala Astronomical Observatory in 1741, and in 1742 proposed (an inverted form of) the Centigrade temperature scale which was later renamed Celsius in his honor. Anders Celsius was born in Uppsala, Sweden on 27 November 1701. His family originated from Ovanåker in the province of Hälsingland. Their family estate was at "Doma", also known as "Höjen" or "Högen" (locally as "Högen 2"). The name "Celsius" is a latinization of the estate's name (Latin "celsus" "mound"). As the son of an astronomy professor, Nils Celsius, and the grandson of the mathematician Magnus Celsius and the astronomer Anders Spole, Celsius chose a career in science. He was a talented mathematician from an early age. Anders Celsius studied at Uppsala University, where his father was a teacher, and in 1730 he, too, became a professor of astronomy there. In 1730, Celsius published the ("New Method for Determining the Distance from the Earth to the Sun"). His research also involved the study of auroral phenomena, which he conducted with his assistant Olof Hiorter, and he was the first to suggest a connection between the aurora borealis and changes in the magnetic field of the Earth. He observed the variations of a compass needle and found that larger deflections correlated with stronger auroral activity. At Nuremberg in 1733, he published a collection of 316 observations of the aurora borealis made by himself and others over the period 1716–1732. Celsius traveled frequently in the early 1730s, including to Germany, Italy and France, when he visited most of the major European observatories. In Paris he advocated the measurement of an arc of the meridian in Lapland. In 1736, he participated in the expedition organized for that purpose by the French Academy of Sciences, led by the French mathematician Pierre Louis Maupertuis (1698–1759) to measure a degree of latitude. The aim of the expedition was to measure the length of a degree along a meridian, close to the pole, and compare the result with a similar expedition to Peru, today in Ecuador, near the equator. The expeditions confirmed Isaac Newton's belief that the shape of the earth is an ellipsoid flattened at the poles. In 1738, he published the ("Observations on Determining the Shape of the Earth"). Celsius's participation in the Lapland expedition won him much respect in Sweden with the government and his peers, and played a key role in generating interest from the Swedish authorities in donating the resources required to construct a new modern observatory in Uppsala. He was successful in the request, and Celsius founded the Uppsala Astronomical Observatory in 1741. The observatory was equipped with instruments purchased during his long voyage abroad, comprising the most modern instrumental technology of the period. He made observations of eclipses and various astronomical objects and published catalogues of carefully determined magnitudes for some 300 stars using his own photometric system (mean error=0.4 mag). He proposed the Celsius temperature scale in a paper to the Royal Society of Sciences in Uppsala, the oldest Swedish scientific society, founded in 1710. His thermometer was calibrated with a value of 0 for the boiling point of water and 100 for the freezing point. In 1745, a year after Celsius's death, the scale was reversed by Carl Linnaeus to facilitate more practical measurement. Celsius conducted many geographical measurements for the Swedish General map, and was one of earliest to note that much of Scandinavia is slowly rising above sea level, a continuous process which has been occurring since the melting of the ice from the latest ice age. However, he wrongly posed the notion that the water was evaporating. In 1725 he became secretary of the Royal Society of Sciences in Uppsala, and served at this post until his death from tuberculosis in 1744. He supported the formation of the Royal Swedish Academy of Sciences in Stockholm in 1739 by Linnaeus and five others, and was elected a member at the first meeting of this academy. It was in fact Celsius who proposed the new academy's name.
https://en.wikipedia.org/wiki?curid=2722
Adam Carolla Adam Carolla (born May 27, 1964) is an American radio personality, actor and podcaster. He hosts "The Adam Carolla Show", a talk show distributed as a podcast which set the record as the "most downloaded podcast" as judged by Guinness World Records in 2011. Carolla co-hosted the syndicated radio call-in program "Loveline" with Drew Pinsky from 1995 to 2005 as well as the show's television incarnation on MTV from 1996 to 2000. He was the co-host and co-creator of the television program "The Man Show" (1999–2004), and the co-creator and a regular performer on the television show "Crank Yankers" (2002–2007). He hosted "The Adam Carolla Project", a home improvement television program which aired on TLC in 2005 and "The Car Show" on Speed TV in 2011. Carolla has also appeared on the network reality television programs "Dancing with the Stars" and "The Celebrity Apprentice". His book "In Fifty Years We'll All Be Chicks" debuted on "The New York Times" Best Seller list in 2010, and his second book, "Not Taco Bell Material", also reached "New York Times" bestseller status. Carolla has made numerous guest appearances on political talk shows as a commentator. He hosted a weekly segment, "Rollin' with Carolla", on Bill O'Reilly's "The O'Reilly Factor". Adam Carolla was born on May 27, 1964, to Jim and Kris ("née" McCall) Carolla. Some sources list his birthplace as Los Angeles County, California, while others list it as Philadelphia, Pennsylvania. He grew up in Los Angeles's San Fernando Valley, and his parents separated the he was young. Carolla was not given a middle name by his parents; on his driver's license application he listed his middle name as "Lakers" as a joke. The application was processed without notice. His maternal step-grandfather was screenwriter László Görög. Adam was raised in the North Hollywood neighborhood of Los Angeles. He attended Colfax Elementary School, Walter Reed Junior High, and North Hollywood High School. Carolla did not receive his high school diploma until years later as it was held by the school until a library fine was paid. Carolla can be seen paying off the book and receiving his diploma in an episode of his 2005 television show, "The Adam Carolla Project". During his youth, Carolla played Pop Warner football for seven years; he later suggested that being involved in sports saved him from a chaotic home life. During his senior year at North Hollywood High School, Carolla distinguished himself in football. In December 1981, he was named to the First Team Offensive Line, Central Valley League, one of 8 leagues at the time in the LA City Section of the California Interscholastic Federation. He began living on his own at the age of 18. He briefly attended Los Angeles Valley College, a community college, where he was placed on academic probation before dropping out to work in a series of jobs, including carpet cleaner, carpenter, boxing instructor, and traffic school instructor. Although broke, Carolla, his friends, and roommates owned a 1963 Cadillac limousine. In the early 1990s, Carolla studied improvisational comedy with The Groundlings and was a member of the ACME Comedy Theatre troupe. In 1994, Carolla volunteered his services as a boxing trainer to prepare Jimmy Kimmel for a bout being staged by KROQ-FM's morning radio program "Kevin and Bean". Kimmel was a regular on the show as "Jimmy the Sports Guy" and he was set to fight another KROQ personality in a boxing exhibition which was being billed as the "Bleeda in Reseda". Carolla parlayed this opportunity into a long-running friendship and business partnership with Kimmel as well as a recurring role on "Kevin and Bean" as cranky woodshop teacher, Mr. Birchum. In October 1995, after being signed to the William Morris Agency by Mark Itkin, Carolla was offered the job of co-hosting the evening radio call-in show "Loveline." His co-hosts were the physician Drew Pinsky ("Dr. Drew") and metal DJ Riki Rachtman. Carolla received the offer after Pinsky heard him on "Kevin and Bean" (Rachtman left the show the following year.) "Loveline" is broadcast on KROQ-FM in Los Angeles and is syndicated nationwide on the Westwood One radio network. While the unique format of the program was primarily a call-in show where listeners would ask questions about sex and relationships, Carolla would often spend much of the show ranting about various topics, from fart jokes to extended parodies of radio morning shows, including mocking the format's penchant for useless and repetitive weather and traffic reports. In contrast to the reserved, thoughtful Pinsky, Carolla served as the loud, funny side of the show. Carolla's character was described by one reviewer as "a toned-down version of Howard Stern minus the huge ego". On September 25, 2001, Carolla and Jimmy Kimmel both had filled in for Howard Stern's head writer Jackie Martling after Martling's departure from the show in March 2001. Carolla had mentioned in various interviews and on his podcast that he was offered the job to replace Jackie Martling, but turned it down primarily due to Stern's show being located in New York as well as his commitment to "Loveline". In October 2005, Carolla was announced as the host of a new morning radio show on the Infinity Broadcasting network. His new show would replace the popular syndicated "Howard Stern Show" (which was moving to satellite radio) in twelve of the 27 markets in which Stern had been broadcast including Los Angeles, Las Vegas, San Francisco, San Diego, Phoenix, and Portland, Oregon. "The Adam Carolla Show" debuted in January 2006. In early 2008, actor Gerard Butler sat in and observed Adam Carolla on "The Adam Carolla Show" in order to prepare for his role in "The Ugly Truth" as a cynical and crass talk-radio host allegedly based on Carolla. On February 18, 2009, "The Adam Carolla Show" was canceled as part of a format switch at KLSX to AMP FM, a new top 40 station. The final show was Friday, February 20, 2009. Carolla started a daily podcast on February 23, 2009, at his personal website, which would evolve into the ACE Broadcasting Network. The first Adam Carolla podcast was downloaded more than 250,000 times in the initial 24 hours, and by the third podcast, it was the number one podcast on iTunes in both the U.S. and Canada. During the debut week, the Adam Carolla podcast recorded 1.6 million downloads. In the second week it recorded 2.4 million downloads. By the fourth episode of the second week, featuring former "Adam Carolla Show" sidekick Dave Dameshek, the show was downloaded more than 500,000 times. Adam stated that bandwidth cost more than $9,000 a month as of May 2009. At the end of 2009, "The Adam Carolla Podcast" was selected by iTunes for its end-of-the-year awards as the Best Audio Podcast of 2009. On May 18, 2011, Carolla made the claim on "Jimmy Kimmel Live!" that "The Adam Carolla Show" had taken the Guinness World Record for the most downloaded podcast ever from Ricky Gervais after he claims it received 59,574,843 unique downloads from March 2009 to March 16, 2011. In 2013, Personal Audio filed a patent infringement lawsuit against Lotzi Digital, Inc., producers of "The Adam Carolla Show" and several other podcasts on the Carolla Digital Network, in the U.S. District Court for the Eastern District of Texas. The suit alleged that owner Adam Carolla and his network of content infringed on Personal Audio's patent 8,112,504. Using the crowdfunding site FundAnything.com, listeners contributed more than $475,000 (as of August 2014) to support Carolla throughout the legal proceedings. Personal Audio dropped the lawsuit July 29, 2014, stating that the defendants were not "making significant money from infringing Personal Audio's patents". However, Carolla countersued, having already spent hundreds of thousands of dollars mounting a defense against claims he deemed unfounded. Among claims sought by the countersuit is a request that the initial patent be invalidated. On August 15, 2014, Carolla and Personal Audio filed a joint motion to dismiss after reaching a settlement, the details of which are confidential, but include a six-week "quiet period" during which neither party could speak to the media. Both parties' claims were dropped without prejudice, and as such could be re-filed at a later date. After co-hosting (with Dr. Drew) a television version of "Loveline" on MTV during the late 1990s, Carolla began his first original television series with "The Man Show", along with partner and friend Jimmy Kimmel, on Comedy Central from 1999 to 2003. He left "The Man Show" at the same time as Kimmel. Carolla has continued his work with Kimmel as a writer and guest on "Jimmy Kimmel Live!". He also appeared on an episode of "Space Ghost Coast to Coast" around this time. Carolla and partner Daniel Kellison are the heads of Jackhole Productions. The two created the television show "Crank Yankers" for Comedy Central, which revived the Mr. Birchum character. (The show premiered in 2002 on Comedy Central and returned to MTV2 on February 9, 2007, running again until March 30, 2007. The show screened in Australia on SBS Television and The Comedy Channel between 2003 and 2008.) From August 2005 to November 2005, Carolla hosted the talk show "Too Late with Adam Carolla" on Comedy Central. Also in 2005, Carolla was featured in a home remodeling program called "The Adam Carolla Project" wherein he and a crew of old friends renovated his childhood home. The 13 episodes aired on the cable channel TLC (The Learning Channel) from October through December 2005. The house was then sold for 1.2 million dollars. In 2006, Carolla appeared on the special summer series "Gameshow Marathon" as a celebrity panelist on the Match Game episode. On the February 18, 2008 broadcast of his radio show, Carolla announced that he would be one of the contestants on the next season of "Dancing with the Stars". Later in the broadcast, it was revealed to Adam that his partner would be Julianne Hough. He was voted off on the April 8, 2008 episode after his performance of the Paso Doble, after incorporating a demonstration of unicycle riding in his dance routine. On June 16, 2008 Carolla was selected to host a pilot of an American version of the popular BBC show "Top Gear" for NBC. In December 2008, NBC decided not to pick up the show. On February 21, 2009, a day after his Los Angeles-based morning radio show was canceled – as part of a format change at KLSX-FM – CBS ordered a comedy pilot starring the actor/comedian. "Ace in the Hole" was to star Carolla as a husband and father who works as a driving instructor. Carolla created and wrote the pilot with Kevin Hench ("Jimmy Kimmel Live!"). Carolla stated that Pamela Adlon was to play his wife and Windell Middlebrooks of the Miller High Life commercial fame will play his best friend. During his March 30, 2009, podcast, Carolla briefly described the show as being ""All in the Family", essentially", with Carolla playing a similar role to that of Archie Bunker. On the July 23, 2009, episode of the Adam Carolla Podcast, Carolla announced that CBS was not picking up the pilot for the 2009 season, "in any way, shape or form". On October 22, 2009, it was reported in "Variety" that Carolla had struck a deal with NBC to produce a half-hour pilot for a sitcom. The report was later confirmed on January 4, 2010, and was the first comedy pilot ordered by NBC for the season. The untitled project, written by Carolla and Kevin Hench, was a single-camera endeavor that starred Carolla as a contractor and father who attempts to rebuild his life after his wife leaves him. Carolla was set to executive produce the NBC project along with frequent collaborators Kimmel and Hench, as well as his agent James "Babydoll" Dixon, Jon Pollack from "30 Rock", Gail Berman, Daniel Kellison, and Lloyd Braun. Universal Media Studios, BermanBraun, and Carolla and Jimmy Kimmel's own Jackhole Industries. On the February 13, 2010 episode of Carolla's CarCast podcast, he broke the news that The History Channel had picked up "Top Gear US", which NBC had decided against in 2008. On the March 26, 2010, episode of CarCast, Carolla said that he would not be co-hosting "Top Gear US" because of scheduling conflicts with his NBC sitcom project. In June 2010, Carolla said that his NBC pilot had not been picked up and was now "dead". Premiering on February 19, 2012, Carolla was also one of the contestants in the 12th season of NBC's "The Celebrity Apprentice". He was fired in Week 4, because Trump perceived that Carolla did not utilize Andretti's car background during a Buick presentation, even though it was clear that Andretti did not like public speaking. Carolla's "The Car Show" debuted on Speed TV July 13, 2011. Appearing Wednesdays at 10 pm Eastern, it featured Carolla as the host, along with Dan Neil, John Salley, and Matt Farah. It had a format similar to Top Gear, mixing car reviews, tests and humor. The show was initially met with largely positive reviews from car enthusiasts and comedy fans. Talk show host and comedian Jay Leno called The Car Show, "a lot of fun". The Car Show was cancelled after one season, after undergoing format changes due to low ratings, as Carolla mentioned on his podcast January 13, 2012. "Catch a Contractor" is a non-scripted, original series on Spike, hosted by Adam Carolla along with "no-nonsense contractor" Skip Bedell and his wife, investigator Alison Bedell. Together they expose crooked contractors and seek retribution for wronged homeowners. The show premiered on March 9, 2014, to 1.2 million viewers, the largest audience for a series debut on Spike since "Coal" in March 2011. The show was cancelled in 2015. Premiering on Spike TV on March 14, 2017, "Adam Carolla and Friends Build Stuff Live" features Carolla building projects live and in studio with some of his Hollywood friends, and tackling viewers' home improvement projects via social media. Carolla has also done voice acting on cartoons, including Commander Nebula on the Disney animated series "Buzz Lightyear of Star Command", Death on "Family Guy" (replacing Norm Macdonald) and Spanky Ham on "Drawn Together". He was also the voice of the éclair police officer, Wynchell, in the Disney film "Wreck-It Ralph". In 2008 and 2009, he was the spokesperson for T.G.I. Friday's. In 2003, he appeared in "Windy City Heat" as himself. In 2006, Carolla finished work on "The Hammer", a semi-autobiographical independent film he co-wrote and co-produced, in which he stars opposite Heather Juergensen. The film is based loosely on his own life and is filmed at a gym he helped build with his co-star, Ozzie, played by Oswaldo Castillo, his friend in real life whom he met while building the gym when they both worked in construction. The film made its world premiere at the 2007 Tribeca Film Festival in New York City and shortly thereafter received a positive review in Variety. The film was released on March 21, 2008. The film is rated 80% on Rotten Tomatoes. Adam made a short appearance in Jeff Balis' "Still Waiting..." (a sequel to "Waiting...") playing a pick-up artist guru. Adam helped write an unproduced screenplay for a movie entitled "Deaf Frat Guy: Showdown at Havasu". He is the voice of Virgil in the independent short film "Save Virgil". In July 2013, Carolla used crowdfunding for "Road Hard"; a film he directed and starred in, about the lives of aging road comics. Adam confirmed through a press conference that the film would co-star David Alan Grier, Illeana Douglas, Diane Farr, and Larry Miller. It had limited theatrical release in the United States. Several minutes of the credits are devoted to listing the names of those who helped crowdfund the movie. Carolla also directed the documentary "" on the 35-year car racing career of Paul Newman. The documentary showcases Newman's racing life as both a prolific driver and owner. In 2017, Carolla and Dennis Prager began filming "No Safe Spaces", a documentary about political correctness at universities. "No Safe Spaces" had a limited opening on October 25th 2019, and did well enough to open nationwide on December 6th 2019. Carolla and Drew Pinsky co-wrote (with Marshall Fine) the self-help book "The Dr. Drew and Adam Book: A Survival Guide to Life and Love," published in 1998. The book is a compilation of some of the advice the pair compiled while producing Loveline. In November 2010, Carolla's "In Fifty Years We'll All Be Chicks... And Other Complaints from an Angry Middle-Aged White Guy" was published by Crown Archetype and debuted at number eight on the "New York Times" Best Seller list for hardcover non-fiction on November 21, 2010. The book was compiled from rants Carolla had delivered on his radio show and podcast along with some new material and was dictated to and ghost-written by Mike Lynch. Carolla published a short, illustrated e-book entitled "Rich Man, Poor Man" in January 2012. The book details some similarities in the experiences of the very rich and the very poor which are not shared by the middle class. The book was illustrated by Michael Narren. Carolla's book "Not Taco Bell Material" was published by Crown Archetype on June 12, 2012. It climbed the "New York Times" bestseller list and received widespread praise for its satirical humor and honest, self-deprecating style. In "President Me: The America That's in My Head," Carolla presents the comedian's fantasy of the United States with him at the helm. When asked in separate interviews, both before and after the book's release, about whether the "if-I-were-king" critique of America was a serious piece, he said it's both: "Well, there's a lot of jokes in it, but you know, it's like... Well, if you have a fat friend you may make a lot of fat jokes about your fat friend, but he's still fat". In "Daddy, Stop Talking!: And Other Things My Kids Want But Won't Be Getting," Carolla writes about modern parenting. Carolla describes what he believes adults must do if they don't want to have to support their kids forever. Carolla uses his own childhood as a cautionary tale, and decries helicopter parenting. Corolla's latest book, "I'm Your Emotional Support Animal: Navigating Our All Woke, No Joke Culture", was published by Post Hill Press on June 16, 2020. Carolla has occasionally made generalizations about various groups, leading to criticism. For example, in a late-2003 "Loveline" episode, Carolla said that Hawaiians are "dumb", "in-bred", "retarded" people who are among the "dumbest people we have". The comments were met with anger in Hawaii and resulted in "Loveline"s cancellation on Hawaiian affiliate KPOI. Carolla had made similar comments about Hawaiians in previous shows as well. On the April 4, 2010, episode of "The Adam Carolla Show", Carolla referred to Filipino boxer Manny Pacquiao as a "fucking idiot" and said of the Philippines: "They got this and sex tours, that's all they have over there. Get your shit together, Philippines." A spokesman for President Gloria Macapagal-Arroyo called Carolla an "ignorant fool". Carolla subsequently apologized via Twitter. In 2010, Carolla posed for the NOH8 Campaign. In August 2011, Carolla released a podcast where he mocked a petition to the producers of "Sesame Street" that demanded Bert and Ernie get married on air. He said on air that gay activists should "[j]ust get married, and please shut up" and that "Y.U.C.K." would be more memorable acronym than LGBT, and referring to transgender people he asked: "When did we start giving a shit about these people?" The Gay & Lesbian Alliance Against Defamation characterized the previous remarks by Carolla as offensive, including an assertion that "all things being equal", heterosexual parents make better parents than homosexual parents. Carolla responded: "I'm sorry my comments were hurtful. I'm a comedian, not a politician." GLAAD called the apology "empty." In June 2012, Carolla became the target of charges of sexism due to his remarks during an interview published in the "New York Post": Carolla's comments were widely reported and criticized as sexist. Carolla criticized coverage of his comments as over-simplistic and misleading. On September 28, 2002, Carolla married Lynette Paradise. The couple's twins Natalia and Santino "Sonny" Richard Carolla were born June 7, 2006. On "The Adam Carolla Show", Carolla repeatedly mentioned the birth was originally scheduled for June 6, but that he and his wife pushed it one day as to avoid the symbolic 666 (06–06–06). Carolla was a part owner of Amalfi, an Italian restaurant in Los Angeles, saying, "I own about two percent of it, but I’ve never seen a penny." Carolla won the 2013 Pro/Celebrity Race as a professional and the 2012 Pro/Celebrity Race at the Toyota Grand Prix of Long Beach as an amateur. The 2012 race was run on April 14, 2012 and was broadcast on Speed TV. Carolla has previously participated in the race in 2010 and 2003. He finished ninth among 19 racers (fifth among the ten celebrities) in 2010 despite being regarded as a pre-race favorite. He is also a serious automobile collector with over 20 cars. His collection includes several Lamborghinis from the 1960s and early 1970s, including two Miuras (of 764 examples ever produced), one of which he has loaned to the Petersen Automotive Museum in Los Angeles, two 400GT 2+2s (of 247 units produced) and a 1965 350GT (one of 135 built). At least one Ferrari and Aston Martin and several vintage race cars round out the collection. Regarding his political views, Carolla has stated, "I guess I would be Republican, in the sense that I want a secure border, I'm not into the welfare state, I'm not into all those freebie lunch programs. It just kind of demeans people." He goes on to state, however, that he is also in favor of typically liberal causes such as the legalization of marijuana (he is a member of the advisory board of the Marijuana Policy Project) and support for as well some progressive causes such as "...  [being] against semi-automatic and automatic weapons. I'm not an NRA guy by any stretch of the imagination. I'd like alternative energy to be explored and electric cars to be used, but I want them to be powered by nuclear power plants." Elsewhere, he has stated, "My feeling is this whole country is founded on the principle of 'If you are not hurting anyone, and you're not fucking with someone else's shit, and you are paying your taxes, you should be able to just do what you want to do.' It's the freedom and the independence." In an interview with "Reason TV", Carolla described his views as libertarian. Carolla and Pinsky received a Sexual Health in Entertainment (SHINE) Award from The Media Project in 2000 for "incorporating accurate and honest portrayals of sexuality" in the talk show category for "Loveline". Asteroid 4535 Adamcarolla is named in his honor.
https://en.wikipedia.org/wiki?curid=2723
Autocorrelation Autocorrelation, also known as serial correlation, is the correlation of a signal with a delayed copy of itself as a function of delay. Informally, it is the similarity between observations as a function of the time lag between them. The analysis of autocorrelation is a mathematical tool for finding repeating patterns, such as the presence of a periodic signal obscured by noise, or identifying the missing fundamental frequency in a signal implied by its harmonic frequencies. It is often used in signal processing for analyzing functions or series of values, such as time domain signals. Different fields of study define autocorrelation differently, and not all of these definitions are equivalent. In some fields, the term is used interchangeably with autocovariance. Unit root processes, trend stationary processes, autoregressive processes, and moving average processes are specific forms of processes with autocorrelation. In statistics, the autocorrelation of a real or complex random process is the Pearson correlation between values of the process at different times, as a function of the two times or of the time lag. Let formula_1 be a random process, and formula_2 be any point in time (formula_2 may be an integer for a discrete-time process or a real number for a continuous-time process). Then formula_4 is the value (or realization) produced by a given run of the process at time formula_2. Suppose that the process has mean formula_6 and variance formula_7 at time formula_2, for each formula_2. Then the definition of the auto-correlation function between times formula_10 and formula_11 is where formula_12 is the expected value operator and the bar represents complex conjugation. Note that the expectation may not be well defined. Subtracting the mean before multiplication yields the auto-covariance function between times formula_10 and formula_11: Note that this expression is not well-defined for all-time series or processes, because the mean may not exist, or the variance may be zero (for a constant process) or infinite (for processes with distribution lacking well-behaved moments, such as certain types of power law). If formula_1 is a wide-sense stationary process then the mean formula_16 and the variance formula_17 are time-independent, and further the autocovariance function depends only on the lag between formula_10 and formula_11: the autocovariance depends only on the time-distance between the pair of values but not on their position in time. This further implies that the autocovariance and auto-correlation can be expressed as a function of the time-lag, and that this would be an even function of the lag formula_20. This gives the more familiar forms for the auto-correlation function and the auto-covariance function: It is common practice in some disciplines (e.g. statistics and time series analysis) to normalize the autocovariance function to get a time-dependent Pearson correlation coefficient. However, in other disciplines (e.g. engineering) the normalization is usually dropped and the terms "autocorrelation" and "autocovariance" are used interchangeably. The definition of the auto-correlation coefficient of a stochastic process is If the function formula_22 is well-defined, its value must lie in the range formula_23, with 1 indicating perfect correlation and −1 indicating perfect anti-correlation. For a weak-sense stationarity, wide-sense stationarity (WSS) process, the definition is where The normalization is important both because the interpretation of the autocorrelation as a correlation provides a scale-free measure of the strength of statistical dependence, and because the normalization has an effect on the statistical properties of the estimated autocorrelations. The fact that the auto-correlation function formula_26 is an even function can be stated as Respectively for a WSS process: For a WSS process: Notice that formula_30 is always real. The Cauchy–Schwarz inequality, inequality for stochastic processes: The autocorrelation of a continuous-time white noise signal will have a strong peak (represented by a Dirac delta function) at formula_32 and will be exactly 0 for all other formula_33. The Wiener–Khinchin theorem relates the autocorrelation function formula_26 to the power spectral density formula_35 via the Fourier transform: For real-valued functions, the symmetric autocorrelation function has a real symmetric transform, so the Wiener–Khinchin theorem can be re-expressed in terms of real cosines only: The auto-correlation matrix (also called second moment) of a random vector formula_40 is an formula_41 matrix containing as elements the autocorrelations of all pairs of elements of the random vector formula_42. The autocorrelation matrix is used in various digital signal processing algorithms. For a random vector formula_40 containing random elements whose expected value and variance exist, the auto-correlation matrix is defined by where formula_44 denotes transposition and has dimensions formula_41. Written component-wise: If formula_47 is a complex random vector, the autocorrelation matrix is instead defined by Here formula_49 denotes Hermitian transposition. For example, if formula_50 is a random vectors, then formula_51 is a formula_52 matrix whose formula_53-th entry is formula_54. In signal processing, the above definition is often used without the normalization, that is, without subtracting the mean and dividing by the variance. When the autocorrelation function is normalized by mean and variance, it is sometimes referred to as the autocorrelation coefficient or autocovariance function. Given a signal formula_59, the continuous autocorrelation formula_60 is most often defined as the continuous cross-correlation integral of formula_59 with itself, at lag formula_33. where formula_63 represents the complex conjugate of formula_59. Note that the parameter formula_2 in the integral is a dummy variable and is only necessary to calculate the integral. It has no specific meaning. The discrete autocorrelation formula_66 at lag formula_67 for a discrete-time signal formula_68 is The above definitions work for signals that are square integrable, or square summable, that is, of finite energy. Signals that "last forever" are treated instead as random processes, in which case different definitions are needed, based on expected values. For wide-sense-stationary random processes, the autocorrelations are defined as For processes that are not stationary, these will also be functions of formula_2, or formula_72. For processes that are also ergodic, the expectation can be replaced by the limit of a time average. The autocorrelation of an ergodic process is sometimes defined as or equated to These definitions have the advantage that they give sensible well-defined single-parameter results for periodic functions, even when those functions are not the output of stationary ergodic processes. Alternatively, signals that "last forever" can be treated by a short-time autocorrelation function analysis, using finite time integrals. (See short-time Fourier transform for a related process.) If formula_75 is a continuous periodic functions of period formula_76, the integration from formula_77 to formula_78 is replaced by integration over any interval formula_79 of length formula_76: which is equivalent to In the following, we will describe properties of one-dimensional autocorrelations only, since most properties are easily transferred from the one-dimensional case to the multi-dimensional cases. These properties hold for wide-sense stationary processes. Multi-dimensional autocorrelation is defined similarly. For example, in three dimensions the autocorrelation of a square-summable discrete signal would be When mean values are subtracted from signals before computing an autocorrelation function, the resulting function is usually called an auto-covariance function. For data expressed as a discrete sequence, it is frequently necessary to compute the autocorrelation with high computational efficiency. A brute force method based on the signal processing definition formula_98 can be used when the signal size is small. For example, to calculate the autocorrelation of the real signal sequence formula_99 (i.e. formula_100, and formula_101 for all other values of ) by hand, we first recognize that the definition just given is the same as the "usual" multiplication, but with right shifts, where each vertical addition gives the autocorrelation for particular lag values: Thus the required autocorrelation sequence is formula_103, where formula_104 formula_105 and formula_106 the autocorrelation for other lag values being zero. In this calculation we do not perform the carry-over operation during addition as is usual in normal multiplication. Note that we can halve the number of operations required by exploiting the inherent symmetry of the autocorrelation. If the signal happens to be periodic, i.e. formula_107 then we get a circular autocorrelation (similar to circular convolution) where the left and right tails of the previous autocorrelation sequence will overlap and give formula_108 which has the same period as the signal sequence formula_109 The procedure can be regarded as an application of the convolution property of z-transform of a discrete signal. While the brute force algorithm is order , several efficient algorithms exist which can compute the autocorrelation in order . For example, the Wiener–Khinchin theorem allows computing the autocorrelation from the raw data with two fast Fourier transforms (FFT): where IFFT denotes the inverse fast Fourier transform. The asterisk denotes complex conjugate. Alternatively, a multiple correlation can be performed by using brute force calculation for low values, and then progressively binning the data with a logarithmic density to compute higher values, resulting in the same efficiency, but with lower memory requirements. For a discrete process with known mean and variance for which we observe formula_72 observations formula_112, an estimate of the autocorrelation may be obtained as for any positive integer formula_114. When the true mean formula_16 and variance formula_17 are known, this estimate is unbiased. If the true mean and variance of the process are not known there are a several possibilities: The advantage of estimates of the last type is that the set of estimated autocorrelations, as a function of formula_123, then form a function which is a valid autocorrelation in the sense that it is possible to define a theoretical process having exactly that autocorrelation. Other estimates can suffer from the problem that, if they are used to calculate the variance of a linear combination of the formula_124's, the variance calculated may turn out to be negative. In regression analysis using time series data, autocorrelation in a variable of interest is typically modeled either with an autoregressive model (AR), a moving average model (MA), their combination as an autoregressive-moving-average model (ARMA), or an extension of the latter called an autoregressive integrated moving average model (ARIMA). With multiple interrelated data series, vector autoregression (VAR) or its extensions are used. In ordinary least squares (OLS), the adequacy of a model specification can be checked in part by establishing whether there is autocorrelation of the regression residuals. Problematic autocorrelation of the errors, which themselves are unobserved, can generally be detected because it produces autocorrelation in the observable residuals. (Errors are also known as "error terms" in econometrics.) Autocorrelation of the errors violates the ordinary least squares assumption that the error terms are uncorrelated, meaning that the Gauss Markov theorem does not apply, and that OLS estimators are no longer the Best Linear Unbiased Estimators (BLUE). While it does not bias the OLS coefficient estimates, the standard errors tend to be underestimated (and the t-scores overestimated) when the autocorrelations of the errors at low lags are positive. The traditional test for the presence of first-order autocorrelation is the Durbin–Watson statistic or, if the explanatory variables include a lagged dependent variable, Durbin's h statistic. The Durbin-Watson can be linearly mapped however to the Pearson correlation between values and their lags. A more flexible test, covering autocorrelation of higher orders and applicable whether or not the regressors include lags of the dependent variable, is the Breusch–Godfrey test. This involves an auxiliary regression, wherein the residuals obtained from estimating the model of interest are regressed on (a) the original regressors and (b) "k" lags of the residuals, where 'k' is the order of the test. The simplest version of the test statistic from this auxiliary regression is "TR"2, where "T" is the sample size and "R"2 is the coefficient of determination. Under the null hypothesis of no autocorrelation, this statistic is asymptotically distributed as formula_125 with "k" degrees of freedom. Responses to nonzero autocorrelation include generalized least squares and the Newey–West HAC estimator (Heteroskedasticity and Autocorrelation Consistent). In the estimation of a moving average model (MA), the autocorrelation function is used to determine the appropriate number of lagged error terms to be included. This is based on the fact that for an MA process of order "q", we have formula_126, for formula_127, and formula_128, for formula_129. Serial dependence is closely linked to the notion of autocorrelation, but represents a distinct concept (see Correlation and dependence). In particular, it is possible to have serial dependence but no (linear) correlation. In some fields however, the two terms are used as synonyms. A time series of a random variable has serial dependence if the value at some time formula_2 in the series is statistically dependent on the value at another time formula_131. A series is serially independent if there is no dependence between any pair. If a time series formula_1 is stationary, then statistical dependence between the pair formula_133 would imply that there is statistical dependence between all pairs of values at the same lag formula_134.
https://en.wikipedia.org/wiki?curid=2724
Atlas Autocode Atlas Autocode (AA) is a programming language developed around 1965 at the University of Manchester. A variant of the language ALGOL, it was developed by Tony Brooker and Derrick Morris for the Atlas computer. The word "Autocode" was basically an early term for "programming language". Different autocodes could vary greatly. AA featured explicitly typed variables, subroutines, and functions. It omitted some ALGOL features such as "passing parameters by name", which in ALGOL 60 means passing the memory address of a short subroutine to recalculate a parameter each time it is mentioned. The AA compiler generated range-checking for array accesses, and allowed an array to have dimensions that were determined at runtime, i.e., an array could be declared as codice_1, where codice_2 and codice_3 were calculated values. AA high-level routines can include machine code, either to make an inner loop more efficient or to effect some operation which otherwise cannot be done easily. AA includes a codice_4 data type to represent complex numbers, partly because of pressure from the electrical engineering department, as complex numbers are used to represent the behavior of alternating current. The imaginary unit square root of -1 was represented by codice_2, which was treated as a fixed complex constant = "i". The codice_4 data type was dropped when Atlas Autocode later evolved into the language Edinburgh IMP. IMP was an extension of AA and was used to write the Edinburgh Multiple Access System (EMAS) operating system. AA's second-greatest claim to fame (after being the progenitor of IMP and EMAS) was that it had many of the features of the original "Compiler Compiler". A variant of the AA compiler included run-time support for a top-down recursive descent parser. The style of parser used in the Compiler Compiler was in use continuously at Edinburgh from the 60's until almost the year 2000. Other Autocodes were developed for the Titan computer, a prototype Atlas 2 at Cambridge, and the Ferranti Mercury. Atlas Autocode's syntax was largely similar to ALGOL, though it was influenced by the output device which the author had available, a Friden Flexowriter. Thus, it allowed symbols like codice_7 for .5 and the superscript 2 for "to the power of 2". The Flexowriter supported overstriking and thus, AA did also: up to three characters could be overstruck as a single symbol. For example, the character set had no codice_8 symbol, so exponentiation was an overstrike of codice_9 and codice_10. The aforementioned underlining of reserved words (keywords) could also be done using overstriking. The language is described in detail in the Atlas Autocode Reference Manual. Other Flexowriter characters that were found a use in AA were: codice_11 in floating-point numbers, "e.g.", codice_12 for modern codice_13 ; codice_14 to mean "the second half of a 48-bit Atlas memory word"; codice_15 for the mathematical constant pi. When AA was ported to the English Electric KDF9 computer, the character set was changed to International Organization for Standardization (ISO) and that compiler has been recovered from an old paper tape by the Edinburgh Computer History Project and is available online, as is a high-quality scan of the original Edinburgh version of the Atlas Autocode manual. Keywords in AA were distinguishable from other text by being underlined, which was implemented via overstrike in the Flexowriter (compare to bold in ALGOL). There were also two stropping regimes. First, there was an "uppercasedelimiters" mode where all uppercase letters (outside strings) were treated as underlined lowercase. Second, in some versions (but not in the original Atlas version), it was possible to strop keywords by placing a "codice_16" sign in front of them, for example the keyword codice_17 could be typed as codice_18 or codice_19. This significantly reduced typing, due to only needing one character, rather than overstriking the whole keyword. As in ALGOL, there were no reserved words in the language as keywords were identified by underlining (or stropping), not by recognising reserved character sequences. In the statement codice_20, there is both a keyword codice_21 and a variable named codice_21. As in ALGOL, AA allowed spaces in variable names, such as codice_23. Spaces were not significant and were removed before parsing in a trivial pre-lexing stage called "line reconstruction". What the compiler would see in the above example would be "codice_24". Spaces were possible due partly to keywords being distinguished in other ways, and partly because the source was processed by scannerless parsing, without a separate lexing phase, which allowed the lexical syntax to be context-sensitive. The syntax for expressions let the multiplication operator be omitted, e.g., codice_25 was treated as codice_26, and codice_27 was treated as codice_28 if codice_29 was not an array. In ambiguous uses, the longest possible name was taken (maximal munch), for example codice_30 was not treated as codice_31, whether or not codice_29 and codice_33 had been declared. In the original Atlas Autocode for the Atlas computer, Atlas machine code instructions could be interpolated between the AA statements.
https://en.wikipedia.org/wiki?curid=2726
Au file format The Au file format is a simple audio file format introduced by Sun Microsystems. The format was common on NeXT systems and on early Web pages. Originally it was headerless, being simply 8-bit μ-law-encoded data at an 8000 Hz sample rate. Hardware from other vendors often used sample rates as high as 8192 Hz, often integer multiples of video clock signal frequencies. Newer files have a header that consists of six unsigned 32-bit words, an optional information chunk and then the data (in big endian format). Although the format now supports many audio encoding formats, it remains associated with the μ-law logarithmic encoding. This encoding was native to the SPARCstation 1 hardware, where SunOS exposed the encoding to application programs through the /dev/audio interface. This encoding and interface became a de facto standard for Unix sound. All fields are stored in big-endian format, including the sample data. The type of encoding depends on the value of the "encoding" field (word 3 of the header). Formats 2 through 7 are uncompressed linear PCM, therefore technically lossless (although not necessarily free of quantization error, especially in 8-bit form). Formats 1 and 27 are μ-law and A-law, respectively, both companding logarithmic representations of PCM, and arguably lossy as they pack what would otherwise be almost 16 bits of dynamic range into 8 bits of encoded data, even though this is achieved by an altered dynamic response and no data is actually "thrown away". Formats 23 through 26 are ADPCM, which is an early form of lossy compression, usually but not always with 4 bits of encoded data per audio sample (for 4:1 efficiency with 16-bit input, or 2:1 with 8-bit; equivalent to e.g. encoding CD quality MP3 at a 352kbit rate using a low quality encoder). Several of the others are DSP commands or data, designed to be processed by the NeXT Music Kit software. Note: PCM formats are encoded as signed data (as opposed to unsigned). Following the header structure is a variable-length annotation field. The contents of this field are currently undefined, except that its length must be a multiple of eight bytes and it must be terminated with at least one null (zero) byte. The audio data segment begins on an eight-byte boundary immediately following the annotation field. Audio data is encoded in the format identified by the file header. The current implementation supports only a single audio data segment per file. The variable-length annotation field is currently ignored by most audio applications.
https://en.wikipedia.org/wiki?curid=2732
Andalusia Andalusia region (, ; ) is the southern autonomous community in Peninsular Spain. It is the most populous, and the second largest autonomous community in the country. The Andalusian autonomous community is officially recognised as a "historical nationality". The territory is divided into eight provinces: Almería, Cádiz, Córdoba, Granada, Huelva, Jaén, Málaga and Seville. Its capital is the city of Seville. The seat of the High Court of Justice of Andalusia is located in the city of Granada. Andalusia is located in the south of the Iberian peninsula, in southwestern Europe, immediately south of the autonomous communities of Extremadura and Castilla-La Mancha; west of the autonomous community of Murcia and the Mediterranean Sea; east of Portugal and the Atlantic Ocean; and north of the Mediterranean Sea and the Strait of Gibraltar. Andalusia is the only European region with both Mediterranean and Atlantic coastlines. The small British overseas territory of Gibraltar shares a three-quarter-mile land border with the Andalusian province of Cádiz at the eastern end of the Strait of Gibraltar. The main mountain ranges of Andalusia are the Sierra Morena and the Baetic System, consisting of the Subbaetic and Penibaetic Mountains, separated by the Intrabaetic Basin. In the north, the Sierra Morena separates Andalusia from the plains of Extremadura and Castile–La Mancha on Spain's Meseta Central. To the south the geographic subregion of Upper Andalusia lies mostly within the Baetic System, while Lower Andalusia is in the Baetic Depression of the valley of the Guadalquivir. The name "Andalusia" is derived from the Arabic word "Al-Andalus" (الأندلس). The toponym "al-Andalus" is first attested by inscriptions on coins minted in 716 by the new Muslim government of Iberia. These coins, called "dinars", were inscribed in both Latin and Arabic. The etymology of the name ""al-Andalus"" has traditionally been derived from the name of the "Vandals"; however, a number of proposals since the 1980s have challenged this contention. Halm in 1989 derived the name from a Gothic term, "*", and in 2002, Bossong suggested its derivation from a pre-Roman substrate. The region's history and culture have been influenced by the native Iberians, Phoenicians, Carthaginians, Greeks, Romans, Vandals, Visigoths, Byzantines, Berbers of North Africa, Jews, Romani, Arab Umayyads, and Moors. During the Islamic Golden Age, Cordoba surpassed Constantinople to be Europe's biggest city, and became the capital of the Al Andalus and a prominent center of education and learning in the world, producing numerous philosophers and scientists. The Castilian and other Christian North Iberian nationalities reconquered and settled the area in the latter phases of the "Reconquista". Andalusia has historically been an agricultural region, compared to the rest of Spain and the rest of Europe. However, the growth of the community especially in the sectors of industry and services was above average in Spain and higher than many communities in the Eurozone. The region has a rich culture and a strong identity. Many cultural phenomena that are seen internationally as distinctively Spanish are largely or entirely Andalusian in origin. These include flamenco and, to a lesser extent, bullfighting and Hispano-Moorish architectural styles, both of which are also prevalent in some other regions of Spain. Andalusia's hinterland is the hottest area of Europe, with cities like Córdoba and Seville averaging above 36 °C (97 °F) in summer high temperatures. Late evening temperatures can sometimes stay around 35 °C (95 °F) until close to midnight and daytime highs of over 40 °C (104 °F) are common. Seville also has the highest average annual temperature in mainland Spain and mainland Europe (19.2 °C), closely followed by Almería (19.1 °C). Its present form is derived from the Arabic name for Muslim Iberia, "Al-Andalus". However, the etymology of the name "Al-Andalus" is disputed, and the extent of Iberian territory encompassed by the name has changed over the centuries. The Spanish place name "Andalucía" (immediate source of the English "Andalusia") was introduced into the Spanish languages in the 13th century under the form "el Andalucía". The name was adopted to refer to those territories still under Moorish rule, and generally south of Castilla Nueva and Valencia, and corresponding with the former Roman province hitherto called Baetica in Latin sources. This was a Castilianization of "Al-Andalusiya", the adjectival form of the Arabic language "al-Andalus", the name given by the Arabs to all of the Iberian territories under Muslim rule from 711 to 1492. The etymology of "al-Andalus" is itself somewhat debated (see al-Andalus), but in fact it entered the Arabic language before this area came under Muslim rule. Like the Arabic term "al-Andalus", in historical contexts the Spanish term "Andalucía" or the English term "Andalusia" do not necessarily refer to the exact territory designated by these terms today. Initially, the term referred exclusively to territories under Muslim control; later, it was applied to some of the last Iberian territories to be regained from the Muslims, though not always to exactly the same ones. In the "Estoria de España" (also known as the "Primera Crónica General") of Alfonso X of Castile, written in the second half of the 13th century, the term "Andalucía" is used with three different meanings: From an administrative point of view, Granada remained separate for many years even after the completion of the "Reconquista" due, above all, to its emblematic character as the last territory regained, and as the seat of the important Real Chancillería de Granada, a court of last resort. Still, the reconquest and repopulation of Granada was accomplished largely by people from the three preexisting Christian kingdoms of Andalusia, and Granada came to be considered a fourth kingdom of Andalusia. The often-used expression "Four Kingdoms of Andalusia" dates back in Spanish at least to the mid-18th century. The Andalusian emblem shows the figure of Hercules and two lions between the two pillars of Hercules that tradition situates on either side of the Strait of Gibraltar. An inscription below, superimposed on an image of the flag of Andalusia reads "Andalucía por sí, para España y la Humanidad" ("Andalusia for herself, Spain and Humanity"). Over the two columns is a semicircular arch in the colours of the flag of Andalusia, with the Latin words "Dominator Hercules Fundator" (Lord Hercules is the Founder) superimposed. The official flag of Andalusia consists of three equal horizontal stripes, coloured green, white, and green respectively; the Andalusian coat of arms is superimposed on the central stripe. Its design was overseen by Blas Infante and approved in the Assembly of Ronda (a 1918 gathering of Andalusian nationalists at Ronda). Blas Infante considered these to have been the colours most used in regional symbols throughout the region's history. According to him, the green came in particular from the standard of the Umayyad Caliphate and represented the call for a gathering of the populace. The white symbolised pardon in the Almohad dynasty, interpreted in European heraldry as parliament or peace. Other writers have justified the colours differently, with some Andalusian nationalists referring to them as the "Arbonaida", meaning white-and-green in Mozarabic, a Romance language that was spoken in the region in Muslim times. Nowadays, the Andalusian government states that the colours of the flag evoke the Andalusian landscape as well as values of purity and hope for the future. The anthem of Andalusia was composed by José del Castillo Díaz (director of the Municipal Band of Seville, commonly known as Maestro Castillo) with lyrics by Blas Infante. The music was inspired by "Santo Dios", a popular religious song sung at harvest time by peasants and day labourers in the provinces of Málaga, Seville, and Huelva. Blas Infante brought the song to Maestro Castillo's attention; Maestro Castillo adapted and harmonized the traditional melody. The lyrics appeal to the Andalusians to mobilise and demand "tierra y libertad" ("land and liberty") by way of agrarian reform and a statute of autonomy within Spain. The Parliament of Andalusia voted unanimously in 1983 that the preamble to the Statute of Autonomy recognise Blas Infante as the Father of the Andalusian Nation ("Padre de la Patria Andaluza"), which was reaffirmed in the reformed Statute of Autonomy submitted to popular referendum 18 February 2007. The preamble of the present 2007 Statute of Autonomy says that Article 2 of the present Spanish Constitution of 1978 recognises Andalusia as a nationality. Later, in its articulation, it speaks of Andalusia as a "historic nationality" (Spanish: "nacionalidad histórica"). It also cites the 1919 Andalusianist Manifesto of Córdoba describing Andalusia as a "national reality" ("realidad nacional"), but does not endorse that formulation. Article 1 of the earlier 1981 Statute of Autonomy defined it simply as a "nationality" ("nacionalidad"). The national holiday, the Día de Andalucía, is celebrated on 28 February, commemorating the 1980 autonomy referendum. In spite of this, nationalist groups celebrate the holiday on 4 December, commemorating the 1977 demonstrations to demand autonomy. The honorific title of "Hijo Predilecto de Andalucía" ("Favourite Son of Andalucia") is granted by the Autonomous Government of Andalusia to those whose exceptional merits benefited Andalusia, for work or achievements in natural, social, or political science. It is the highest distinction given by the Autonomous Community of Andalusia. The Sevillian historian Antonio Domínguez Ortiz wrote that: Andalusia has a surface area of , 17.3 percent of the territory of Spain. Andalusia alone is comparable in extent and in the variety of its terrain to any of several of the smaller European countries. To the east is the Mediterranean Sea; to the west the Atlantic Ocean; to the north the Sierra Morena constitutes the border with the Meseta Central; to the south, the self-governing British overseas territory of Gibraltar and the Strait of Gibraltar separate it from Morocco. Andalusia is home to the hottest and driest summers in Spain, but in the west, weather systems sweeping in from the Atlantic ensure that it is relatively wet in the winter, with some areas receiving copious amounts. Contrary to what many people think, as a whole, the region enjoys above-average yearly rainfall in the context of Spain. Andalusia sits at a latitude between 36° and 38° 44' N, in the warm-temperate region. In general, it experiences a hot-summer Mediterranean climate, with dry summers influenced by the Azores High, but subject to occasional torrential rains and extremely hot temperatures. In the winter, the tropical anticyclones move south, allowing cold polar fronts to penetrate the region. Still, within Andalusia there is considerable climatic variety. From the extensive coastal plains one may pass to the valley of the Guadalquivir, barely above sea level, then to the highest altitudes in the Iberian peninsula in the peaks of the Sierra Nevada. In a mere one can pass from the subtropical coast of the province of Granada to the snowy peaks of Mulhacén. Andalusia also includes both the dry Tabernas Desert in the province of Almería and the Sierra de Grazalema Natural Park in the province of Cádiz, which experiences Spain's greatest rainfall. Annual rainfall in the Sierra de Grazalema has been measured as high as in 1963, the highest ever recorded for any location in Iberia. Andalusia is also home to the driest place in continental Europe, the Cabo de Gata, with only of rain per year. In general, as one goes from west to east, away from the Atlantic, there is less precipitation. "Wet Andalusia" includes most of the highest points in the region, above all the Sierra de Grazalema but also the Serranía de Ronda in western Málaga. The valley of the Guadalquivir has moderate rainfall. The Tabernas Desert in Almería, Europe's only true desert, has less than 75 days with any measurable precipitation, and some particular places in the desert have as few as 50 such days. Much of "dry Andalusia" has more than 300 "sunny" days a year. The average temperature in Andalusia throughout the year is over . Averages in the cities range from in Baeza to in Almería. Much of the Guadalquivir valley and the Mediterranean coast has an average of about . The coldest month is January when Granada at the foot of the Sierra Nevada experiences an average temperature of . The hottest are July and August, with an average temperature of for Andalusia as a whole. Córdoba is the hottest provincial capital, followed by Seville. The Guadalquivir valley has experienced some of the highest temperatures recorded in Europe, with a maximum of recorded at Córdoba and Seville. The mountains of Granada and Jaén have the coldest temperatures in southern Iberia, but do not reach continental extremes (and, indeed are surpassed by some mountains in northern Spain). In the cold snap of January 2005, Santiago de la Espada (Jaén) experienced a temperature of and the ski resort at Sierra Nevada National Park—the southernmost ski resort in Europe—dropped to . Sierra Nevada Natural Park has Iberia's lowest average annual temperature, ( at Pradollano) and its peaks remain snowy practically year-round. Mountain ranges affect climate, the network of rivers, soils and their erosion, bioregions, and even human economies insofar as they rely on natural resources. The Andalusian terrain offers a range of altitudes and slopes. Andalusia has the Iberian peninsula's highest mountains and nearly 15 percent of its terrain over . The picture is similar for areas under (with the Baetic Depression), and for the variety of slopes. The Atlantic coast is overwhelmingly beach and gradually sloping coasts; the Mediterranean coast has many cliffs, above all in the Malagan Axarquía and in Granada and Almería. This asymmetry divides the region naturally into Upper Andalusia (two mountainous areas) and Lower Andalusia (the broad basin of the Guadalquivir). The Sierra Morena separates Andalusia from the plains of Extremadura and Castile–La Mancha on Spain's Meseta Central. Although sparsely populated, this is not a particularly high range, and its highest point, the peak of La Bañuela in the Sierra Madrona, lies outside of Andalusia. Within the Sierra Morena, the gorge of Despeñaperros forms a natural frontier between Castile and Andalusia. The Baetic Cordillera consists of the parallel mountain ranges of the Cordillera Penibética near the Mediterranean coast and the Cordillera Subbética inland, separated by the Surco Intrabético. The Cordillera Subbética is quite discontinuous, offering many passes that facilitate transportation, but the Penibético forms a strong barrier between the Mediterranean coast and the interior. The Sierra Nevada, part of the Cordillera Penibética in the Province of Granada, has the highest peaks in Iberia: El Mulhacén at and El Veleta at . Lower Andalusia, the Baetic Depression, the basin of the Guadalquivir, lies between these two mountainous areas. It is a nearly flat territory, open to the Gulf of Cádiz in the southeast. Throughout history, this has been the most populous part of Andalusia. Andalusia has rivers that flow into both the Atlantic and the Mediterranean. Flowing to the Atlantic are the Guadiana, Odiel-Tinto, Guadalquivir, Guadalete, and Barbate. Flowing to the Mediterranean are the Guadiaro, Guadalhorce, Guadalmedina, Guadalfeo, Andarax (also known as the Almería) and Almanzora. Of these, the Guadalquivir is the longest in Andalusia and fifth longest on the Iberian peninsula, at . The rivers of the Atlantic basin are characteristically long, run through mostly flat terrain, and have broad river valleys. As a result, at their mouths are estuaries and wetlands, such as the marshes of Doñana in the delta of the Guadalquivir, and wetlands of the Odiel. In contrast, the rivers of the Mediterranean Basin are shorter, more seasonal, and make a precipitous descent from the mountains of the Baetic Cordillera. Their estuaries are small, and their valleys are less suitable for agriculture. Also, being in the rain shadow of the Baetic Cordillera means that they receive a lesser volume of water. The following hydrographic basins can be distinguished in Andalusia. On the Atlantic side are the Guadalquivir basin; the Andalusian Atlantic Basin with the sub-basins Guadalete-Barbate and Tinto-Odiel; and the Guadiana basin. On the Mediterranean side is the Andalusian Mediterranean Basin and the upper portion of the basin of the Segura. The soils of Andalusia can be divided into three large areas: the Sierra Morena, Cordillera Subbética, and the Baetic Depression and the Surco Intrabético. The Sierra Morena, due to its morphology and the acidic content of its rocks, developed principally relatively poor, shallow soils, suitable only for forests. In the valleys and in some areas where limestone is present, deeper soils allowed farming of cereals suitable for livestock. The more complicated morphology of the Baetic Cordillera makes it more heterogeneous, with the most heterogeneous soils in Andalusia. Very roughly, in contrast to the Sierra Morena, a predominance of basic (alkaline) materials in the Cordillera Subbética, combined with a hilly landscape, generates deeper soils with greater agricultural capacity, suitable to the cultivation of olives. Finally, the Baetic Depression and the Surco Intrabético have deep, rich soils, with great agricultural capacity. In particular, the alluvial soils of the Guadalquivir valley and plain of Granada have a loamy texture and are particularly suitable for intensive irrigated crops. In the hilly areas of the countryside, there is a double dynamic: the depressions have filled with older lime-rich material, developing the deep, rich, dark clay soils the Spanish call "bujeo", or "tierras negras andaluzas", excellent for dryland farming. In other zones, the whiter "albariza" provides an excellent soil for vineyards. Despite their marginal quality, the poorly consolidated soils of the sandy coastline of Huelva and Almería have been successfully used in recent decades for hothouse cultivation under clear plastic of strawberries, raspberries, blueberries, and other fruits. Biogeographically, Andalusia forms part of the Western Mediterranean subregion of the Mediterranean Basin, which falls within the Boreal Kingdom. Five floristic provinces lie, in whole or in part, within Andalusia: along much of the Atlantic coast, the Lusitanian-Andalusian littoral or Andalusian Atlantic littoral; in the north, the southern portion of the Luso-Extremaduran floristic province; covering roughly half of the region, the Baetic floristic province; and in the extreme east, the Almerian portion of the Almerian-Murcian floristic province and (coinciding roughly with the upper Segura basin) a small portion of the Castilian-Maestrazgan-Manchegan floristic province. These names derive primarily from past or present political geography: "Luso" and "Lusitanian" from Lusitania, one of three Roman provinces in Iberia, most of the others from present-day Spanish provinces, and Maestrazgo being a historical region of northern Valencia. In broad terms, the typical vegetation of Andalusia is Mediterranean woodland, characterized by leafy xerophilic perennials, adapted to the long, dry summers. The dominant species of the climax community is the holly oak ("Quercus ilex"). Also abundant are cork oak ("Quercus suber"), various pines, and Spanish fir ("Abies pinsapo"). Due to cultivation, olive ("Olea europaea") and almond ("Prunus dulcis") trees also abound. The dominant understory is composed of thorny and aromatic woody species, such as rosemary ("Rosmarinus officinalis"), thyme ("Thymus"), and "Cistus". In the wettest areas with acidic soils, the most abundant species are the oak and cork oak, and the cultivated "Eucalyptus". In the woodlands, leafy hardwoods of genus "Populus" (poplars, aspens, cottonwoods) and "Ulmus" (elms) are also abundant; poplars are cultivated in the plains of Granada. The Andalusian woodlands have been much altered by human settlement, the use of nearly all of the best land for farming, and frequent wildfires. The degraded forests become shrubby and combustible garrigue. Extensive areas have been planted with non-climax trees such as pines. There is now a clear conservation policy for the remaining forests, which survive almost exclusively in the mountains. The biodiversity of Andalusia extends to its fauna as well. More than 400 of the 630 vertebrate species extant in Spain can be found in Andalusia. Spanning the Mediterranean and Atlantic basins, and adjacent to the Strait of Gibraltar, Andalusia is on the migratory route of many of the numerous flocks of birds that travel annually from Europe to Africa and back. The Andalusian wetlands host a rich variety of birds. Some are of African origin, such as the red-knobbed coot ("Fulica cristata"), the purple swamphen ("Porphyrio porphyrio"), and the greater flamingo ("Phoenicopterus roseus"). Others originate in Northern Europe, such as the greylag goose ("Anser anser"). Birds of prey (raptors) include the Spanish imperial eagle ("Aquila adalberti"), the griffon vulture ("Gyps fulvus"), and both the black and red kite ("Milvus migrans" and "Milvus milvus"). Among the herbivores, are several deer (Cervidae) species, notably the fallow deer ("Dama dama") and roe deer ("Capreolus capreolus"); the European mouflon ("Ovis orientalis musimon"), a type of sheep; and the Spanish ibex ("Capra pyrenaica", which despite its scientific name is no longer found in the Pyrenees). The Spanish ibex has recently been losing ground to the Barbary sheep ("Ammotragus lervia"), an invasive species from Africa, introduced for hunting in the 1970s. Among the small herbivores are rabbits—especially the European rabbit ("Oryctolagus cuniculus")—which form the most important part of the diet of the carnivorous species of the Mediterranean woodlands. The large carnivores such as the Iberian wolf ("Canis lupus signatus") and the Iberian lynx ("Lynx pardinus") are quite threatened, and are limited to the Sierra de Andújar, inside of Sierra Morena, Doñana and Despeñaperros. Stocks of the wild boar ("Sus scrofa"), on the other hand, have been well preserved because they are popular with hunters. More abundant and in varied situations of conservation are such smaller carnivores as otters, dogs, foxes, the European badger ("Meles meles"), the European polecat ("Mustela putorius"), the least weasel ("Mustela nivalis"), the wildcat ("Felis silvestris"), the common genet ("Genetta genetta"), and the Egyptian mongoose ("Herpestes ichneumon)". Other notable species are "Acherontia atropos" (a variety of death's-head hawkmoth), "Vipera latasti" (a venomous snake), and the endemic (and endangered) fish "Aphanius baeticus". Andalusia has many unique ecosystems. In order to preserve these areas in a manner compatible with both conservation and economic exploitation, many of the most representative ecosystems have been given protected status. The various levels of protection are encompassed within the Network of Protected Natural Spaces of Andalusia (Red de Espacios Naturales Protegidos de Andalucía, RENPA) which integrates all protected natural spaces located in Andalusia, whether they are protected at the level of the local community, the autonomous community of Andalusia, the Spanish state, or by international conventions. RENPA consists of 150 protected spaces, consisting of two national parks, 24 natural parks, 21 periurban parks (on the fringes of cities or towns), 32 natural sites, two protected countrysides, 37 natural monuments, 28 nature reserves, and four concerted nature reserves (in which a government agency coordinates with the owner of the property for its management), all part of the European Union's Natura 2000 network. Under the international ambit are the nine Biosphere Reserves, 20 Ramsar wetland sites, four Specially Protected Areas of Mediterranean Importance and two UNESCO Geoparks. In total, nearly 20 percent of the territory of Andalusia lies in one of these protected areas, which constitute roughly 30 percent of the protected territory of Spain. Among these many spaces, some of the most notable are the Sierras de Cazorla, Segura y Las Villas Natural Park, Spain's largest natural park and the second largest in Europe, the Sierra Nevada National Park, Doñana National Park and Natural Park, the Tabernas Desert, and the Cabo de Gata-Níjar Natural Park, the largest terrestrial-maritime reserve in the European Western Mediterranean Sea. The geostrategic position of Andalusia in the extreme south of Europe, providing (along with Morocco) a gateway between Europe and Africa, added to its position between the Atlantic Ocean and the Mediterranean Sea, as well as its rich deposits of minerals and its agricultural wealth, have made Andalusia a tempting prize for civilizations since prehistoric times. Add to this its area of (larger than many European countries), and it can be no surprise that Andalusia has figured prominently in the history of Europe and the Mediterranean. Several theories postulate that the first hominids in Europe were in Andalusia, having passed across the Strait of Gibraltar; the earliest known paintings of humanity have been found in the Caves of Nerja, Málaga. The first settlers, based on artifacts from the archaeological sites at Los Millares, El Argar, and Tartessos, were clearly influenced by cultures of the Eastern Mediterranean who arrived on the Andalusian coast. Andalusia then went through a period of protohistory, when the region did not have a written language of its own, but its existence was known to and documented by literate cultures, principally the Phoenicians and Ancient Greeks, wide historical moment in which Cádiz was founded, regarded by many as the oldest city still standing in Western Europe; another city among the oldest is Málaga. During the second millennium BCE, the kingdom of Tartessos developed in Andalusia. With the fall of the original Phoenician cities in the East, Carthage — itself the most significant Phoenician colony — became the dominant sea power of the western Mediterranean and the most important trading partner for the Phoenician towns along the Andalusian coast. Between the First and Second Punic Wars, Carthage extended its control beyond Andalucia to include all of Iberia except the Basque Country. Andalusia was the major staging ground for the war with Rome led by the Carthaginian general Hannibal. The Romans defeated the Carthaginians and conquered Andalusia, the region being renamed Baetica. It was fully incorporated into the Roman Empire, and from this region came many Roman magistrates and senators, as well as the emperors Trajan and (most likely) Hadrian. The Vandals moved briefly through the region during the 5th century AD before settling in North Africa, after which the region fell into the hands of the Visigothic Kingdom. The Visigoths in this region were practically independent of the Visigothic Catholic Kingdom of Toledo. This is the era of Saints Isidore of Seville and Hermenegild. During this period, around 555 AD, the Eastern Roman Empire conquered Andalusia under Justinian I, the Eastern Roman Emperor. They established Spania, a province of the Byzantine Empire from 552 until 624. Though their holdings were quickly reduced, they continued to have interests in the region until it was lost altogether in 624. The Visigothic era came to an abrupt end in 711 with the Umayyad conquest of Hispania by the Muslim Umayyad general Tariq ibn Ziyad. Tariq is known in Umayyad history and legend as a formidable conqueror who dared and bore the nerve to burn his fleet of ships, when he landed with his troops on the coast of Gibraltar – an acronym of "Jabel alTariq" meaning "the mountain of Tariq". When the Muslim invaders seized control and consolidated their dominion of the region, they remained tolerant of the local faiths, but they also needed a place for their own faith. In the 750s, they forcibly rented half of Cordoba's Cathedral of San Vicente (Visigothic) to use as a mosque. In this period, the name "Al-Andalus" was applied to the Iberian Peninsula, and later it referred to the parts not controlled by the Gothic states in the North. The Muslim rulers in Al-Andalus were economic invaders and interested in collecting taxes; social changes imposed on the native populace were mainly confined to geographical, political and legal conveniences. Al-Andalus remained connected to other states under Muslim rule; also trade routes between it and Constantinople and Alexandria remained open, while many cultural features of the Roman Empire were transmitted throughout Europe and the Near East by its successor state, the Byzantine Empire. Byzantine architecture is an example of such cultural diffusion continuing even after the collapse of the empire. Nevertheless, the Guadalquivir River valley became the point of power projection in the peninsula with the Caliphate of Cordoba making Cordoba its capital. The Umayyad Caliphate produced such leaders as Caliph Abd-ar-Rahman III (ruled 912–961) and his son, Caliph Al-Hakam II (ruled 961–976) and built the magnificent Great Mosque of Cordoba. Under these rulers, Cordoba was the center of economic and cultural significance. By the 10th century, the northern Kingdoms of Spain and other European Crowns had begun what would eventually become the Reconquista: the reconquest of the Iberian Peninsula for Christendom. Caliph Abd-ar-Rahman suffered some minor military defeats, but often managed to manipulate the Gothic northern kingdoms to act against each other's interests. Al-Hakam achieved military successes, but at the expense of uniting the north against him. After the conquest of Toledo in 1086 by Alfonso VI, the Crown of Castille and the Crown of Aragon dominated large parts of the peninsula. The main "Taifas" therefore had to resort to assistance from various other powers across the Mediterranean. A number of different Muslim dynasties of North African origin—notably Almoravid dynasty and Almohad dynasty—dominated a slowly diminishing Al-Andalus over the next several centuries. After the victory at the Battle of Sagrajas (1086) put a temporary stop to Castile expansion, the Almoravid dynasty reunified Al-Andalus with its capital in Granada, ruling until the mid-12th century. The various "Taifa" kingdoms were assimilated. the Almohad dynasty expansion in North Africa weakened Al-Andalus, and in 1170 the Almohads transferred their capital from Marrakesh to Seville. The victory at the Battle of Las Navas de Tolosa (1212) marked the beginning of the end of the Almohad dynasty. The weakness caused by the collapse of Almohad power and the subsequent creation of new "Taifas", each with its own ruler, led to the rapid Castile reconquest of the valley of the Guadalquivir. Córdoba was regained in 1236 and Seville in 1248. The fall of Granada on 2 January 1492 put an end to the Nasrid rule, event that marks the beginning of Andalusia, the southern four territories of the Crown of Castile in the Iberian Peninsula. Seven months later, on 3 August 1492 Christopher Columbus left the town of Palos de la Frontera, Huelva, with the first expedition that resulted in the Discovery of the Americas, that would end the Middle Ages and signal the beginning of modernity. Many Castilians participated in this and other expeditions that followed, some of them known as the Minor or Andalusian Journeys. Contacts between Spain and the Americas, including royal administration and the shipping trade from Asia and America for over three hundred years, came almost exclusively through the south of Spain, specially Seville and Cadiz ports. As a result, it became the wealthiest, most influential region in Spain and amongst the most influential in Europe. For example, the Habsburg diverted much of this trade wealth to control its European territories. In the first half of the 16th century plague was still prevalent in Spain. According to George C. Kohn, "One of the worst epidemics of the century, whose miseries were accompanied by severe drought and food shortage, started in 1505; by 1507, about 100,000 people had died in Andalusia alone. Andalusia was struck once again in 1646. For three years, plague haunted the entire region, causing perhaps as many as 200,000 deaths, especially in Málaga and Seville." Following the Second Rebellion of the Alpujarras in 1568–1571, the Moorish population—that is, unconverted Moriscos—were expelled from Kingdom of Castile (and Aragon). However, by order of the Spanish crown, two Moorish families were required to remain in each village in order to demonstrate to the new inhabitants, introduced from northern Spain, the workings of the terracing and irrigation systems on which the district's agriculture depends. In 1810–12 Spanish troops strongly resisted the French occupation during the Peninsular War (part of the Napoleonic Wars). Andalusia profited from the Spanish overseas empire, although much trade and finance eventually came to be controlled by other parts of Europe to where it was ultimately destined. In the 18th century, commerce from other parts of Spain began to displace Andalusian commerce when the Spanish government ended Andalusia's trading monopoly with the colonies in the Americas. The loss of the empire in the 1820s hurt the economy of the region, particularly the cities that had benefited from the trade and ship building. The construction of railways in the latter part of the 19th century enabled Andalusia to better develop its agricultural potential and it became an exporter of food. While industrialisation was taking off in the northern Spanish regions of Catalonia and the Basque country, Andalusia remained traditional and displayed a deep social division between a small class of wealthy landowners and a population made up largely of poor agricultural labourers and tradesmen. Andalusia was one of the worst affected regions of Spain by Francisco Franco's brutal campaign of mass-murder and political suppression called the White Terror during and after the Spanish Civil War. The Nationalist rebels bombed and seized the working-class districts of the main Andalusian cities in the first days of the war, and afterwards went on to execute thousands of workers and militants of the leftist parties: in the city of Cordoba 4,000; in the city of Granada 5,000; in the city of Seville 3,028; and in the city of Huelva 2,000 killed and 2,500 disappeared. The city of Málaga, occupied by the Nationalists in February 1937 following the Battle of Málaga, experienced one of the harshest repressions following Francoist victory with an estimated total of 17,000 people summarily executed. Carlos Arias Navarro, then a young lawyer who as public prosecutor signed thousands of execution warrants in the trials set up by the triumphant rightists, became known as "The Butcher of Málaga" ("Carnicero de Málaga"). Paul Preston estimates the total number of victims of deliberately killed by the Nationalists in Andalusia at 55,000. Andalusia is one of the 17 autonomous communities of Spain. The Regional Government of Andalusia (Spanish: "Junta de Andalucía") includes the Parliament of Andalusia, its chosen president, a Consultative Council, and other bodies. The Autonomous Community of Andalusia was formed in accord with a referendum of 28 February 1980 and became an autonomous community under the 1981 Statute of Autonomy known as the "Estatuto de Carmona". The process followed the Spanish Constitution of 1978, still current as of 2009, which recognizes and guarantees the right of autonomy for the various regions and nationalities of Spain. The process to establish Andalusia as an autonomous region followed Article 151 of the Constitution, making Andalusia the only autonomous community to take that particular course. That article was set out for regions like Andalusia that had been prevented by the outbreak of the Spanish Civil War from adopting a statute of autonomy during the period of the Second Spanish Republic. Article 1 of the 1981 Statute of Autonomy justifies autonomy based on the region's "historical identity, on the self-government that the Constitution permits every nationality, on outright equality to the rest of the nationalities and regions that compose Spain, and with a power that emanates from the Andalusian Constitution and people, reflected in its Statute of Autonomy". In October 2006 the constitutional commission of the Cortes Generales (the national legislature of Spain), with favorable votes from the left-of-center Spanish Socialist Workers' Party (PSOE), the leftist United Left (IU) and the right-of-center People's Party (PP), approved a new Statute of Autonomy for Andalusia, whose preamble refers to the community as a "national reality" ("realidad nacional"): On 2 November 2006 the Spanish Chamber Deputies ratified the text of the Constitutional Commission with 306 votes in favor, none opposed, and 2 abstentions. This was the first time a Spanish Organic Law adopting a Statute of Autonomy was approved with no opposing votes. The Senate, in a plenary session of 20 December 2006, ratified the referendum to be voted upon by the Andalusian public 18 February 2007. The Statute of Autonomy spells out Andalusia's distinct institutions of government and administration. Chief among these is the Andalusian Autonomous Government ("Junta de Andalucía"). Other institutions specified in the Statute are the Defensor del Pueblo Andaluz (literally "Defender of the Andalusian People", basically an ombudsperson), the Consultative Council, the Chamber of Accounts, the Audiovisual Council of Andalusia, and the Economic and Social Council. The Andalusian Statute of Autonomy recognizes Seville as the region's capital. The Andalusian Autonomous Government is located there. However, the region's highest court, the High Court of Andalusia ("Tribunal Superior de Justicia de Andalucía") is not part of the Autonomous Government, and has its seat in Granada. The Andalusian Autonomous Government ("Junta de Andalucía") is the institution of self-government of the Autonomous Community of Andalusia. Within the government, the President of the Regional Government of Andalusia is the supreme representative of the autonomous community, and the ordinary representative of the Spanish state in the autonomous community. The president is formally named to the position by the Monarch of Spain and then confirmed by a majority vote of the Parliament of Andalusia. In practice, the monarch always names a person acceptable to the ruling party or coalition of parties in the autonomous region. In theory, were the candidate to fail to gain the needed majority, the monarch could propose a succession of candidates. After two months, if no proposed candidate could gain the parliament's approval, the parliament would automatically be dissolved and the acting president would call new elections. On 5 September 2013 Susana Díaz was elected president. The Council of Government, the highest political and administrative organ of the Community, exercises regulatory and executive power. The President presides over the council, which also includes the heads of various departments ("Consejerías"). In the current legislature (2008–2012), there are 15 of these departments. In order of precedence, they are Presidency, Governance, Economy and Treasury, Education, Justice and Public Administration, Innovation, Science and Business, Public Works and Transportation, Employment, Health, Agriculture and Fishing, Housing and Territorial Planning, Tourism, Commerce and Sports, Equality and Social Welfare, Culture, and Environment. The Parliament of Andalusia, its Autonomic Legislative Assembly, develops and approves laws and elects and removes the President. Elections to the Andalusian Parliament follow a democratic formula through which the citizens elect 109 representatives. After the approval of the Statute of Autonomy through Organic Law 6/1981 on 20 December 1981, the first elections to the autonomic parliament took place 23 May 1982. Further elections have occurred in 1986, 1990, 1994, 1996, 2000, 2004, and 2008. The current (2008–2012) legislature includes representatives of the PSOE-A (Andalusian branch of the left-of-center PSOE), PP-A (Andalusian branch of the right-of-center PP) and IULV-CA (Andalusian branch of the leftist IU). The High Court of Andalusia ("Tribunal Superior de Justicia de Andalucía") in Granada is subject only to the higher jurisdiction of the Supreme Court of Spain. The High Court is not an organ of the Autonomous Community, but rather of the Judiciary of Spain, which is unitary throughout the kingdom and whose powers are not transferred to the autonomous communities. The Andalusian territory is divided into 88 legal/judicial districts ("partidos judiciales"). Andalusia consists of eight provinces. The latter were established by Javier de Burgos in the 1833 territorial division of Spain. Each of the Andalusian provinces bears the same name as its capital: Andalusia is traditionally divided into two historical subregions: Upper Andalusia" or "Eastern Andalusia ("Andalucía Oriental"), consisting of the provinces of Almería, Granada, Jaén, and Málaga, and Lower Andalusia" or "Western Andalusia ("Andalucía Occidental"), consisting of the provinces of Cádiz, Córdoba, Huelva and Seville. Within the various autonomous communities of Spain, "comarcas" are comparable to shires (or, in some countries, counties) in the English-speaking world. Unlike in some of Spain's other autonomous communities, under the original 1981 Statute of Autonomy, the "comarcas" of Andalusia had no formal recognition, but, in practice, they still had informal recognition as geographic, cultural, historical, or in some cases administrative entities. The 2007 Statute of Autonomy echoes this practice, and mentions "comarcas" in Article 97 of Title III, which defines the significance of "comarcas" and establishes a basis for formal recognition in future legislation. The current statutory entity that most closely resembles a "comarca" is the "mancomunidad", a freely chosen, bottom-up association of municipalities intended as an instrument of socioeconomic development and coordination between municipal governments in specific areas. Beyond the level of provinces, Andalusia is further divided into 774 municipalities ("municipios"). The municipalities of Andalusia are regulated by Title III of the Statute of Autonomy, Articles 91–95, which establishes the municipality as the basic territorial entity of Andalusia, each of which has legal personhood and autonomy in many aspects of its internal affairs. At the municipal level, representation, government and administration is performed by the "ayuntamiento" (municipal government), which has competency for urban planning, community social services, supply and treatment of water, collection and treatment of waste, and promotion of tourism, culture, and sports, among other matters established by law. Among the more important Andalusian cities besides the provincial capitals are: In conformity with the intent to devolve control as locally as possible, in many cases, separate nuclei of population within municipal borders each administer their own interests. These are variously known as "pedanías" ("hamlets"), "villas" ("villages"), "aldeas" (also usually rendered as "villages"), or other similar names. Andalusia ranks first by population among the 17 autonomous communities of Spain. The estimated population at the beginning of 2009 was 8,285,692. The population is concentrated, above all, in the provincial capitals and along the coasts, so that the level of urbanization is quite high; half the population is concentrated in the 28 cities of more than 50,000 inhabitants. The population is aging, although the process of immigration is countering the inversion of the population pyramid. At the end of the 20th century, Andalusia was in the last phase of demographic transition. The death rate stagnated at around 8–9 per thousand, and the population came to be influenced mainly by birth and migration. In 1950, Andalusia had 20.04 percent of the national population of Spain. By 1981, this had declined to 17.09 percent. Although the Andalusian population was not declining in absolute terms, these relative losses were due to emigration great enough to nearly counterbalance having the highest birth rate in Spain. Since the 1980s, this process has reversed on all counts, and as of 2009, Andalusia has 17.82 percent of the Spanish population. The birth rate is sharply down, as is typical in developed economies, although it has lagged behind much of the rest of the world in this respect. Furthermore, prior emigrants have been returning to Andalusia. Beginning in the 1990s, others have been immigrating in large numbers as well, as Spain has become a country of net immigration. At the beginning of the 21st century, statistics show a slight increase in the birth rate, due in large part to the higher birth rate among immigrants. The result is that as of 2009, the trend toward rejuvenation of the population is among the strongest of any autonomous community of Spain, or of any comparable region in Europe. At the beginning of the 21st century, the population structure of Andalusia shows a clear inversion of the population pyramid, with the largest cohorts falling between ages 25 and 50. Comparison of the population pyramid in 2008 to that in 1986 shows: As far as composition by sex, two aspects stand out: the higher percentage of women in the elderly population, owing to women's longer life expectancy, and, on the other hand, the higher percentage of men of working age, due in large part to a predominantly male immigrant population. In 2005, 5.35 percent of the population of Andalusia were born outside of Spain. This is a relatively low number for a Spanish region, the national average being three percentage points higher. The immigrants are not evenly distributed among the Andalusian provinces: Almería, with a 15.20 percent immigrant population, is third among all provinces in Spain, while at the other extreme Jaén is only 2.07 percent immigrants and Córdoba 1.77 percent. The predominant nationalities among the immigrant populations are Moroccan (92,500, constituting 17.79 percent of the foreigners living in Andalusia) and British (15.25 percent across the region). When comparing world regions rather than individual countries, the single largest immigrant block is from the region of Latin America, outnumbering not only all North Africans, but also all non-Spanish Western Europeans. Demographically, this group has provided an important addition to the Andalusian labor force. Andalusia is traditionally an agricultural area, but the service sector (particularly tourism, retail sales, and transportation) now predominates. The once booming construction sector, hit hard by the 2009 recession, was also important to the region's economy. The industrial sector is less developed than most other regions in Spain. Between 2000–2006 economic growth per annum was 3.72%, one of the highest in the country. Still, according to the Spanish Instituto Nacional de Estadística (INE), the GDP per capita of Andalusia (€17,401; 2006) remains the second lowest in Spain, with only Extremadura lagging behind. The Gross domestic product (GDP) of the autonomous community was 160.6 billion euros in 2018, accounting for 13.4% of Spanish economic output. GDP per capita adjusted for purchasing power was 20,500 euros or 68% of the EU27 average in the same year. The primary sector, despite adding the least of the three sectors to the regional GDP remains important, especially when compared to typical developed economies. The primary sector produces 8.26 percent of regional GDP, 6.4 percent of its GVA and employs 8.19 percent of the workforce. In monetary terms it could be considered a rather uncompetitive sector, given its level of productivity compared to other Spanish regions. In addition to its numeric importance relative to other regions, agriculture and other primary sector activities have strong roots in local culture and identity. The primary sector is divided into a number of subsectors: agriculture, commercial fishing, animal husbandry, hunting, forestry, mining, and energy. For many centuries, agriculture dominated Andalusian society, and, with 44.3 percent of its territory cultivated and 8.4 percent of its workforce in agriculture as of 2016, it remains an integral part of Andalusia's economy. However, its importance is declining, as is that of the primary and secondary sectors generally, as the service sector is increasingly taking over. The primary cultivation is dryland farming of cereals and sunflowers without artificial irrigation, especially in the vast countryside of the Guadalquivir valley and the high plains of Granada and Almería-with a considerably lesser and more geographically focused cultivation of barley and oats. Using irrigation, maize, cotton and rice are also grown on the banks of the Guadalquivir and Genil. The most important tree crops are olives, especially in the Subbetic regions of the provinces of Córdoba and Jáen, where irrigated olive orchards constitute a large component of agricultural output. There are extensive vineyards in various zones such as Jerez de la Frontera (sherry), Condado de Huelva, Montilla-Moriles and Málaga. Fruits—mainly citrus fruits—are grown near the banks of the Guadalquivir; almonds, which require far less water, are grown on the high plains of Granada and Almería. In monetary terms, by far the most productive and competitive agriculture in Andalusia is the intensive forced cultivation of strawberries, raspberries, blueberries, and other fruits grown under hothouse conditions under clear plastic, often in sandy zones, on the coasts, in Almería and Huelva. Organic farming has recently undergone rapid expansion in Andalusia, mainly for export to European markets but with increasing demand developing in Spain. Andalusia has a long tradition of animal husbandry and livestock farming, but it is now restricted mainly to mountain meadows, where there is less pressure from other potential uses. Andalusians have a long and colourful history of dog breeding that can be observed throughout the region today. The raising of livestock now plays a semi-marginal role in the Andalusian economy, constituting only 15 percent of the primary sector, half the number for Spain taken as a whole. "Extensive" raising of livestock grazes the animals on natural or cultivated pastures, whereas "intensive" raising of livestock is based in fodder rather than pasture. Although the productivity is higher than with extensive techniques, the economics are quite different. While intensive techniques now dominate in Europe and even in other regions of Spain, most of Andalusia's cattle, virtually all of its sheep and goats, and a good portion of its pigs are raised by extensive farming in mountain pastures. This includes the Black Iberian pigs that are the source of "Jamón ibérico". Andalusia's native sheep and goats present a great economic opportunity in a Europe where animal products are generally in strong supply, but the sheep and goat meat, milk, and leather (and the products derived from these) are relatively scarce. Dogs are bred not just as companion animals, but also as herding animals used by goat and sheep herders. Hunting remains relatively important in Andalusia, but has largely lost its character as a means of obtaining food. It is now more of a leisure activity linked to the mountain areas and complementary to forestry and the raising of livestock. Dogs are frequently used as hunting companions to retrieve killed game. The Andalusian forests are important for their extent—50 percent of the territory of Andalusia—and for other less quantifiable environmental reasons, such as their value in preventing erosion, regulating the flow of water necessary for other flora and fauna. For these reasons, there is legislation in place to protect the Andalusian forests. The value of forest products as such constitutes only 2 percent of agricultural production. This comes mostly from cultivated species—eucalyptus in Huelva and poplar in Granada—as well as naturally occurring cork oak in the Sierra Morena. Fishing is a longstanding tradition on the Andalusian coasts. Fish and other seafood have long figured prominently in the local diet and in the local gastronomic culture: fried fish ("pescaito frito" in local dialect), white prawns, "almadraba" tuna, among others. The Andalusian fishing fleet is Spain's second largest, after Galicia, and Andalusia's 38 fishing ports are the most of any Spanish autonomous community. Commercial fishing produces only 0.5 percent of the product of the regional primary sector by value, but there are areas where it has far greater importance. In the province of Huelva it constitutes 20 percent of the primary sector, and locally in Punta Umbría 70 percent of the work force is involved in commercial fishing. Failure to comply with fisheries laws regarding the use of trawling, urban pollution of the seacoast, destruction of habitats by coastal construction (for example, alteration of the mouths of rivers, construction of ports), and diminution of fisheries by overexploitation have created a permanent crisis in the Andalusian fisheries, justifying attempts to convert the fishing fleet. The decrease in fish stocks has led to the rise of aquaculture, including fish farming both on the coasts and in the interior. Despite the general poor returns in recent years, mining retains a certain importance in Andalusia. Andalusia produces half of Spain's mining product by value. Of Andalusia's production, roughly half comes from the province of Huelva. Mining for precious metals at Minas de Riotinto in Huelva ("see Rio Tinto Group") dates back to pre-Roman times; the mines were abandoned in the Middle Ages and rediscovered in 1556. Other mining activity is coal mining in the Guadiato valley in the province of Córdoba; various metals at Aznalcóllar in the province of Seville, and iron at Alquife in the province of Granada. In addition, limestone, clay, and other materials used in construction are well distributed throughout Andalusia. The Andalusian industrial sector has always been relatively small. Nevertheless, in 2007, Andalusian industry earned 11.979 million euros and employed more than 290,000 workers. This represented 9.15 percent of regional GDP, far below the 15.08 the secondary sector represents in the economy of Spain as a whole. By analyzing the different subsectors of the food industry Andalusian industry accounts for more than 16% of total production. In a comparison with the Spanish economy, this subsector is virtually the only food that has some weight in the national economy with 16.16%. Lies far behind the manufacturing sector of shipping materials just over 10% of the Spanish economy. Companies like Cruzcampo (Heineken Group), Puleva, Domecq, Santana Motors or Renault-Andalusia, are exponents of these two subsectors. Of note is the Andalusian aeronautical sector, which is second nationally only behind Madrid and represents approximately 21% of total turnover in terms of employment, highlighting companies like Airbus, Airbus Military, or the newly formed Aerospace Alestis. On the contrary it is symptomatic of how little weight the regional economy in such important sectors such as textiles or electronics at the national level. Andalusian industry is also characterized by a specialization in industrial activities of transforming raw agricultural and mineral materials. This is largely done by small enterprises without the public or foreign investment more typical of a high level of industrialization. In recent decades the Andalusian tertiary (service) sector has grown greatly, and has come to constitute the majority of the regional economy, as is typical of contemporary economies in developed nations. In 1975 the service sector produced 51.1 percent of local GDP and employed 40.8 percent of the work force. In 2007, this had risen to 67.9 percent of GDP and 66.42 percent of jobs. This process of "tertiarization" of the economy has followed a somewhat unusual course in Andalusia. This growth occurred somewhat earlier than in most developed economies and occurred independently of the local industrial sector. There were two principal reasons that "tertiarization" followed a different course in Andalusia than elsewhere: 1. Andalusian capital found it impossible to compete in the industrial sector against more developed regions, and was obligated to invest in sectors that were easier to enter. 2. The absence of an industrial sector that could absorb displaced agricultural workers and artisans led to the proliferation of services with rather low productivity. This unequal development compared to other regions led to a hypertrophied and unproductive service sector, which has tended to reinforce underdevelopment, because it has not led to large accumulations of capital. Due in part to the relatively mild winter and spring climate, the south of Spain is attractive to overseas visitors–especially tourists from Northern Europe. While inland areas such as Jaén, Córdoba and the hill villages and towns remain relatively untouched by tourism, the coastal areas of Andalusia have heavy visitor traffic for much of the year. Among the autonomous communities, Andalusia is second only to Catalonia in tourism, with nearly 30 million visitors every year. The principal tourist destinations in Andalusia are the Costa del Sol and (secondarily) the Sierra Nevada. As discussed above, Andalusia is one of the sunniest and warmest places in Europe, making it a center of "sun and sand" tourism, but not only it. Around 70 percent of the lodging capacity and 75 percent of the nights booked in Andalusian hotels are in coastal municipalities. The largest number of tourists come in August—13.26 percent of the nights booked throughout the year—and the smallest number in December—5.36 percent. On the west (Atlantic) coast are the Costa de la Luz (provinces of Huelva and Cádiz), and on the east (Mediterranean) coast, the Costa del Sol (provinces of Cádiz y Málaga), Costa Tropical (Granada and part of Almería) and the Costa de Almería. In 2004, the Blue Flag beach program of the non-profit Foundation for Environmental Education recognized 66 Andalusian beaches and 18 pleasure craft ports as being in a good state of conservation in terms of sustainability, accessibility, and quality. Nonetheless, the level of tourism on the Andalusian coasts has been high enough to have a significant environmental impact, and other organizations—such as the Spanish Ecologists in Action ("Ecologistas en Acción") with their description of "Black Flag beaches" or Greenpeace—have expressed the opposite sentiment. However, Hotel chains such as Fuerte Hotels have ensured that sustainability within the tourism industry is one of their highest priorities. Together with "sand and sun" tourism, there has also been a strong increase in nature tourism in the interior, as well as cultural tourism, sport tourism, and conventions. One example of sport and nature tourism is the ski resort at Sierra Nevada National Park. As for cultural tourism, there are hundreds of cultural tourist destinations: cathedrals, castles, forts, monasteries, and historic city centers and a wide variety of museums. It can be highlighted that Spain has seven of its 42 cultural UNESCO World Heritage Sites in Andalucia: Further, there are the "Lugares colombinos", significant places in the life of Christopher Columbus: Palos de la Frontera, La Rábida Monastery, and Moguer) in the province of Huelva. There are also archeological sites of great interest: the Roman city of Italica, birthplace of Emperor Trajan and (most likely) Hadrian or Baelo Claudia near Tarifa. Andalusia was the birthplace of such great painters as Velázquez and Murillo (Seville) and, more recently, Picasso (Málaga); Picasso is memorialized by his native city at the Museo Picasso Málaga and Natal House Foundation; the Casa de Murillo was a house museum 1982–1998, but is now mostly offices for the Andalusian Council of Culture. The CAC Málaga (Museum of Modern Art) is the most visited museum of Andalusia and has offered exhibitions of artists such as Louise Bourgeois, Jake and Dinos Chapman, Gerhard Richter, Anish Kapoor, Ron Mueck or Rodney Graham. Malaga is also located part of the private Carmen Thyssen-Bornemisza Collection at Carmen Thyssen Museum. There are numerous other significant museums around the region, both of paintings and of archeological artifacts such as gold jewelry, pottery and other ceramics, and other works that demonstrate the region's artisanal traditions. The Council of Government has designated the following "Municipios Turísticos": in Almería, Roquetas de Mar; in Cádiz, Chiclana de la Frontera, Chipiona, Conil de la Frontera, Grazalema, Rota, and Tarifa; in Granada, Almuñécar; in Huelva, Aracena; in Jaén, Cazorla; in Málaga, Benalmádena, Fuengirola, Nerja, Rincón de la Victoria, Ronda, and Torremolinos; in Seville, Santiponce. The unemployment rate stood at 25.5% in 2017 and was one of the highest in Spain and Europe. As in any modern society, transport systems are an essential structural element of the functioning of Andalusia. The transportation network facilitates territorial coordination, economic development and distribution, and intercity transportation. In urban transport, underdeveloped public transport systems put pedestrian traffic and other non-motorized traffic are at a disadvantage compared to the use of private vehicles. Several Andalusian capitals—Córdoba, Granada and Seville—have recently been trying to remedy this by strengthening their public transport systems and providing a better infrastructure for the use of bicycles. There are now three rapid transit systems operating in Andalucia – the Seville Metro, Málaga Metro and Granada Metro. Cercanías commuter rail networks operate in Seville, Málaga and Cádiz. For over a century, the conventional rail network has been centralized on the regional capital, Seville, and the national capital, Madrid; in general, there are no direct connections between provincial capitals. High-speed AVE trains run from Madrid via Córdoba to Seville and Málaga, from which a branch from Antequera to Granada opened in 2019. Further AVE routes are under construction. The Madrid-Córdoba-Seville route was the first high-velocity route in Spain (operating since 1992). Other principal routes are the one from Algeciras to Seville and from Almería via Granada to Madrid. Most of the principal roads have been converted into limited access highways known as "autovías". The Autovía del Este (Autovía A-4) runs from Madrid through the Despeñaperros Natural Park, then via Bailén, Córdoba, and Seville to Cádiz, and is part of European route E05 in the International E-road network. The other main road in the region is the portion of European route E15, which runs as the Autovia del Mediterráneo along the Spanish Mediterranean coast. Parts of this constitute the superhighway Autopista AP-7, while in other areas it is Autovía A-7. Both of these roads run generally east–west, although the Autovía A-4 turns to the south in western Andalusia. Other first-order roads include the Autovía A-48 roughly along the Atlantic coast from Cádiz to Algeciras, continuing European route E05 to meet up with European route E15; the Autovía del Quinto Centenario (Autovía A-49), which continues west from Seville (where the Autovía A-4 turns toward the south) and goes on to Huelva and into Portugal as European route E01; the Autovía Ruta de la Plata (Autovía A-66), European route E803, which roughly corresponds to the ancient Roman 'Silver Route' from the mines of northern Spain, and runs north from Seville; the Autovía de Málaga (Autovía A-45), which runs south from Córdoba to Málaga; and the Autovía de Sierra Nevada (Autovía A-44), part of European route E902, which runs south from Jaén to the Mediterranean coast at Motril. As of 2008 Andalusia has six public airports, all of which can legally handle international flights; however the Málaga Airport is dominant, handling 60.67 percent of passengers and 85 percent of its international traffic. The Seville Airport handles another 20.12 percent of traffic, and the Jerez Airport 7.17 percent, so that these three airports account for 87.96 percent of traffic. Málaga Airport is the international airport that offers a wide variety of international destinations. It has a daily link with twenty cities in Spain and over a hundred cities in Europe (mainly in Great Britain, Central Europe and the Nordic countries but also the main cities of Eastern Europe: Moscow, Saint Petersburg, Sofia, Riga or Bucharest), North Africa, Middle East (Riyadh, Jeddah and Kuwait) and North America (New York, Toronto and Montreal). The main ports are Algeciras (for freight and container traffic) and Málaga for cruise ships. Algeciras is Spain's leading commercial port, with of cargo in 2004. Seville has Spain's only commercial river port. Other significant commercial ports in Andalusia are the ports of the Bay of Cádiz, Almería and Huelva. The Council of Government has approved a Plan of Infrastructures for the Sustainability of Transport in Andalusia (PISTA) 2007–2013, which plans an investment of 30 billion euros during that period. The lack of high-quality fossil fuels in Andalusia has led to a strong dependency on petroleum imports. Still, Andalusia has a strong potential for the development of renewable energy, above all wind energy. The Andalusian Energy Agency established in 2005 by the autonomous government, is a new governmental organ charged with the development of energy policy and provision of a sufficient supply of energy for the community. The infrastructure for production of electricity consists of eight large thermal power stations, more than 70 hydroelectric power plants, two wind farms, and 14 major cogeneration facilities. Historically, the largest Andalusian business in this sector was the Compañía Sevillana de Electricidad, founded in 1894, absorbed into Endesa in 1996. The Solar power tower PS10 was built by the Andalusian firm Abengoa in Sanlúcar la Mayor in the province of Seville, and began operating in March 2007. It is the largest existing solar power facility in Europe. Smaller solar power stations, also recent, exist at Cúllar and Galera, Granada, inaugurated by Geosol and Caja Granada. Two more large thermosolar facilities, Andasol I y II, planned at Hoya de Guadix in the province of Granada are expected to supply electricity to half a million households. The Plataforma Solar de Almería (PSA) in the Tabernas Desert is an important center for the exploration of the solar energy. The largest wind power firm in the region is the Sociedad Eólica de Andalucía, formed by the merger of Planta Eólica del Sur S.A. and Energía Eólica del Estrecho S.A. As throughout Spain, basic education in Andalusia is free and compulsory. Students are required to complete ten years of schooling, and may not leave school before the age of 16, after which students may continue on to a baccalaureate, to intermediate vocational education, to intermediate-level schooling in arts and design, to intermediate sports studies, or to the working world. Andalusia has a tradition of higher education dating back to the Modern Age and the University of Granada, University of Baeza, and University of Osuna. As of 2009, there are ten private or public universities in Andalucia. University studies are structured in cycles, awarding degrees based on ECTS credits in accord with the Bologna process, which the Andalusian universities are adopting in accord with the other universities of the European Higher Education Area. Responsibility for healthcare jurisdictions devolved from the Spanish government to Andalusia with the enactment of the Statute of Autonomy. Thus, the Andalusian Health Service ("Servicio Andaluz de Salud") currently manages almost all public health resources of the Community, with such exceptions as health resources for prisoners and members of the military, which remain under central administration. According to the Outreach Program for Science in Andalusia, Andalusia contributes 14 percent of Spain's scientific production behind only Madrid and Catalonia among the autonomous communities, even though regional investment in research and development (R&D) as a proportion of GDP is below the national average. The lack of research capacity in business and the low participation of the private sector in research has resulted in R&D taking place largely in the public sector. The Council of Innovation, Science and Business is the organ of the autonomous government responsible for universities, research, technological development, industry, and energy. The council coordinates and initiates scientific and technical innovation through specialized centers an initiatives such as the Andalusian Center for Marine Science and Technology ("Centro Andaluz de Ciencia y Tecnología Marina") and Technological Corporation of Andalusia ("Corporación Tecnológica de Andalucía"). Within the private sphere, although also promoted by public administration, technology parks have been established throughout the Community, such as the Technological Park of Andalucia ("Parque Tecnológico de Andalucía") in Campanillas on the outskirts of Málaga, and Cartuja 93 in Seville. Some of these parks specialize in specific sector, such as Aerópolis in aerospace or Geolit in food technology. The Andalusian government deployed 600,000 Ubuntu desktop computers in their schools. Andalusia has international, national, regional, and local media organizations, which are active gathering and disseminating information (as well as creating and disseminating entertainment). The most notable is the public Radio y Televisión de Andalucía (RTVA), broadcasting on two regional television channels, Canal Sur and Canal Sur 2, four regional radio stations, Canal Sur Radio, Canal Fiesta Radio, Radio Andalucía Información and Canal Flamenco Radio, as well as various digital signals, most notably Canal Sur Andalucía available on cable TV throughout Spain. Different newspapers are published for each Andalusian provincial capital, comarca, or important city. Often, the same newspaper organization publishes different local editions with much shared content, with different mastheads and different local coverage. There are also popular papers distributed without charge, again typically with local editions that share much of their content. No single Andalusian newspaper is distributed throughout the region, not even with local editions. In eastern Andalusia the "Diario Ideal" has editions tailored for the provinces of Almería, Granada, and Jaén. Grupo Joly is based in Andalucia, backed by Andalusian capital, and publishes eight daily newspapers there. Efforts to create a newspaper for the entire autonomous region have not succeeded (the most recent as of 2009 was the "Diario de Andalucía"). The national press ("El País", "El Mundo", "ABC", etc.) include sections or editions specific to Andalusia. Andalusia has two public television stations, both operated by Radio y Televisión de Andalucía (RTVA): In addition, RTVA also operates the national and international cable channel Canal Sur Andalucía, which first broadcast in 1996 as Andalucía Televisión. There are four public radio stations in the region, all operated by RTVA: The patrimony of Andalusia has been shaped by its particular history and geography, as well as its complex flows of population. Andalusia has been home to a succession of peoples and civilizations, many very different from one another, each impacting the settled inhabitants. The ancient Iberians were followed by Celts, Phoenicians and other Eastern Mediterranean traders, Romans, migrating Germanic tribes, Arabs or Berbers. All have shaped the Spanish patrimony in Andalusia, which was already diffused widely in the literary and pictorial genre of the "costumbrismo andaluz". In the 19th century, Andalusian culture came to be widely viewed as the Spanish culture "par excellence", in part thanks to the perceptions of romantic travellers. In the words of Ortega y Gasset: Andalusia has been the birthplace of many great artists: the classic painters Velázquez, Murillo, and Juan de Valdés Leal; the sculptors Juan Martínez Montañés, Alonso Cano and Pedro de Mena; and such modern painters as Daniel Vázquez Díaz and Pablo Picasso. The Spanish composer Manuel de Falla was from Cádiz and incorporated typical Andalusian melodies in his works, as did Joaquín Turina, from Seville. The great singer Camarón de la Isla was born in San Fernando, Cádiz, and Andrés Segovia who helped shape the romantic-modernist approach to classical guitar, was born in Linares, Jaén. The virtuoso Flamenco guitar player Paco de Lucia who helped internationalize Flamenco, was born in Algeciras, Cadiz. Since the Neolithic era, Andalusia has preserved important megaliths, such as the dolmens at the Cueva de Menga and the Dolmen de Viera, both at Antequera. Archeologists have found Bronze Age cities at Los Millares and El Argar. Archeological digs at Doña Blanca in El Puerto de Santa María have revealed the oldest Phoenicians city in the Iberian peninsula; major ruins have also been revealed at Roman Italica near Seville. Some of the greatest architecture in Andalusia were developed across several centuries and civilizations. Such is the case of the Alhambra complex or the Cathedral of Córdoba. The traditional architecture of Andalusia retains its Roman with added Persian and Egyptian influences brought by Arabs, with a marked Mediterranean character strongly conditioned by the climate. Traditional urban houses are constructed with shared walls to minimize exposure to high exterior temperatures. Solid exterior walls are painted with lime to minimize the heating effects of the sun. In accord with the climate and tradition of each area, the roofs may be terraces or tiled in the Roman imbrex and tegula style. One of the most characteristic elements (and one of the most obviously influenced by Roman architecture) is the interior patio or courtyard; the patios of Córdoba are particularly famous. Other characteristic elements are decorative (and functional) wrought iron gratings and the tiles known as "azulejos". Landscaping—both for common private homes and homes on a more lavish scale—also carries on older traditions, with plants, flowers, and fountains, pools, and streams of water. Beyond these general elements, there are also specific local architectural styles, such as the flat roofs, roofed chimneys, and radically extended balconies of the Alpujarra, the cave dwellings of Guadix and of Granada's Sacromonte, or the traditional architecture of the Marquisate of Zenete. The monumental architecture of the centuries immediately after the Reconquista often displayed an assertion of Christian hegemony through architecture that referenced non-Arab influences. Some of the greatest Renaissance buildings in Andalusia are from the then-kingdom of Jaén: the Jaén Cathedral, designed in part by Andrés de Vandelvira, served as a model for the Cathedral of Malaga and Guadix; the centers of Úbeda and Baeza, dating largely from this era, are UNESCO World Heritage Sites. Seville and its kingdom also figured prominently in this era, as is shown by the Casa consistorial de Sevilla, the Hospital de las Cinco Llagas, or the Charterhouse of Jerez de la Frontera. The Palace of Charles V in Granada is uniquely important for its Italianate purism. Andalusia also has such Baroque-era buildings as the Palace of San Telmo in Seville (seat of the current autonomic presidency), the Church of Our Lady of Reposo in Campillos, and the Granada Charterhouse. Academicism gave the region the Royal Tobacco Factory in Seville and Neoclassicism the nucleus of Cádiz, such as its city hall, Royal Prison, and the Oratorio de la Santa Cueva. Revivalist architecture in the 19th and 20th centuries contributed the buildings of the Ibero-American Exposition of 1929 in Seville, including the Neo-Mudéjar Plaza de España. Andalusia also preserves an important industrial patrimony related to various economic activities. Besides the architecture of the cities, there is also much outstanding rural architecture: houses, as well as ranch and farm buildings and dog houses. The Iberian reliefs of Osuna, Lady of Baza, and León de Bujalance, the Phoenician sarcophagi of Cádiz, and the Roman sculptures of the Baetic cities such as Italica give evidence of traditions of sculpture in Andalusia dating back to antiquity. There are few significant surviving sculptures from the time of al-Andalus; two notable exceptions are the lions of the Alhambra and of the Maristán of Granada (the Nasrid hospital in the Albaicín). The Sevillian school of sculpture dating from the 13th century onward and the Granadan school beginning toward the end of the 16th century both focused primarily on Christian religious subject matter, including many wooden altarpieces. Notable sculptors in these traditions include Lorenzo Mercadante de Bretaña, Pedro Millán, Juan Martínez Montañés, Pedro Roldán, José de Arce, Jerónimo Balbás, Alonso Cano, and Pedro de Mena. Non-religious sculpture has also existed in Andalusia since antiquity. A fine example from the Renaissance era is the decoration of the Casa de Pilatos in Seville. Nonetheless, non-religious sculpture played a relatively minor role until such 19th-century sculptors as Antonio Susillo. As in sculpture, there were Sevillian and the Granadan schools of painting. The former has figured prominently in the history of Spanish art since the 15th century and includes such important artists as Zurbarán, Velázquez and Murillo, as well as theoreticians of art such as Francisco Pacheco. The Museum of Fine Arts of Seville and the Prado contain numerous representative works of the Sevillian school of painting. A specific romantic genre known as "costumbrismo andaluz" depicts traditional and folkloric Andalusian subjects, such as bullfighting scenes, dogs, and scenes from Andalusia's history. Important artists in this genre include Manuel Barrón, José García Ramos, Gonzalo Bilbao and Julio Romero de Torres. The genre is well represented in the private Carmen Thyssen-Bornemisza Collection, part of which is on display at Madrid's Thyssen-Bornemisza Museum and Carmen Thyssen Museum in Málaga. Málaga also has been and is an important artistic center. Its most illustrious representative was Pablo Picasso, one of the most influential artists of the 20th century. The city has a Museum and Natal House Foundation, dedicated to the painter. Andalusia plays a significant role in the history of Spanish-language literature, however not all of the important literature associated with Andalusia was written in Spanish. Before 1492, there was the literature written in Andalusian Arabic. Hispano-Arabic authors native to the region include Ibn Hazm, Ibn Zaydún, Ibn Tufail, Al-Mu'tamid, Ibn al-Khatib, Ibn al-Yayyab, and Ibn Zamrak or Andalusian Hebrew poets as Solomon ibn Gabirol. Ibn Quzman, of the 12th century, crafted poems in the colloquial Andalusian language. In 1492 Antonio de Nebrija published his celebrated "Gramática de la lengua castellana" ("Grammar of the Castilian language"), the first such work for a modern European language. In 1528 Francisco Delicado wrote "", a novel in the orbit of "La Celestina", and in 1599 the Sevillian Mateo Alemán wrote the first part of "Guzmán de Alfarache", the first picaresque novel with a known author. The prominent humanist literary school of Seville included such writers as Juan de Mal Lara, Fernando de Herrera, Gutierre de Cetina, Luis Barahona de Soto, Juan de la Cueva, Gonzalo Argote de Molina, and Rodrigo Caro. The Cordoban Luis de Góngora was the greatest exponent of the "culteranismo" of Baroque poetry in the Siglo de Oro; indeed, the style is often referred to as "Góngorismo". Literary Romanticism in Spain had one of its great centers in Andalusia, with such authors as Ángel de Saavedra, 3rd Duke of Rivas, José Cadalso and Gustavo Adolfo Bécquer. "Costumbrismo andaluz" existed in literature as much as in visual art, with notable examples being the "Escenas andaluzas" of Serafín Estébanez Calderón and the works of Pedro Antonio de Alarcón. Andalusian authors Ángel Ganivet, Manuel Gómez-Moreno, Manuel and Antonio Machado, and Francisco Villaespesa are all generally counted in the Generation of '98. Also of this generation were the Quintero brothers, dramatists who faithfully captured Andalusian dialects and idiosyncrasies. Also of note, 1956 Nobel Prize-winning poet Juan Ramón Jiménez was a native of Moguer, near Huelva. A large portion of the "avant-garde" Generation of '27 who gathered at the Ateneo de Sevilla on the 300th anniversary of Góngora's death were Andalusians: Federico García Lorca, Luis Cernuda, Rafael Alberti, Manuel Altolaguirre, Emilio Prados, and 1977 Nobel laureate Vicente Aleixandre. Certain Andalusian fictional characters have become universal archetypes: Prosper Mérimée's gypsy "Carmen", P. D. Eastman's "Perro", Pierre Beaumarchais's "Fígaro", and Tirso de Molina's "Don Juan". As in most regions of Spain, the principal form of popular verse is the romance, although there are also strophes specific to Andalusia, such as the "soleá" or the "". Ballads, lullabies, street vendor's cries, nursery rhymes, and work songs are plentiful. Among the philosophers native to the region can be counted Seneca, Avicebron, Maimonides, Averroes, Fernán Pérez de Oliva, Sebastián Fox Morcillo, Ángel Ganivet, Francisco Giner de los Ríos and María Zambrano. The music of Andalusia includes traditional and contemporary music, folk and composed music, and ranges from flamenco to rock. Conversely, certain metric, melodic and harmonic characteristics are considered Andalusian even when written or performed by musicians from elsewhere. Flamenco, perhaps the most characteristically Andalusian genre of music and dance, originated in the 18th century, but is based in earlier forms from the region. The influence of the traditional music and dance of the Romani people or Gypsies is particularly clear. The genre embraces distinct vocal ("cante flamenco"), guitar ("toque flamenco"), and dance ("baile flamenco") styles. The Andalusian Statute of Autonomy reflects the cultural importance of flamenco in its Articles 37.1.18 and 68: Fundamental in the history of Andalusian music are the composers Cristóbal de Morales, Francisco Guerrero, Francisco Correa de Arauxo, Manuel García, Manuel de Falla, Joaquín Turina, and Manuel Castillo, as well as one of the fathers of modern classical guitar, the guitarist Andrés Segovia. Mention should also be made of the great folk artists of the "copla (music)" and the "cante hondo", such as Rocío Jurado, Lola Flores ("La Faraona", "the pharaoh"), Juanito Valderrama and the revolutionary Camarón de la Isla. Prominent Andalusian rock groups include Triana and Medina Azahara. The duo Los del Río from Dos Hermanas had international success with their "Macarena", including playing at a Super Bowl half-time show in the United States, where their song has also been used as campaign music by the Democratic Party. Other notables include the singer, songwriter, and poet Joaquín Sabina, Isabel Pantoja, Rosa López, who represented Spain at Eurovision in 2002, and David Bisbal. The portrayal of Andalusia in film is often reduced to archetypes: flamenco, bullfighting, Catholic pageantry, brigands, the property-rich and cash-poor "señorito andaluz" and emigrants. These images particularly predominated from the 1920s through the 1960s, and helped to consolidate a cliched image of the region. In a very different vein, the province of Almería was the filming location for many Westerns, especially (but by no means exclusively) the Italian-directed Spaghetti Westerns. During the dictatorship of Francisco Franco, this was the extent of the film industry in Andalusia. Nonetheless, Andalusian film has roots as far back as José Val del Omar in the pre-Franco years, and since the Spanish transition to democracy has brought forth numerous nationally and internationally respected directors: Antonio Cuadri ("Heart of the Earth"), Chus Gutiérrez ("Poniente"), Chiqui Carabante ("Carlos Against the World"), Alberto Rodríguez ("7 Virgins"), Benito Zambrano ("Solas"), and Antonio Banderas ("Summer Rain"). Counting together feature films, documentaries, television programs, music videos etc., Andalusia has boomed from 37 projects shooting in 1999 to 1,054 in 2007, with the figure for 2007 including 19 feature films. Although feature films are the most prestigious, commercials and television are currently more economically important to the region. The Filmoteca de Andalucía, headquartered in Córdoba, is a government-run entity in charge of the investigation, collection and diffusion of Andalusian cinematic heritage. Other important contributors to this last activity are such annual film festivals as the ("Festival de Málaga Cine Español (FMCE)"), the most important festival dedicated exclusively to cinema made in Spain, the Seville Festival of European Film (SFCE), the International Festival of Short Films—Almería in Short, the Huelva Festival of Latin American Film, the Atlantic Film Show in Cádiz, the Islantilla Festival of Film and Television and the African Film Festival of Tarifa. Each sub-region in Andalusia has its own unique customs that represent a fusion of Catholicism and local folklore. Cities like Almería have been influenced historically by both Granada and Murcia in the use of traditional head coverings. The "sombrero de Labrador", a worker's hat made of black velvet, is a signature style of the region. In Cádiz, traditional costumes with rural origins are worn at bullfights and at parties on the large estates. The "tablao flamenco" dance and the accompanying "cante jondo" vocal style originated in Andalusia and traditionally most often performed by the gypsy Romani. One of the most distinctive cultural events in Andalusia is the Romería de El Rocío in May. It consists of a pilgrimage to the Hermitage of El Rocío in the countryside near Almonte, in honor of the Virgin of El Rocío, an image of the Virgin and Child. In recent times the "Romería" has attracted roughly a million pilgrims each year. In Jaén, the saeta is a revered form of Spanish religious song, whose form and style has evolved over many centuries. Saetas evoke strong emotion and are sung most often during public processions. "Verdiales", based upon the fandango, are a flamenco music style and song form originating in Almogia, near Málaga. For this reason, the Verdiales are sometimes known as "Fandangos de Málaga." The region also has a rich musical tradition of flamenco songs, or palos called cartageneras. Seville celebrates "Semana Santa", one of the better known religious events within Spain. During the festival, religious fraternities dress as penitents and carry large floats of lifelike wooden sculptures representing scenes of the Passion, and images of the Virgin Mary. Sevillanas, a type of old folk music sung and written in Seville and still very popular, are performed in fairs and festivals, along with an associated dance for the music, the "Baile por sevillanas". All the different regions of Andalusia have developed their own distinctive customs, but all share a connectedness to Catholicism as developed during baroque Spain society. Andalusian Spanish is one of the most widely spoken forms of Spanish in Spain, and because of emigration patterns was very influential on American Spanish. Rather than a single dialect, it is really a range of dialects sharing some common features; among these is the retention of more Arabic words than elsewhere in Spain, as well as some phonological differences compared with Standard Spanish. The isoglosses that mark the borders of Andalusian Spanish overlap to form a network of divergent boundaries, so there is no clear border for the linguistic region. The territory now known as Andalusia fell within the sphere of influence of ancient Mediterranean mythological beliefs. Phoenician colonization brought the cults of Baal and Melqart; the latter lasted into Roman times as Hercules, mythical founder of both Cádiz and Seville. The Islote de Sancti Petri held the supposed tomb of Hercules, with representations of his Twelve labors; the region was the traditional site of the tenth labor, obtaining the cattle of the monster Geryon. Traditionally, the Pillars of Hercules flank the Strait of Gibraltar. Clearly, the European pillar is the Rock of Gibraltar; the African pillar was presumably either Monte Hacho in Ceuta or Jebel Musa in Morocco. The Roman road that led from Cádiz to Rome was known by several names, one of them being "Via Herculea", Hercules route returning from his tenth labor. The present coat of arms of Andalusia shows Hercules between two lions, with two pillars behind these figures. Roman Catholicism is, by far, the largest religion in Andalusia. In 2012, the proportion of Andalusians that identify themselves as Roman Catholic was 78.8%. The principal characteristic of the local popular form of Catholicism is devotion to the Virgin Mary; Andalusia is sometimes known as "la tierra de María Santísima" ("the land of Most Holy Mary"). Also characteristic are the processions during Holy Week, in which thousands of penitents (known as "nazarenos") sing saetas. Andalusia is the site of such pilgrim destinations as the Santuario de Nuestra Señora de la Cabeza in Andújar and the Hermitage of El Rocío in Almonte. While some trace the lineage of the Spanish Fighting Bull back to Roman times, today's fighting bulls in the Iberian peninsula and in the former Spanish Empire trace back to Andalusia in the 15th and 16th centuries. Andalusia remains a center of bull-rearing and bullfighting: its 227 "fincas de ganado" where fighting bulls are raised cover . In the year 2000, Andalusia's roughly 100 bullrings hosted 1,139 "corridas". The oldest bullring still in use in Spain is the neoclassical "Plaza de toros" in Ronda, built in 1784. The Andalusian Autonomous Government sponsors the "Rutas de Andalucía taurina", a touristic route through the region centered on bullfighting. The Andalusian festivals provide a showcase for popular arts and traditional costume. Among the most famous of these are the Seville Fair or "Feria de Abril" in Seville, now echoed by smaller fairs in Madrid and Barcelona, both of which have many Andalusian immigrants; the "Feria de Agosto" in Málaga; the Feria de Jerez or "Feria del Caballo" in Jerez; the Festival of Corpus Christi in Granada; the Feria de Nuestra Señora de la Salud in Córdoba; the Columbian Festivals ("Fiestas Colombinas") in Huelva; the Feria de la Virgen del Mar in Almería; and the Feria de San Lucas in Jaén, among many others. Festivals of a religious nature are a deep Andalusian tradition and are met with great popular fervor. There are numerous major festivals during Holy Week. An annual pilgrimage brings a million visitors to the Hermitage of El Rocío in Almonte (population 16,914 in 2008); similarly large crowds visit the Santuario de Nuestra Señora de la Cabeza in Andújar every April. Other important festivals are the Carnival of Cádiz and the Fiesta de las Cruces or Cruz de mayo in Granada and Córdoba; in Córdoba this is combined with a competition for among the "patios" (courtyards) of the city. Andalusia hosts an annual festival for the dance of flamenco in the summer-time. The Andalusian diet varies, especially between the coast and the interior, but in general is a Mediterranean diet based on olive oil, cereals, legumes, vegetables, fish, dried fruits and nuts, and meat; there is also a great tradition of drinking wine. Fried fish—"pescaíto frito"—and seafood are common on the coast and also eaten well into the interior under coastal influence. Atlantic bluefin tuna ("Thunnus thynnus") from the Almadraba areas of the Gulf of Cádiz, prawns from Sanlúcar de Barrameda (known as "langostino de Sanlúcar"), and deepwater rose shrimp ("") from Huelva are all highly prized. Fishing for the transparent goby or "chanquete" ("Aphia minuta"), a once-popular small fish from Málaga, is now banned because the techniques used to catch them trap too many immature fish of other species. The mountainous regions of the Sierra Morena and Sierra Nevada produce cured hams, notably including "jamón serrano" and "jamón ibérico". These come from two different types of pig, ("jamón serrano" from white pigs, the more expensive "jamón ibérico" from the Black Iberian pig). There are several denominaciones de origen, each with its own specifications including in just which microclimate region ham of a particular denomination must be cured. "Plato alpujarreño" is another mountain specialty, a dish combining ham, sausage, sometimes other pork, egg, potatoes, and olive oil. Confectionery is popular in Andalusia. Almonds and honey are common ingredients. Many enclosed convents of nuns make and sell pastries, especially Christmas pastries: "mantecados", "polvorones", "pestiños", "alfajores", "yemas de San Leandro", as well as "churros" or ', meringue cookies ("merengadas"), and '. Cereal-based dishes include "migas de harina" in eastern Andalusia (a similar dish to couscous rather than the fried breadcrumb based "migas" elsewhere in Spain) and a sweeter, more aromatic porridge called "poleá" in western Andalusia. Vegetables form the basis of such dishes as ' (similar to "ratatouille") and the chopped salad known as ' or "". Hot and cold soups based in olive oil, garlic, bread, tomato and peppers include "gazpacho", "salmorejo", "porra antequerana", "ajo caliente", "sopa campera", or—using almonds instead of tomato—"ajoblanco". Wine has a privileged place at the Andalusian table. Andalusian wines are known worldwide, especially fortified wines such as sherry ("jerez"), aged in soleras. These are enormously varied; for example, dry sherry may be the very distinct "fino", "manzanilla", "amontillado", "oloroso", or "Palo Cortado" and each of these varieties can each be sweetened with Pedro Ximénez or Moscatel to produce a different variety of sweet sherry. Besides sherry, Andalucía has five other denominaciones de origen for wine: D.O. Condado de Huelva, D.O. Manzanilla-Sanlúcar de Barrameda, D.O. Málaga, D.O. Montilla-Moriles, and D.O. Sierras de Málaga. Most Andalusian wine comes from one of these regions, but there are other historic wines without a Protected Geographical Status, for example Tintilla de Rota, Pajarete, Moscatel de Chipiona and Mosto de Umbrete. Andalusia also produces D.O. vinegar and brandy: D.O. Vinagre de Jerez and D.O. Brandy de Jerez. The traditional dress of 18th-century Andalusia was strongly influenced by "" within the context of "casticismo" (purism, traditionalism, authenticity). The archetype of the "majo" and "maja" was that of a bold, pure Spaniard from a lower-class background, somewhat flamboyant in his or her style of dress. This emulation of lower-class dress also extended to imitating the clothes of brigands and Romani ("Gypsy") women. The Museum of Arts and Traditions of Sevilla has collected representative samples of a great deal of the history of Andalusian dress, including examples of such notable types of hat as the "sombrero cordobés, "sombrero calañés, "sombrero de catite" and the "pavero", as well as the "traje corto" and "traje de flamenca". Andalusia has a great artisan tradition in tile, leather ("see Shell cordovan"), weaving (especially of the heavy "jarapa" cloth), marquetry, and ceramics (especially in Jaén, Granada, and Almería), lace (especially Granada and Huelva), embroidery (in Andévalo), ironwork, woodworking, and basketry in wicker, many of these traditions a heritage of the long period of Muslim rule. Andalusia is also known for its dogs, particularly the Andalusian Hound, which was originally bred in the region. Dogs, not just andalusian hounds, are very popular in the region. Andalusian equestrianism, institutionalized in the Royal Andalusian School of Equestrian Art is known well beyond the borders of Spain. The Andalusian horse is strongly built, compact yet elegant, distinguished in the area of dressage and show jumping, and is also an excellent horse for driving. They are known for their elegant "dancing" gait. In Andalusia, as throughout Spain, football is the predominant sport. Introduced to Spain by British men who worked in mining for Rio Tinto in the province of Huelva, the sport soon became popular with the local population. As Spain's oldest existing football club, Recreativo de Huelva, founded 1889, is known as "El Decano" ("the Dean"). For the 2019/20 season, 3 Andalusian clubs will compete in Spain's First Division "La Liga": Granada CF, Real Betis, and Sevilla FC. Betis won La Liga in 1934–35 and Sevilla in the 1945–46 season. The seven other Andalusian teams, Cádiz CF, Córdoba CF, UD Almería and Málaga CF play in the Segunda División, whilst Recreativo de Huelva, Spain's oldest club and Marbella FC participate in the Segunda División B, and Real Jaén participates in the Tercera División. The Andalusia autonomous football team is not in any league, and plays only friendly matches. In recent years, they have played mostly during the Christmas break of the football leagues. They play mostly against national teams from other countries, but would not be eligible for international league play, where Spain is represented by a single national team. In recent decades, basketball has become increasingly popular, with CB Málaga, also known as Unicaja Málaga who have won the Liga ACB in 2007 and the Korać Cup in 2001 and usually play the Euroleague, CB Sevilla (Banca Cívica) and CB Granada competing at the top level in the Liga ACB. Unlike basketball, handball has never really taken off in Andalusia. There is one Andalusian team in the Liga Asobal, Spain's premier handball league: BM Puente Genil, playing in the province of Córdoba. Andalusia's strongest showing in sports has been in table tennis. There are two professional teams: Cajasur Priego TM and Caja Granada TM, the latter being Spain's leading table tennis team, with more than 20 league championships in nearly consecutive years and 14 consecutive Copas del Rey, dominating the Liga ENEBÉ. Cajasur is also one of the league's leading teams. 220 Andalusian athletes have competed in a total of 16 summer or winter Olympic Games. The first was Leopoldo Sainz de la Maza, part of the silver-medal-winning polo team at the 1920 Summer Olympics in Antwerp, Belgium. In all, Andalusians have won 6 gold medals, 11 silver, and 2 bronze. Winners of multiple medals include the Cordoban boxer Rafael Lozano (bronze in the 1996 Summer Olympics at Atlanta, Georgia, US, and silver in the 2000 Summer Olympics in Sydney, Australia); sailor Theresa Zabell, Malagueña by adoption (gold medals at Barcelona in 1992 and Atlanta in 1996). Other notable winners have been Granadan tennis player Manuel Orantes (silver in the men's singles of the demonstration tournament in Mexico City in 1968), Jerezano riders Ignacio Rambla and Rafael Soto (silver in dressage in Athens in 2004) and the racewalker Paquillo Fernández from Guadix (silver in Athens in 2004). The largest number of Olympic appearances were by the Malagueña swimmer María Peláez (five appearances), the Granadan skier María José Rienda (four), the Sevillian rider Luis Astolfi (four), and the Sevillian rower Fernando Climent (four, including a silver at Los Angeles, California, US, in 1984. Seville has been a pre-candidate to host the Summer Olympics in two occasions, 2004 and 2008, and Granada has been a pre-candidate to host the winter Olympics; neither has ever succeeded in its candidature. The ski resort of Sierra Nevada, near Granada, has however hosted the 1996 Alpine World Ski Championships, and Granada hosted the 2015 Winter Universiade. Other sporting events in Andalusia include surfing, kitesurfing and windsurfing competitions at Tarifa, various golf tournaments at courses along the coast, and horse racing and polo at several locations in the interior. Andalusia hosted the 1999 World Championships in Athletics (Seville), the 2005 Mediterranean Games (Almería) and the FIS Alpine World Ski Championships 1996 (Granada), among other major events. There is also the annual Vuelta a Andalucía bicycle road race and the Linares chess tournament. Andalusia has had a sister region relationship with Buenos Aires (Argentina), since 2001; and with Córdoba (Argentina). Also Andalusia has a collaboration agreement with Guerrero (Mexico).
https://en.wikipedia.org/wiki?curid=2736
Abigail Abigail (, "avigáyil", "’ǎḇîḡayil") was the wife of Nabal; she became a wife of the future King David after Nabal's death (1 Samuel ). Abigail was David's third wife, after Saul's daughter, Michal, whom Saul later married to Palti, son of Laish when David went into hiding, and Ahinoam. She became the mother of one of David's sons, who is listed in the Book of Chronicles under the name "Daniel", in the Masoretic Text of the Books of Samuel as "Chileab," and in the Septuagint text of 2 Samuel 3:3 as Δαλουια, "Dalouia". Her name is spelled Abigal in in the American Standard Version. Derived from the Hebrew word "ab," "father", and the Hebrew root "g-y-l", "to rejoice," the name Abigail has a variety of possible meanings including "my father's joy" and "source of joy". In 1 Samuel 25, Nabal demonstrates ingratitude towards David, the son of Jesse (from the tribe of Judah), and Abigail attempts to placate David, in order to stop the future King from taking revenge. She gives him food, and speaks to him, urging him not to "have on his conscience the staggering burden of needless bloodshed" (verse 31, NIV) and reminding him that God will make him a "lasting dynasty" (verse 28). Jon Levenson calls this an "undeniable adumbration" of Nathan's prophecy in 2 Samuel 7. Alice Bach notes that Abigail pronounces a "crucial prophecy," and the Talmud regards her as one of the Tanakh's seven female prophets. Levenson, however, suggests that she "senses the drift of history" from intelligence rather than from special revelation. After Abigail reveals to Nabal what she has done, "God struck Nabal and he died," (v.38), after which David married her. Abigail is described as intelligent and beautiful. The Talmud amplifies this idea, mentioning her as being one of the "four women of surpassing beauty in the world," (the other three being Rahab, Sarah, and Esther). As the wife of the wealthy Nabal, she is also a woman of high socioeconomic status. Whether David married her because he was attracted to her, or as an astute political move, or both is unclear. Abigail and David's second wife, Ahinoam the Jezreelite, accompany David and his war band as they seek refuge in Philistine territory. While David and his men are encamped near Jezreel, they are captured by Amalekites who raided the town of Ziglak and carried off the women and children. David led the pursuit, and they were subsequently rescued. Both wives then settle with David in Hebron, where Abigail gives birth to David's second son, Chileab (also called Daniel). Abigail is also listed as one of the seven Jewish women prophets, the other six being Miriam, Deborah, Hannah, Sarah, Huldah,and Esther. In terms of her moral character, Abraham Kuyper argues that Abigail's conduct indicates "a most appealing character and unwavering faith," but Alice Bach regards her as subversive. Adele Berlin contrasts the story of Abigail with that of Bathsheba. In one, the wife prevents David from murdering her foolish and greedy husband. In the second, David orders the murder of a good man because he desires his wife. "In the Abigail story, David, the potential king, is seen as increasingly strong and virtuous, whereas in the Bathsheba story, the reigning monarch shows his flaws ever more overtly and begins to lose control of his family." Levenson and Halpern suggest that Abigail may, in fact, also be the same person as Abigail, mother of Amasa. Richard M. Davidson, however, points out that "on the basis of the final form of Old Testament canon, references to Abigail in the biblical accounts indicate two different individuals." Abigail's self-styling as a "handmaid" led to "Abigail" being a traditional term for a waiting-woman, for example as the "waiting gentlewoman" in Beaumont and Fletcher's "The Scornful Lady", published in 1616. Jonathan Swift and Henry Fielding use "Abigail" in this generic sense, as does Charlotte Brontë. Anthony Trollope makes two references to "the abigail" (all lower case) in "The Eustace Diamonds", at the beginning of Chapter 42, whilst Thomas Mann makes the same reference at the start of the second chapter of Part 2 in Buddenbrooks (published in 1901). William Rose Benet notes the notoriety of Abigail Hill, better known as "Mrs Masham", a lady-in-waiting to Queen Anne. George MacDonald Fraser makes mention of "an "abigail" fussing about the room" in his novel "Flashman" from "The Flashman Papers" series. Abigail is a featured figure on Judy Chicago's installation piece "The Dinner Party", being represented in one of the 999 tiles of the "Heritage Floor."
https://en.wikipedia.org/wiki?curid=2741
Azad Kashmir Azad Jammu and Kashmir (, ), abbreviated as AJK and commonly known as Azad Kashmir, is a region administered by Pakistan as a nominally self-governing jurisdiction, and constituting the western portion of the larger Kashmir region which has been the subject of a dispute between India and Pakistan since 1947, and between India and China since 1962. The territory shares a border with Gilgit-Baltistan, together with which it is referred to by the United Nations and other international organisations as "Pakistan administered Kashmir". Azad Kashmir is one-sixth of the size of Gilgit-Baltistan. The territory also borders Pakistan's Punjab province to the south and Khyber Pakhtunkhwa province to the west. To the east, Azad Kashmir is separated from the Indian administered Kashmir by the Line of Control, the "de facto" border between India and Pakistan. Azad Kashmir has a total area of , and a total population of 4,045,366 as per the 2017 Census. The territory has a parliamentary form of government modelled after the Westminster system, with its capital located at Muzaffarabad. The President is the constitutional head of state, while the Prime Minister, supported by a Council of Ministers, is the chief executive. The unicameral Azad Kashmir Legislative Assembly elects both the Prime Minister and President. The state has its own Supreme Court and a High Court, while the Government of Pakistan's Ministry of Kashmir Affairs and Gilgit-Baltistan serves as a link with Azad Kashmir's government, although Azad Kashmir is not represented in the Parliament of Pakistan. An earthquake in 2005 killed 100,000 people and left another three million people displaced, with widespread devastation. Since then, with help from the Government of Pakistan and foreign donors, reconstruction of infrastructure is underway. Azad Kashmir's economy largely depends on agriculture, services, tourism, and remittances sent by members of the British Mirpuri community. Nearly 87% of the households own farms in Azad Kashmir, while the region has a literacy rate of approximately 72% and has the highest school enrollment in Pakistan. "Azad Kashmir" (Free Kashmir) was the title of a pamphlet issued by the Muslim Conference party at its 13th general session held in 1945 at Poonch. It is believed to have been a response to the National Conference's "Naya Kashmir" (New Kashmir) programme. Sources state that it was no more than a compilation of various resolutions passed by the party. But its intent seems to have been to declare that the Muslims of Jammu and Kashmir were committed to the Muslim League's struggle for a separate homeland (Pakistan), and that the Muslim Conference was their sole representative organisation. However, the following year, the party passed an "Azad Kashmir resolution" demanding the Maharaja to institute a Constituent Assembly elected on an extended franchise. According to scholar Chitralekha Zutshi, the organisation's declared goal was to achieve responsible government under the aegis of the Maharaja, without association with either India or Pakistan. The following year, the party workers assembled at the house of Sardar Ibrahim on 19 July 1947 reversed the decision, demanding that the Maharaja accede to Pakistan. Soon afterwards, Sardar Ibrahim escaped to Pakistan and led the Poonch rebellion from there with the assistance of Pakistan's prime minister Liaquat Ali Khan and other officials. Liaquat Ali Khan appointed a committee headed by Mian Iftikharuddin to draft a "declaration of freedom". On 4 October an Azad Kashmir provisional government was declared in Lahore with Ghulam Nabi Gilkar as president under the assumed name "Mr. Anwar" and Sardar Ibrahim as the prime minister. Gilkar travelled to Srinagar and was arrested by Maharaja's government. Pakistani officials subsequently appointed Sardar Ibrahim as the president. The northern part of Azad Jammu and Kashmir encompasses the lower area of the Himalayas, including Jamgarh Peak (). However, Hari Parbat peak in Neelum Valley is the highest peak in the state. The region receives rainfall in both the winter and the summer. Muzaffarabad and Pattan are among the wettest areas of Pakistan. Throughout most of the region, the average rainfall exceeds 1400 mm, with the highest average rainfall occurring near Muzaffarabad (around 1800 mm). During the summer season, monsoon floods of the rivers Jhelum and Leepa are common due to extreme rains and snow melting. At the time of the Partition of India in 1947, the British abandoned their suzerainty over the princely states, which were left with the options of joining India or Pakistan or remaining independent. Hari Singh, the maharaja of Jammu and Kashmir, wanted his state to remain independent. Muslims in Western Jammu province (current day Azad Kashmir) and the Frontier Districts Province (current day Gilgit-Baltistan) had wanted to join Pakistan. In Spring 1947, an uprising against the Maharaja broke out in Poonch, an area bordering the Rawalpindi division of West Punjab. Maharaja's administration is said to have started levying punitive taxes on the peasantry which provoked a local revolt and the administration resorted to brutal suppression. The area's population, swelled by recently demobilised soldiers following World War II, rebelled against the Maharaja's forces and gained control of almost the entire district. Following this victory, the pro-Pakistan chieftains of the western districts of Muzaffarabad, Poonch and Mirpur proclaimed a provisional Azad Jammu and Kashmir government in Rawalpindi on October 3, 1947. Ghulam Nabi Gilkar, under the assumed name "Mr. Anwar," issued a proclamation in the name of the provisional government in Muzaffarabad. However, this government quickly fizzled out with the arrest of Anwar in Srinagar. On October 24, a second provisional government of Azad Kashmir was established at Palandri under the leadership of Sardar Ibrahim Khan. On October 21, several thousand Pashtun tribesmen from North-West Frontier Province poured into Jammu and Kashmir to liberate it from the Maharaja's rule. They were led by experienced military leaders and were equipped with modern arms. The Maharaja's crumbling forces were unable to withstand the onslaught. The raiders captured the towns of Muzaffarabad and Baramulla, the latter northwest of the state capital Srinagar. On October 24, the Maharaja requested military assistance from India, which responded that it was unable to help him unless he acceded to India. Accordingly, on October 26, 1947, Maharaja Hari Singh signed an Instrument of Accession, handing over control of defence, external affairs and communications to the Government of India in return for military aid. Indian troops were immediately airlifted into Srinagar. Pakistan intervened subsequently. Fighting ensued between the Indian and Pakistani armies, with the two areas of control more or less stabilised around what is now known as the "Line of Control". India later approached the United Nations, asking it to resolve the dispute, and resolutions were passed in favour of the holding of a plebiscite with regard to Kashmir's future. However, no such plebiscite has ever been held on either side, since there was a precondition which required the withdrawal of the Pakistani Army along with the non-state elements and the subsequent partial withdrawal of the Indian Army. from the parts of Kashmir under their respective control – a withdrawal that never took place. In 1949, a formal cease-fire line separating the Indian- and Pakistani-controlled parts of Kashmir came into effect. Following the 1949 cease-fire agreement with India, the government of Pakistan divided the northern and western parts of Kashmir that it controlled at the time of cease-fire into the following two separately-controlled political entities: At one time under Pakistani control, Kashmir's Shaksgam tract, a small region along the northeastern border of Gilgit–Baltistan, was provisionally ceded by Pakistan to the People's Republic of China in 1963 and now forms part of China's Xinjiang Uygur Autonomous Region. In 1972, the then current border between the Indian and Pakistani controlled parts of Kashmir was designated as the "Line of Control". This line has remained unchanged since the 1972 Simla Agreement, which bound the two countries "to settle their differences by peaceful means through bilateral negotiations". Some political experts claim that, in view of that pact, the only solution to the issue is mutual negotiation between the two countries without involving a third party such as the United Nations. The 1974 Interim Constitution Act was passed by the 48-member Azad Jammu and Kashmir unicameral assembly. Azad Jammu and Kashmir (AJK) is a self-governing state under Pakistani control, but under Pakistan's constitution, the state is informally part of the country. Pakistan is administering the region as a self-governing territory rather than incorporating it in the federation since the UN-mandated ceasefire. Azad Kashmir has its own elected President, Prime Minister, Legislative Assembly, High Court, with Azam Khan as its present chief justice, and official flag. Azad Kashmir's financial matters, i.e., budget and tax affairs, are dealt with by the Azad Jammu and Kashmir Council rather than by Pakistan's Central Board of Revenue. The Azad Jammu and Kashmir Council is a supreme body consisting of 14 members, 8 from the government of Azad Jammu and Kashmir and 6 from the government of Pakistan. Its chairman/chief executive is the prime minister of Pakistan. Other members of the council are the president and the prime minister of Azad Kashmir (or and individual nominated by her/him) and 6 members of the AJK Legislative Assembly. Azad Kashmir Day is celebrated in Azad Jammu and Kashmir on October 24, which is the day that the Azad Jammu and Kashmir government was created in 1947. Pakistan has celebrated Kashmir Solidarity Day on February 5 of each year since 1990 as a day of protest against India's "de facto" sovereignty over its State of Jammu and Kashmir. That day is a national holiday in Pakistan. Kashmiris in Azad Kashmir observe the Kashmir Black Day on October 27 of each year since 1947 as a day of protest against military occupation in Indian controlled Jammu and Kashmir. Brad Adams the Asia director at the U.S.-based NGO Human Rights Watch has said in 2006; "Although 'azad' means 'free,' the residents of Azad Kashmir are anything but, the Pakistani authorities govern Azad Kashmir government with tight controls on basic freedoms." Scholar Christopher Snedden has observed that despite tight controls the people of Azad Kashmir have generally accepted whatever Pakistan has done to them, which in any case has varied little from how most Pakistanis have been treated (by Pakistan). According to Christopher Snedden, one of the reasons for this was that the people of Azad Kashmir had always wanted to be a part of Pakistan. Consequently, having little to fear from a pro-Pakistan population devoid of options, Pakistan imposed its will through the Federal Ministry of Kashmir Affairs and failed to empower the people of Azad Kashmir, allowing genuine self-government for only a short period in the 1970s. The Interim Constitution of the 1970s only allows the political parties that pay allegiance to Pakistan: "No person or political party in Azad Jammu and Kashmir shall be permitted... activities prejudicial or detrimental to the State's accession to Pakistan." The pro-independence Jammu and Kashmir Liberation Front has never been allowed to contest elections in Azad Kashmir. While the Interim Constitution does not give them a choice, the people of Azad Kashmir have not considered any option other than joining Pakistan. Except in the legal sense, Azad Kashmir has been fully integrated into Pakistan. According to the project report by the Asian Development Bank, the bank has set out development goals for Azad Kashmir in the areas of health, education, nutrition, and social development. The whole project is estimated to cost US$76 million. Germany, between 2006 and 2014, has also donated $38 million towards the AJK Health Infrastructure Programme. The state is administratively divided into three divisions which, in turn, are divided into ten districts. The southern parts of Azad Kashmir including Bhimber, Mirpur and Kotli districts has extremely hot weather in summers and moderate cold weather in winters. It receives rains mostly in monsoon weather. In the central and northern parts of state weather remains moderate hot in summers and very cold and chilly in winter. Snow fall also occurs there in December and January. This region receives rainfall in both winters and summers. Muzaffarabad and Pattan are among the wettest areas of the state. Throughout most of the region, the average rainfall exceeds 1400 mm, with the highest average rainfall occurring near Muzaffarabad (around 1800 mm). During summer, monsoon floods of the Jhelum and Leepa rivers are common, due to high rainfall and melting snow. The population of Azad Kashmir, according to the preliminary results of the 2017 Census, is 4.45 million. The website of the AJK government reports the literacy rate to be 74%, with the enrolment rate in primary school being 98% and 90% for boys and girls respectively. The population of Azad Kashmir is almost entirely Muslim. The people of this region culturally differ from the Kashmiris living in the Kashmir Valley of Jammu and Kashmir, and are closer to the culture of Jammu. Mirpur, Kotli and Bhimber are all old towns of the Jammu region. Azad Jammu and Kashmir has an almost entirely Muslim population. Most residents of the region are not ethnic Kashmiris. The majority of people in Azad Kashmir are ethnically Punjabi. According to a data maintained by Christian community organizations, there are around 4,500 Christian residents in the region. Bhimber is home to most of them, followed by Mirpur and Muzaffarabad. A few dozen families also live in Kotli, Poonch and Bagh. However, the Christian community has been struggling to get residential status and property rights in AJK. There is no official data on the total number of Bahais in AJK. Only six families are known to be living in Muzaffarabad while some of them live in rural areas. The followers of the Ahmadi faith is estimated to be somewhere between 20,000 and 25,000 and most of them live in Kotli, Mirpur Bhimber and Muzaffarabad. The main communities living in this region are: The culture of Azad Kashmir has many similarities to that of northern Punjabi (Potohar) culture in Punjab province, while the Sudhans have oral tradition of Pashtuns. The Peshawari turban is worn by some Sudhans in the area. The traditional dress of the women is the shalwar kameez in Pahari style. The shalwar kameez is commonly worn by both men and women. Women use shawl to cover their head and upper body. The official language of Azad Kashmir is Urdu, while English is used in higher domains. The majority of the population, however, are native speakers of other languages. The foremost among these is Pahari–Pothwari, with its various dialects. There are also sizeable communities speaking Gujari and Kashmiri, as well as pockets of speakers of Shina, Pashto and Kundal Shahi. With the exception of Pashto and English, these languages belong to the Indo-Aryan language family. The dialects of the Pahari-Pothwari language complex cover most of the territory of Azad Kashmir. These are also spoken across the Line of Control in neighbouring areas of Indian Jammu and Kashmir, and are closely related both to Punjabi to the south and Hinko to the northwest. The language variety in the southern districts of Azad Kashmir is known by a variety of names – including "Mirpuri", "Pothwari" and "Pahari" – and is closely related to the Pothwari proper spoken to the east in the Pothohar region of Punjab. The dialects of the central districts of Azad Kashmir are occasionally referred to in the literature as "Chibhali" or "Punchi", but the speakers themselves usually call them "Pahari", an ambiguous name that is also used for several unrelated languages of the Lower Himalayas. Going north, the speech forms gradually change into Hindko. Already in Muzaffarabad District the preferred local name for the language is "Hindko", although it is still apparently more closely related to the core dialects of Pahari. Further north in the Neelam Valley the dialect, locally also known as "Parmi", can more unambiguously be subsumed under Hindko. Another major language of Azad Kashmir is Gujari. It is spoken by several hundred thousand people among the traditionally nomadic Gujars, many of whom are nowadays settled. Not all ethnic Gujars speak Gujari, the proportion of those who have shifted to other languages is probably higher in southern Azad Kashmir. Gujari is most closely related to the Rajasthani languages (particularly Mewati), although it also shares features with Punjabi. It is dispersed over large areas in northern Pakistan and India. Within Pakistan, the Gujari dialects of Azad Kashmir are more similar, in terms of shared basic vocabulary and mutual intelligibility, to the Gujar varieties of the neighbouring Hazara region than to the dialects spoken further to the northwest in Khyber Pakhtunkhwa and north in Gilgit. There are scattered communities of Kashmiri speakers, notably in the Neelam Valley, where they form the second-largest language group after speakers of Hindko. There have been calls for the teaching of Kashmiri (particularly in order to counter India's claim of promoting the culture of Kashmir), but the limited attempts at introducing the language at the secondary school level have not been successful, and it is Urdu, rather than Kashmiri, that Kashmiri Muslims have seen as their identity symbol. There is an ongoing process of gradual shift to larger local languages, but at least in the Neelam Valley there still exist communities for whom Kashmiri is the sole mother tongue. In the northernmost district of Neelam, there are pockets of other languages. Shina, which like Kashmiri belongs to the Dardic group, is present in two distinct varieties spoken altogether in three villages. The Iranian language Pashto, the major language of the neighbouring province of Khyber Pakhtunkhwa, is spoken in two villages in Azad Kashmir, both situated on the Line of Control. The endangered Kundal Shahi is native to the eponymous village and it is the only language not found outside Azad Kashmir. Historically the economy of Azad Kashmir has been agricultural which meant that land was the main source or mean of production. This means that all food for immediate and long term consumption was produced from land. The produce included various crops, fruits, vegetables, etc. Land was also the source of other livelihood necessities such as wood, fuel, grazing for animals which then turned into dairy products. Because of this land was also the main source of revenue for the governments whose primary purpose for centuries was to accumulate revenue. Agriculture is a major part of Azad Kashmir's economy. Low-lying areas that have high populations grow crops like barley, mangoes, millet, corn (maize), and wheat, and also raise cattle. In the elevated areas that are less populated and more spread-out, forestry, corn, and livestock are the main sources of income. There are mineral and marble resources in Azad Kashmir close to Mirpur and Muzaffarabad. There are also graphite deposits at Mohriwali. There are also reservoirs of low-grade coal, chalk, bauxite, and zircon. Local household industries produce carved wooden objects, textiles, and dhurrie carpets. There is also an arts and crafts industry that produces such cultural goods as namdas, shawls, pashmina, pherans, Papier-mâché, basketry copper, rugs, wood carving, silk and woolen clothing, patto, carpets, namda gubba, and silverware. Agricultural goods produced in the region include mushrooms, honey, walnuts, apples, cherries, medicinal herbs and plants, resin, deodar, kail, chir, fir, maple, and ash timber. The migration to UK was accelerated and by the completion of Mangla Dam in 1967 the process of 'chain migration' became in full flow. Today, remittances from British Mirpuri community make a critical role in AJK's economy. In the mid-1950s various economic and social development processes were launched in Azad Kashmir. In the 1960s, with the construction of the Mangla Dam in Mirpur District, the Azad Jammu and Kashmir Government began to receive royalties from the Pakistani government for the electricity that the dam provided to Pakistan. During the mid-2000s, a multibillion-dollar reconstruction began in the aftermath of the 2005 Kashmir earthquake. In addition to agriculture, textiles, and arts and crafts, remittances have played a major role in the economy of Azad Kashmir. One analyst estimated that the figure for Azad Kashmir was 25.1% in 2001. With regard to annual household income, people living in the higher areas are more dependent on remittances than are those living in the lower areas. In the latter part of 2006, billions of dollars for development were mooted by international aid agencies for the reconstruction and rehabilitation of earthquake-hit zones in Azad Kashmir, though much of that amount was subsequently lost in bureaucratic channels, leading to considerable delays in help getting to the most neediest. Hundreds of people continued to live in tents long after the earthquake. A land-use plan for the city of Muzaffarabad was prepared by the Japan International Cooperation Agency. Tourist destinations in the area include the following: The literacy rate in Azad Kashmir was 62% in 2004, higher than in any other region of Pakistan. However, only 2.2% were graduates, compared to the average of 2.9% for Pakistan. The following is a list of universities recognised by Higher Education Commission of Pakistan (HEC): * Granted university status. The following is a list of undergraduate medical institutions recognised by Pakistan Medical and Dental Council (PMDC) . Football, cricket and volleyball are very popular in Azad Kashmir. Many tournaments are also held throughout the year and in the holy month of Ramazan, night-time flood-lit tournaments are also organised. Azad Kashmir has a T20 cricket team in Pakistan's T20 domestic tournament New Mirpur City has a cricket stadium (Quaid-e-Azam Stadium) which has been taken over by the Pakistan Cricket Board for renovation to bring it up to International standards. There is also a cricket stadium in Muzaffarabad with the capacity of 8,000 people. This stadium has hosted 8 matches of Inter-District Under 19 Tournament 2013. There are also registered football clubs:
https://en.wikipedia.org/wiki?curid=2745
Arabian Sea The Arabian Sea is a region of the northern Indian Ocean bounded on the north by Pakistan and Iran, on the west by the Gulf of Aden, Guardafui Channel and the Arabian Peninsula, on the southeast by the Laccadive Sea, on the southwest by the Somali Sea, and on the east by India. Its total area is 3,862,000 km2 (1,491,000 sq mi) and its maximum depth is 4,652 meters (15,262 ft). The Gulf of Aden in the west connects the Arabian Sea to the Red Sea through the strait of Bab-el-Mandeb, and the Gulf of Oman is in the northwest, connecting it to the Persian Gulf. The Arabian Sea has been crossed by many important marine trade routes since the third or second millennium BCE. Major seaports include Kandla Port, Okha Port, Mumbai Port, Nhava Sheva Port (Navi Mumbai), Mormugão Port (Goa), New Mangalore Port and Kochi Port in India, the Port of Karachi, Port Qasim, and the Gwadar Port in Pakistan, Chabahar Port in Iran and the Port of Salalah in Salalah, Oman. The largest islands in the Arabian Sea include Socotra (Yemen), Masirah Island (Oman), Lakshadweep (India) and Astola Island (Pakistan). The Arabian Sea's surface area is about . The maximum width of the Sea is approximately , and its maximum depth is . The biggest river flowing into the Sea is the Indus River. The Arabian Sea has two important branches — the Gulf of Aden in the southwest, connecting with the Red Sea through the strait of Bab-el-Mandeb; and the Gulf of Oman to the northwest, connecting with the Persian Gulf. There are also the gulfs of Khambhat and Kutch on the Indian Coast. The countries with coastlines on the Arabian Sea are Yemen, Oman, Pakistan, Iran, India and the Maldives. The International Hydrographic Organization defines the limits of the Arabian Sea as follows: The Arabian sea was called Erythraean Sea in Roman times. The Arabian Sea historically and geographically has been referred to with many different names by Arabian and European geographers and travelers, including Sindhu Sagar, Arabbi Samudra, and the Erythraean Sea, there are many official maps showing the different names of arabian sea :Imperia Persarum et Macedonum, The Arabian Sea has been an important marine trade route since the era of the "coastal sailing vessels" from possibly as early as the 3rd millennium BCE, certainly the late 2nd millennium BCE through the later days known as the Age of Sail. By the time of Julius Caesar, several well-established combined land-sea trade routes depended upon water transport through the Sea around the rough inland terrain features to its north. These routes usually began in the Far East or down river from Madhya Pradesh with transshipment via historic Bharuch (Bharakuccha), traversed past the inhospitable coast of today's Iran then split around Hadhramaut into two streams north into the Gulf of Aden and thence into the Levant, or south into Alexandria via Red Sea ports such as Axum. Each major route involved transhipping to pack animal caravan, travel through desert country and risk of bandits and extortionate tolls by local potentates. This southern coastal route past the rough country in the southern Arabian Peninsula was significant, and the Egyptian Pharaohs built several shallow canals to service the trade, one more or less along the route of today's Suez canal, and another from the Red Sea to the Nile River, both shallow works that were swallowed up by huge sand storms in antiquity. Later the kingdom of Axum arose in Ethiopia to rule a mercantile empire rooted in the trade with Europe via Alexandria. The Port of Karachi is Pakistan's largest and busiest seaport. It is located between the Karachi towns of Kiamari and Saddar. The Gwadar Port is a warm-water, deep-sea port situated at Gwadar in Balochistan, Pakistan at the apex of the Arabian Sea and at the entrance of the Persian Gulf, about 460 km west of Karachi and approximately 75 km (47 mi) east of Pakistan's border with Iran. The port is located on the eastern bay of a natural hammerhead-shaped peninsula jutting out into the Arabian Sea from the coastline. Port of Salalah in Salalah, Oman is also a major port in the area. The International Task Force often uses the port as a base. There is a significant number of warships of all nations coming in and out of the port, which makes it a very safe bubble. The port handled just under 3.5m teu (twenty-foot equivalent unit, a measure used for capacity in container transportation) in 2009. Jawaharlal Nehru Port in Mumbai is the largest port in the Arabian Sea, and the largest container port in India. Major Indian ports in the Arabian Sea are Mundra Port, Kandla Port, Nava Sheva, Kochi Port, Mumbai Port, and Mormugão. There are several islands in the Arabian Sea, with the most important ones being Lakshadweep Islands (India), Socotra (Yemen), Masirah (Oman) and Astola Island (Pakistan). The Lakshadweep Islands (formerly known as the Laccadive, Minicoy, and Aminidivi Islands) is a group of islands in the Laccadive Sea region of Arabian Sea, 200 to 440 km (120 to 270 mi) off the southwestern coast of India. The archipelago is a Union Territory and is governed by the Union Government of India. The islands form the smallest Union Territory of India: their total surface area is just . The lagoon area covers about , the territorial waters area and the exclusive economic zone area . The islands are the northernmost of the Lakshadweep-Maldives-Chagos group of islands. Astola Island, also known as "Jezira Haft Talar" in Balochi, or 'Island of the Seven Hills', is a small, uninhabited island in the northern tip of the Arabian Sea in Pakistan's territorial waters. Socotra, also spelled "Soqotra", is the largest island, being part of a small archipelago of four islands. It lies some east of the Horn of Africa and south of the Arabian Peninsula. Masirah is an island off the east coast of Oman. The dead zone is an area in the Gulf of Oman that is completely depleted of oxygen, as a result of which it does not support marine life. It is the world's largest-known dead zone with an area larger than that of Scotland.
https://en.wikipedia.org/wiki?curid=2747
Aspartame Aspartame is an artificial non-saccharide sweetener 200 times sweeter than sucrose, and is commonly used as a sugar substitute in foods and beverages. It is a methyl ester of the aspartic acid/phenylalanine dipeptide with the trade names, "NutraSweet", "Equal", and "Canderel". Aspartame was first made in 1965 and approved for use in food products by the U.S. Food and Drug Administration (FDA) in 1981. Aspartame is one of the most rigorously tested food ingredients. Reviews by over 100 governmental regulatory bodies found the ingredient safe for consumption at current levels. As of 2018, several reviews of clinical trials showed that using aspartame in place of sugar reduces calorie intake and body weight in adults and children. Aspartame is around 180 to 200 times as sweet as sucrose (table sugar). Due to this property, even though aspartame produces of energy per gram when metabolized, the quantity of aspartame needed to produce a sweet taste is so small that its caloric contribution is negligible. The taste of aspartame and other artificial sweeteners differs from that of table sugar in the times of onset and how long the sweetness lasts, though aspartame comes closest to sugar's taste profile among approved artificial sweeteners. The sweetness of aspartame lasts longer than that of sucrose, so it is often blended with other artificial sweeteners such as acesulfame potassium to produce an overall taste more like that of sugar. Like many other peptides, aspartame may hydrolyze (break down) into its constituent amino acids under conditions of elevated temperature or high pH. This makes aspartame undesirable as a baking sweetener, and prone to degradation in products hosting a high pH, as required for a long shelf life. The stability of aspartame under heating can be improved to some extent by encasing it in fats or in maltodextrin. The stability when dissolved in water depends markedly on pH. At room temperature, it is most stable at pH 4.3, where its half-life is nearly 300 days. At pH 7, however, its half-life is only a few days. Most soft-drinks have a pH between 3 and 5, where aspartame is reasonably stable. In products that may require a longer shelf life, such as syrups for fountain beverages, aspartame is sometimes blended with a more stable sweetener, such as saccharin. Descriptive analyses of solutions containing aspartame report a sweet aftertaste as well as bitter and off-flavor aftertastes. In products such as powdered beverages, the amine in aspartame can undergo a Maillard reaction with the aldehyde groups present in certain aroma compounds. The ensuing loss of both flavor and sweetness can be prevented by protecting the aldehyde as an acetal. The safety of aspartame has been studied since its discovery and is one of the most rigorously tested food ingredients. Aspartame has been deemed safe for human consumption by over 100 regulatory agencies in their respective countries, including the United States Food and Drug Administration, UK Food Standards Agency, the European Food Safety Authority (EFSA), Health Canada, Australia, and New Zealand. , reviews of clinical trials showed that using aspartame (or other non-nutritive sweeteners) in place of sugar reduces calorie intake and body weight in adults and children. A 2017 review of metabolic effects by consuming aspartame found that it did not affect blood glucose, insulin, total cholesterol, triglycerides, calorie intake, or body weight, while high-density lipoprotein levels were higher. High levels of the naturally occurring essential amino acid phenylalanine are a health hazard to those born with phenylketonuria (PKU), a rare inherited disease that prevents phenylalanine from being properly metabolized. Because aspartame contains a small amount of phenylalanine, foods containing aspartame sold in the United States must state: "Phenylketonurics: Contains Phenylalanine" on product labels. In the UK, foods that contain aspartame are required by the Food Standards Agency to list the substance as an ingredient, with the warning, "Contains a source of phenylalanine". Manufacturers are also required to print '"with sweetener(s)" on the label close to the main product name on foods that contain "sweeteners such as aspartame" or "with sugar and sweetener(s)" on "foods that contain both sugar and sweetener". In Canada, foods that contain aspartame are required to list aspartame among the ingredients, include the amount of aspartame per serving, and state that the product contains phenylalanine. Phenylalanine is one of the essential amino acids and is required for normal growth and maintenance of life. Concerns about the safety of phenylalanine from aspartame for those without phenylketonuria center largely on hypothetical changes in neurotransmitter levels as well as ratios of neurotransmitters to each other in the blood and brain that could lead to neurological symptoms. Reviews of the literature have found no consistent findings to support such concerns, and while high doses of aspartame consumption may have some biochemical effects, these effects are not seen in toxicity studies to suggest aspartame can adversely affect neuronal function. As with methanol and aspartic acid, common foods in the typical diet such as milk, meat, and fruits, will lead to ingestion of significantly higher amounts of phenylalanine than would be expected from aspartame consumption. Reviews have found no association between aspartame and cancer. This position is supported by multiple regulatory agencies like the FDA and EFSA as well as scientific bodies such as the National Cancer Institute. The EFSA and FDA state that aspartame is safe for human consumption. Numerous allegations have been made via the Internet and in consumer magazines purporting neurotoxic effects of aspartame leading to neurological or psychiatric symptoms such as seizures, headaches, and mood changes. Review of the biochemistry of aspartame has found no evidence that low doses consumed would plausibly lead to neurotoxic effects. Comprehensive reviews have not found any evidence for aspartame as a cause for these symptoms. A review of studies on children did not show any significant findings for safety concerns with regard to neuropsychiatric conditions such as panic attacks, mood changes, hallucinations, ADHD, or seizures by consuming aspartame. Headaches are the most common symptom reported by consumers. While one small review noted aspartame is likely one of many dietary triggers of migraines, in a list that includes "cheese, chocolate, citrus fruits, hot dogs, monosodium glutamate, aspartame, fatty foods, ice cream, caffeine withdrawal, and alcoholic drinks, especially red wine and beer," other reviews have noted conflicting studies about headaches and still more reviews lack any evidence and references to support this claim. The perceived sweetness of aspartame (and other sweet substances like acesulfame K) in humans is due to its binding of the heterodimer G protein-coupled receptor formed by the proteins TAS1R2 and TAS1R3. Aspartame is rapidly hydrolyzed in the small intestines. Even with ingestion of very high doses of aspartame (over 200 mg/kg), no aspartame is found in the blood due to the rapid breakdown. Upon ingestion, aspartame breaks down into residual components, including aspartic acid, phenylalanine, methanol, and further breakdown products including formaldehyde and formic acid. Human studies show that formic acid is excreted faster than it is formed after ingestion of aspartame. In some fruit juices, "higher" concentrations of methanol can be found than the amount produced from aspartame in beverages. Aspartame's major decomposition products are its cyclic dipeptide (in a 2,5-diketopiperazine, or DKP, form), the non-esterified dipeptide (aspartylphenylalanine), and its constituent components, phenylalanine, aspartic acid, and methanol. At 180 °C, aspartame undergoes decomposition to form a diketopiperazine derivative. Aspartic acid (aspartate) is one of the most common amino acids in the typical diet. As with methanol and phenylalanine, intake of aspartic acid from aspartame is less than would be expected from other dietary sources. At the 90th percentile of intake, aspartame provides only between 1% and 2% of the daily intake of aspartic acid. There has been some speculation that aspartame, in conjunction with other amino acids like glutamate, may lead to excitotoxicity, inflicting damage on brain and nerve cells. However, clinical studies have shown no signs of neurotoxic effects, and studies of metabolism suggest it is not possible to ingest enough aspartic acid and glutamate through food and drink to levels that would be expected to be toxic. The methanol produced by the metabolism of aspartame is absorbed and quickly converted into formaldehyde and then completely oxidized to formic acid. The methanol from aspartame is unlikely to be a safety concern for several reasons. Fruit juices and citrus fruits contain methanol, and there are other dietary sources for methanol such as fermented beverages and the amount of methanol produced from aspartame-sweetened foods and beverages is likely to be less than that from these and other sources that are already in people's diets. With regard to formaldehyde, it is rapidly converted in the body, and the amounts of formaldehyde from the metabolism of aspartame are trivial when compared to the amounts produced routinely by the human body and from other foods and drugs. At the highest expected human doses of consumption of aspartame, there are no increased blood levels of methanol or formic acid, and ingesting aspartame at the 90th percentile of intake would produce 25 times less methanol than what would be considered toxic. Aspartame is a methyl ester of the dipeptide of the natural amino acids L-aspartic acid and L-phenylalanine. Under strongly acidic or alkaline conditions, aspartame may generate methanol by hydrolysis. Under more severe conditions, the peptide bonds are also hydrolyzed, resulting in free amino acids. While known aspects of synthesis are covered by patents, many details are proprietary. Two approaches to synthesis are used commercially. In the chemical synthesis, the two carboxyl groups of aspartic acid are joined into an anhydride, and the amino group is protected with a formyl group as the formamide, by treatment of aspartic acid with a mixture of formic acid and acetic anhydride. Phenylalanine is converted to its methyl ester and combined with the "N"-formyl aspartic anhydride; then the protecting group is removed from aspartic nitrogen by acid hydrolysis. The drawback of this technique is that a byproduct, the bitter-tasting β-form, is produced when the wrong carboxyl group from aspartic acid anhydride links to phenylalanine, with desired and undesired isomer forming in a 4:1 ratio. A process using an enzyme from "Bacillus thermoproteolyticus" to catalyze the condensation of the chemically altered amino acids will produce high yields without the β-form byproduct. A variant of this method, which has not been used commercially, uses unmodified aspartic acid, but produces low yields. Methods for directly producing aspartyl-phenylalanine by enzymatic means, followed by chemical methylation, have also been tried, but not scaled for industrial production. The acceptable daily intake (ADI) value for aspartame, as well as other food additives studied, is defined as the "amount of a food additive, expressed on a body weight basis, that can be ingested daily over a lifetime without appreciable health risk." The Joint FAO/WHO Expert Committee on Food Additives (JECFA) and the European Commission's Scientific Committee on Food has determined this value is 40 mg/kg of body weight for aspartame, while FDA has set its ADI for aspartame at 50 mg/kg. The primary source for exposure to aspartame in the United States is diet soft drinks, though it can be consumed in other products, such as pharmaceutical preparations, fruit drinks, and chewing gum among others in smaller quantities. A 12 US fluid ounce (355 ml) can of diet soda contains of aspartame, and for a adult, it takes approximately 21 cans of diet soda daily to consume the of aspartame that would surpass the FDA's 50 milligrams per kilogram of body weight ADI of aspartame from diet soda alone. Reviews have analyzed studies which have looked at the consumption of aspartame in countries worldwide, including the United States, countries in Europe, and Australia, among others. These reviews have found that even the high levels of intake of aspartame, studied across multiple countries and different methods of measuring aspartame consumption, are well below the ADI for safe consumption of aspartame. Reviews have also found that populations that are believed to be especially high consumers of aspartame such as children and diabetics are below the ADI for safe consumption, even considering extreme worst-case scenario calculations of consumption. In a report released on 10 December 2013, the EFSA said that, after an extensive examination of evidence, it ruled out the "potential risk of aspartame causing damage to genes and inducing cancer," and deemed the amount found in diet sodas safe to consume. Aspartame was discovered in 1965 by James M. Schlatter, a chemist working for G.D. Searle & Company. Schlatter had synthesized aspartame as an intermediate step in generating a tetrapeptide of the hormone gastrin, for use in assessing an anti-ulcer drug candidate. He discovered its sweet taste when he licked his finger, which had become contaminated with aspartame, to lift up a piece of paper. Torunn Atteraas Garin participated in the development of aspartame as an artificial sweetener. In 1975, prompted by issues regarding Flagyl and Aldactone, a U.S. FDA task force team reviewed 25 studies submitted by the manufacturer, including 11 on aspartame. The team reported "serious deficiencies in Searle's operations and practices". The FDA sought to authenticate 15 of the submitted studies against the supporting data. In 1979, the Center for Food Safety and Applied Nutrition (CFSAN) concluded, since many problems with the aspartame studies were minor and did not affect the conclusions, the studies could be used to assess aspartame's safety. In 1980, the FDA convened a Public Board of Inquiry (PBOI) consisting of independent advisors charged with examining the purported relationship between aspartame and brain cancer. The PBOI concluded aspartame does not cause brain damage, but it recommended against approving aspartame at that time, citing unanswered questions about cancer in laboratory rats. Citing data from a Japanese study that had not been available to the members of the PBOI, and after seeking advice from an expert panel that found fault with statistical analyses underlying the PBOI's hesitation, yet argued against approval, FDA commissioner Hayes approved aspartame for use in dry goods. In 1983, the FDA further approved aspartame for use in carbonated beverages, and for use in other beverages, baked goods, and confections in 1993. In 1996, the FDA removed all restrictions from aspartame, allowing it to be used in all foods. Several European Union countries approved aspartame in the 1980s, with EU-wide approval in 1994. The European Commission Scientific Committee on Food reviewed subsequent safety studies and reaffirmed the approval in 2002. The European Food Safety Authority reported in 2006 that the previously established Acceptable daily intake was appropriate, after reviewing yet another set of studies. Under the trade names Equal, NutraSweet, and Canderel, aspartame is an ingredient in approximately 6,000 consumer foods and beverages sold worldwide, including (but not limited to) diet sodas and other soft drinks, instant breakfasts, breath mints, cereals, sugar-free chewing gum, cocoa mixes, frozen desserts, gelatin desserts, juices, laxatives, chewable vitamin supplements, milk drinks, pharmaceutical drugs and supplements, shake mixes, tabletop sweeteners, teas, instant coffees, topping mixes, wine coolers and yogurt. It is provided as a table condiment in some countries. Aspartame is less suitable for baking than other sweeteners, because it breaks down when heated and loses much of its sweetness. In 1985, Monsanto Company bought G.D.Searle, and the aspartame business became a separate Monsanto subsidiary, the NutraSweet Company. In March 2000, Monsanto sold it to J.W. Childs Equity Partners II L.P. European use patents on aspartame expired starting in 1987, and the U.S. patent expired in 1992. Since then, the company has competed for market share with other manufacturers, including Ajinomoto, Merisant and the Holland Sweetener Company. Many aspects of industrial synthesis of aspartame were established by Ajinomoto. In 2004, the market for aspartame, in which Ajinomoto, the world's largest aspartame manufacturer, had a 40 percent share, was 14,000 metric tons a year, and consumption of the product was rising by 2 percent a year. Ajinomoto acquired its aspartame business in 2000 from Monsanto for $67M. In 2008, Ajinomoto sued British supermarket chain Asda, part of Wal-Mart, for a malicious falsehood action concerning its aspartame product when the substance was listed as excluded from the chain's product line, along with other "nasties". In July 2009, a British court found in favour of Asda. In June 2010, an appeals court reversed the decision, allowing Ajinomoto to pursue a case against Asda to protect aspartame's reputation. Asda said that it would continue to use the term "no nasties" on its own-label products, but the suit was settled in 2011 with Asda choosing to remove references to aspartame from its packaging. In November 2009, Ajinomoto announced a new brand name for its aspartame sweetener – AminoSweet.
https://en.wikipedia.org/wiki?curid=2752
AutoCAD AutoCAD is a commercial computer-aided design (CAD) and drafting software application. Developed and marketed by Autodesk, AutoCAD was first released in December 1982 as a desktop app running on microcomputers with internal graphics controllers. Before AutoCAD was introduced, most commercial CAD programs ran on mainframe computers or minicomputers, with each CAD operator (user) working at a separate graphics terminal. Since 2010, AutoCAD was released as a mobile- and web app as well, marketed as AutoCAD 360. AutoCAD is used in industry, by architects, project managers, engineers, graphic designers, city planners and other professionals. It was supported by 750 training centers worldwide in 1994. AutoCAD was derived from a program that began in 1977, and then released in 1979 called Interact CAD, also referred to in early Autodesk documents as MicroCAD, which was written prior to Autodesk's (then Marinchip Software Partners) formation by Autodesk cofounder Michael Riddle. The first version by Autodesk was demonstrated at the 1982 Comdex and released that December. AutoCAD supported CP/M-80 computers. As Autodesk's flagship product, by March 1986 AutoCAD had become the most ubiquitous CAD program worldwide. The 2020 release marked the 34th major release of AutoCAD for Windows. The 2019 release marked the ninth consecutive year of AutoCAD for Mac. The native file format of AutoCAD is .dwg. This and, to a lesser extent, its interchange file format "DXF", have become de facto, if proprietary, standards for CAD data interoperability, particularly for 2D drawing exchange. AutoCAD has included support for .dwf, a format developed and promoted by Autodesk, for publishing CAD data. ESRI ArcMap 10 permits export as AutoCAD drawing files. Civil 3D permits export as AutoCAD objects and as LandXML. Third-party file converters exist for specific formats such as Bentley MX GENIO Extension, PISTE Extension (France), ISYBAU (Germany), OKSTRA and Microdrainage (UK); also, conversion of .pdf files is feasible, however, the accuracy of the results may be unpredictable or distorted. For example, jagged edges may appear. Several vendors provide online conversions for free such as Cometdocs. Auto CAD and AutoCAD LT are available for English, German, French, Italian, Spanish, Japanese, Korean, Chinese Simplified, Chinese Traditional, Brazilian Portuguese, Russian, Czech, Polish and Hungarian (also through additional language packs). The extent of localization varies from full translation of the product to documentation only. The AutoCAD command set is localized as a part of the software localization. AutoCAD supports a number of APIs for customization and automation. These include AutoLISP, Visual LISP, VBA, .NET and ObjectARX. ObjectARX is a C++ class library, which was also the base for: There are a large number of AutoCAD plugins (add-on applications) available on the application store Autodesk Exchange Apps. AutoCAD's DXF, drawing exchange format, allows importing and exporting drawing information. Autodesk has also developed a few vertical programs for discipline-specific enhancements such as: Since AutoCAD 2019 several verticals are included with AutoCAD subscription as Industry-Specific Toolset. For example, AutoCAD Architecture (formerly Architectural Desktop) permits architectural designers to draw 3D objects, such as walls, doors, and windows, with more intelligent data associated with them rather than simple objects, such as lines and circles. The data can be programmed to represent specific architectural products sold in the construction industry, or extracted into a data file for pricing, materials estimation, and other values related to the objects represented. Additional tools generate standard 2D drawings, such as elevations and sections, from a 3D architectural model. Similarly, Civil Design, Civil Design 3D, and Civil Design Professional support data-specific objects facilitating easy standard civil engineering calculations and representations. Softdesk Civil was developed as an AutoCAD add-on by a company in New Hampshire called Softdesk (originally DCA). Softdesk was acquired by Autodesk, and Civil became Land Development Desktop (LDD), later renamed Land Desktop. Civil 3D was later developed and Land Desktop was retired. AutoCAD LT is the lower-cost version of AutoCAD, with reduced capabilities, first released in November 1993. Autodesk developed AutoCAD LT to have an entry-level CAD package to compete in the lower price level. Priced at $495, it became the first AutoCAD product priced below $1000. It was sold directly by Autodesk and in computer stores unlike the full version of AutoCAD, which must be purchased from official Autodesk dealers. AutoCAD LT 2015 introduced Desktop Subscription from $360 per year; as of 2018, three subscription plans were available, from $50 a month to a 3-year, $1170 license. While there are hundreds of small differences between the full AutoCAD package and AutoCAD LT, there are a few recognized major differences in the software's features: AutoCAD 360 (formerly AutoCAD WS) is an account-based mobile and web application enabling registered users to view, edit, and share AutoCAD files via mobile device and web using a limited AutoCAD feature set — and using cloud-stored drawing files. The program, which is an evolution and combination of previous products, uses a freemium business model with a free plan and two paid levels — marketed as Pro ($4.99 monthly or $49.99 yearly) and Pro Plus ($99.99 yearly) — including various amounts of storage, tools, and online access to drawings. 360 includes new features such as a "Smart Pen" mode and linking to third-party cloud-based storage such as Dropbox. Having evolved from Flash-based software, AutoCAD 360 uses HTML5 browser technology available in newer browsers including Firefox and Google Chrome. AutoCAD WS began with a version for the iPhone and subsequently expanded to include versions for the iPod Touch, iPad, Android phones, and Android tablets. Autodesk released the iOS version in September 2010, following with the Android version on April 20, 2011. The program is available via download at no cost from the App Store (iOS), Google Play (Android) and Amazon Appstore (Android). In its initial iOS version, AutoCAD WS supported drawing of lines, circles, and other shapes; creation of text and comment boxes; and management of color, layer, and measurements — in both landscape and portrait modes. Version 1.3, released August 17, 2011, added support for unit typing, layer visibility, area measurement and file management. The Android variant includes the iOS feature set along with such unique features as the ability to insert text or captions by voice command as well as manually. Both Android and iOS versions allow the user to save files on-line — or off-line in the absence of an Internet connection. In 2011, Autodesk announced plans to migrate the majority of its software to "the cloud", starting with the AutoCAD WS mobile application. According to a 2013 interview with Ilai Rotbaein, an AutoCAD WS Product Manager for Autodesk, the name "AutoCAD WS" had no definitive meaning, and was interpreted variously as "Autodesk Web Service", "White Sheet" or "Work Space." In 2013, "AutoCAD WS" was renamed to "AutoCAD 360". AutoCAD is licensed, for free, to students, educators, and educational institutions, with a 36-month renewable license available. The student version of AutoCAD is functionally identical to the full commercial version, with one exception: DWG files created or edited by a student version have an internal bit-flag set (the "educational flag"). When such a DWG file is printed by any version of AutoCAD (commercial or student) older than AutoCAD 2014 SP1 or AutoCAD 2019 and newer, the output includes a plot stamp/banner on all four sides. Objects created in the Student Version cannot be used for commercial use. Student Version objects "infect" a commercial version DWG file if they are imported in versions older than AutoCAD 2015 or newer than AutoCAD 2018. AutoCAD is a software package created for Windows and usually, any new AutoCAD version supports the current Windows version and some older ones. AutoCAD 2016 to 2020 support Windows 7 up to Windows 10. Autodesk stopped supporting Apple's Macintosh computers in 1994. Over the next several years, no compatible versions for the Mac were released. In 2010 Autodesk announced that it would once again support Apple's Mac OS X software in the future. Most of the features found in the 2012 Windows version can be found in the 2012 Mac version. The main difference is the user interface and layout of the program. The interface is designed so that users who are already familiar with Apple's macOS software will find it similar to other Mac applications. Autodesk has also built-in various features in order to take full advantage of Apple's Trackpad capabilities as well as the full-screen mode in Apple's OS X Lion. AutoCAD 2012 for Mac supports both the editing and saving of files in DWG formatting that will allow the file to be compatible with other platforms besides the OS X. AutoCAD 2019 for Mac requires Apple OS X v10.11 (El Capitan) or later. AutoCAD LT 2013 was available through the Mac App Store for $899.99. The full-featured version of AutoCAD 2013 for Mac, however, wasn't available through the Mac App Store due to the price limit of $999 set by Apple. AutoCAD 2014 for Mac was available for purchase from Autodesk's Web site for $4,195 and AutoCAD LT 2014 for Mac for $1,200, or from an Autodesk Authorized Reseller. The latest version available for Mac is AutoCAD 2019 as of December 2018. As of 2019, no AutoCAD release is listed for purchase on the Mac App Store.
https://en.wikipedia.org/wiki?curid=2753
AutoCAD DXF AutoCAD DXF (Drawing Interchange Format, or Drawing Exchange Format) is a CAD data file format developed by Autodesk for enabling data interoperability between AutoCAD and other programs. DXF was originally introduced in December 1982 as part of AutoCAD 1.0, and was intended to provide an exact representation of the data in the AutoCAD native file format, DWG (Drawing), for which Autodesk for many years did not publish specifications. Because of this, correct imports of DXF files have been difficult. Autodesk now publishes the DXF specifications as a PDF on its website. Versions of AutoCAD from Release 10 (October 1988) and up support both ASCII and binary forms of DXF. Earlier versions support only ASCII. As AutoCAD has become more powerful, supporting more complex object types, DXF has become less useful. Certain object types, including ACIS solids and regions, are not documented. Other object types, including AutoCAD 2006's dynamic blocks, and all of the objects specific to the vertical market versions of AutoCAD, are partially documented, but not well enough to allow other developers to support them. For these reasons many CAD applications use the DWG format which can be licensed from Autodesk or non-natively from the Open Design Alliance. DXF coordinates are always without dimensions so that the reader or user needs to know the drawing unit or has to extract it from the textual comments in the sheets. ASCII versions of DXF can be read with any text editor. The basic organization of a DXF file is as follows: The data format of a DXF is called a "tagged data" format, which "means that each data element in the file is preceded by an integer number that is called a group code. A group code's value indicates what type of data element follows. This value also indicates the meaning of a data element for a given object (or record) type. Virtually all user-specified information in a drawing file can be represented in DXF format."
https://en.wikipedia.org/wiki?curid=2754
Asexual reproduction Asexual reproduction is a type of reproduction which does not involve the fusion of gametes or change in the number of chromosomes. In asexual reproduction the new plants are produced from a single parent. The offspring that arise by asexual reproduction from a single cell or from a multicellular organism inherit the genes of that parent. Asexual reproduction is the primary form of reproduction for single-celled organisms such as archaea and bacteria. Many multicellular animals, plants and fungi can also reproduce asexually. While all prokaryotes reproduce without the formation and fusion of gametes, mechanisms for lateral gene transfer such as conjugation, transformation and transduction can be likened to sexual reproduction in the sense of genetic recombination in meiosis. Prokaryotes (Archaea and Bacteria) reproduce asexually through binary fission, in which the parent organism divides in two to produce two genetically identical daughter organisms. Eukaryotes (such as protists and unicellular fungi) may reproduce in a functionally similar manner by mitosis; most of these are also capable of sexual reproduction. Multiple fission at the cellular level occurs in many protists, e.g. sporozoans and algae. The nucleus of the parent cell divides several times by mitosis, producing several nuclei. The cytoplasm then separates, creating multiple daughter cells. In apicomplexans, multiple fission, or schizogony appears either as merogony, sporogony or gametogony. Merogony results in merozoites, which are multiple daughter cells, that originate within the same cell membrane, sporogony results in sporozoites, and gametogony results in microgametes. Some cells divide by budding (for example baker's yeast), resulting in a "mother" and a "daughter" cell that is initially smaller than the parent. Budding is also known on a multicellular level; an animal example is the hydra, which reproduces by budding. The buds grow into fully matured individuals which eventually break away from the parent organism. Internal budding is a process of asexual reproduction, favoured by parasites such as "Toxoplasma gondii". It involves an unusual process in which two ("endodyogeny") or more ("endopolygeny") daughter cells are produced inside a mother cell, which is then consumed by the offspring prior to their separation. Also, budding (external or internal) occurs in some worms like "Taenia" or "Echinococcus"; these worms produce cysts and then produce (invaginated or evaginated) protoscolex with budding. Sometimes, the process of budding is so fast that the daughter cell remains attached to the mother cell. This creates a chain of buds. Vegetative propagation is a type of asexual reproduction found in plants where new individuals are formed without the production of seeds or spores and thus without syngamy or meiosis. Examples of vegetative reproduction include the formation of miniaturized plants called plantlets on specialized leaves, for example in kalanchoe ("Bryophyllum daigremontianum") and many produce new plants from rhizomes or stolon (for example in strawberry). Other plants reproduce by forming bulbs or tubers (for example tulip bulbs and "Dahlia" tubers). Some plants produce adventitious shoots and may form a clonal colony. In these examples, all the individuals are clones, and the clonal population may cover a large area. Many multicellular organisms form spores during their biological life cycle in a process called "sporogenesis". Exceptions are animals and some protists, which undergo "meiosis" immediately followed by fertilization. Plants and many algae on the other hand undergo "sporic meiosis" where meiosis leads to the formation of haploid spores rather than gametes. These spores grow into multicellular individuals (called gametophytes in the case of plants) without a fertilization event. These haploid individuals give rise to gametes through mitosis. Meiosis and gamete formation therefore occur in separate generations or "phases" of the life cycle, referred to as alternation of generations. Since sexual reproduction is often more narrowly defined as the fusion of gametes (fertilization), spore formation in plant sporophytes and algae might be considered a form of asexual reproduction (agamogenesis) despite being the result of meiosis and undergoing a reduction in ploidy. However, both events (spore formation and fertilization) are necessary to complete sexual reproduction in the plant life cycle. Fungi and some algae can also utilize true asexual spore formation, which involves mitosis giving rise to reproductive cells called mitospores that develop into a new organism after dispersal. This method of reproduction is found for example in conidial fungi and the red algae "Polysiphonia", and involves sporogenesis without meiosis. Thus the chromosome number of the spore cell is the same as that of the parent producing the spores. However, mitotic sporogenesis is an exception and most spores, such as those of plants, most Basidiomycota, and many algae, are produced by meiosis. Fragmentation is a form of asexual reproduction where a new organism grows from a fragment of the parent. Each fragment develops into a mature, fully grown individual. Fragmentation is seen in many organisms. Animals that reproduce asexually include planarians, many annelid worms including polychaetes and some oligochaetes, turbellarians and sea stars. Many fungi and plants reproduce asexually. Some plants have specialized structures for reproduction via fragmentation, such as "gemmae" in liverworts. Most lichens, which are a symbiotic union of a fungus and photosynthetic algae or cyanobacteria, reproduce through fragmentation to ensure that new individuals contain both symbionts. These fragments can take the form of "soredia", dust-like particles consisting of fungal hyphen wrapped around photobiont cells. Clonal Fragmentation in multicellular or colonial organisms is a form of asexual reproduction or cloning where an organism is split into fragments. Each of these fragments develop into mature, fully grown individuals that are clones of the original organism. In echinoderms, this method of reproduction is usually known as "fissiparity". Due to many environmental and epigenetic differences, clones originating from the same ancestor might actually be genetically and epigenetically different. Agamogenesis is any form of reproduction that does not involve a male gamete. Examples are parthenogenesis and apomixis. Parthenogenesis is a form of agamogenesis in which an unfertilized egg develops into a new individual. Parthenogenesis occurs naturally in many plants, invertebrates (e.g. water fleas, rotifers, aphids, stick insects, some ants, bees and parasitic wasps), and vertebrates (e.g. some reptiles, amphibians, rarely sharks and birds). In plants, apomixis may or may not involve parthenogenesis. Apomixis in plants is the formation of a new sporophyte without fertilization. It is important in ferns and in flowering plants, but is very rare in other seed plants. In flowering plants, the term "apomixis" is now most often used for agamospermy, the formation of seeds without fertilization, but was once used to include vegetative reproduction. An example of an apomictic plant would be the triploid European dandelion. Apomixis mainly occurs in two forms: In gametophytic apomixis, the embryo arises from an unfertilized egg within a diploid embryo sac that was formed without completing meiosis. In nucellar embryony, the embryo is formed from the diploid nucellus tissue surrounding the embryo sac. Nucellar embryony occurs in some citrus seeds. Male apomixis can occur in rare cases, such as the Saharan Cypress "Cupressus dupreziana", where the genetic material of the embryo are derived entirely from pollen. Some species can alternate between sexual and asexual strategies, an ability known as "heterogamy", depending on many conditions. Alternation is observed in several rotifer species (cyclical parthenogenesis e.g. in Brachionus species) and a few types of insects, such as aphids which will, under certain conditions, produce eggs that have not gone through meiosis, thus cloning themselves. The cape bee "Apis mellifera" subsp. "capensis" can reproduce asexually through a process called thelytoky. The freshwater crustacean "Daphnia" reproduces by parthenogenesis in the spring to rapidly populate ponds, then switches to sexual reproduction as the intensity of competition and predation increases. Monogonont rotifers of the genus "Brachionus" reproduce via cyclical parthenogenesis: at low population densities females produce asexually and at higher densities a chemical cue accumulates and induces the transition to sexual reproduction. Many protists and fungi alternate between sexual and asexual reproduction. A few species of amphibians, reptiles, and birds have a similar ability. The slime mold "Dictyostelium" undergoes binary fission (mitosis) as single-celled amoebae under favorable conditions. However, when conditions turn unfavorable, the cells aggregate and follow one of two different developmental pathways, depending on conditions. In the social pathway, they form a multi-cellular slug which then forms a fruiting body with asexually generated spores. In the sexual pathway, two cells fuse to form a giant cell that develops into a large cyst. When this macrocyst germinates, it releases hundreds of amoebic cells that are the product of meiotic recombination between the original two cells. The hyphae of the common mold ("Rhizopus") are capable of producing both mitotic as well as meiotic spores. Many algae similarly switch between sexual and asexual reproduction. A number of plants use both sexual and asexual means to produce new plants, some species alter their primary modes of reproduction from sexual to asexual under varying environmental conditions. In the rotifer "Brachionus calyciflorus" asexual reproduction (obligate parthenogenesis) can be inherited by a recessive allele, which leads to loss of sexual reproduction in homozygous offspring. Inheritance of asexual reproduction by a single recessive locus has also been found in the parasitoid wasp "Lysiphlebus fabarum". Asexual reproduction is found in nearly half of the animal phyla. Parthenogenesis occurs in the hammerhead shark and the blacktip shark. In both cases, the sharks had reached sexual maturity in captivity in the absence of males, and in both cases the offspring were shown to be genetically identical to the mothers. The New Mexico whiptail is another example. Some reptiles use the ZW sex-determination system, which produces either males (with ZZ sex chromosomes) or females (with ZW or WW sex chromosomes). Until 2010, it was thought that the ZW chromosome system used by reptiles was incapable of producing viable WW offspring, but a (ZW) female boa constrictor was discovered to have produced viable female offspring with WW chromosomes. The female boa could have chosen any number of male partners (and had successfully in the past) but on these occasions she reproduced asexually, creating 22 female babies with WW sex-chromosomes. Polyembryony is a widespread form of asexual reproduction in animals, whereby the fertilized egg or a later stage of embryonic development splits to form genetically identical clones. Within animals, this phenomenon has been best studied in the parasitic Hymenoptera. In the 9-banded armadillos, this process is obligatory and usually gives rise to genetically identical quadruplets. In other mammals, monozygotic twinning has no apparent genetic basis, though its occurrence is common. There are at least 10 million identical human twins and triplets in the world today. Bdelloid rotifers reproduce exclusively asexually, and all individuals in the class Bdelloidea are females. Asexuality evolved in these animals millions of years ago and has persisted since. There is evidence to suggest that asexual reproduction has allowed the animals to evolve new proteins through the Meselson effect that have allowed them to survive better in periods of dehydration. Molecular evidence strongly suggests that several species of the stick insect genus "Timema" have used only asexual (parthenogenetic) reproduction for millions of years, the longest period known for any insect. In the grass thrips genus "Aptinothrips" there have been several transitions to asexuality, likely due to different causes. A complete lack of sexual reproduction is relatively rare among multicellular organisms, particularly animals. It is not entirely understood why the ability to reproduce sexually is so common among them. Current hypotheses suggest that asexual reproduction may have short term benefits when rapid population growth is important or in stable environments, while sexual reproduction offers a net advantage by allowing more rapid generation of genetic diversity, allowing adaptation to changing environments. Developmental constraints may underlie why few animals have relinquished sexual reproduction completely in their life-cycles. Another constraint on switching from sexual to asexual reproduction would be the concomitant loss of meiosis and the protective recombinational repair of DNA damage afforded as one function of meiosis.
https://en.wikipedia.org/wiki?curid=2756
Aelbert Cuyp Aelbert Jacobsz Cuyp (20 October 1620 – 15 November 1691) was one of the leading Dutch Golden Age painters, producing mainly landscapes. The most famous of a family of painters, the pupil of his father Jacob Gerritsz Cuyp (1594–1651/52), he is especially known for his large views of Dutch riverside scenes in a golden early morning or late afternoon light. Cuyp was born in Dordrecht on 20 October 1620, and died there on 15 November 1691. Known as the Dutch equivalent of Claude Lorrain, he inherited a considerable fortune. His family were all artists, with his uncle Benjamin and grandfather Gerrit being stained glass cartoon designers. Jacob Gerritsz Cuyp, his father, was a portraitist. Cuyp's father was his first teacher and they collaborated on many paintings throughout his lifetime. Little is known about Aelbert Cuyp's life. Even Arnold Houbraken, a noted historian of Dutch Golden Age paintings and the sole authority on Cuyp for the hundred years following his death, paints a very thin biographical picture. His period of activity as a painter is traditionally limited to the two decades between 1639 and 1660, fitting within the generally accepted limits of the Dutch Golden Age's most significant period, 1640–1665. He is known to have been married to Cornelia Bosman in 1658, a date coinciding so directly with the end of his productivity as a painter that it has been accepted that his marriage played a role in the end of his artistic career. The year after his marriage Cuyp became the deacon of the reformed church. Houbraken recalled that Cuyp was a devout Calvinist and the fact that when he died, there were no paintings of other artists found in his home. The development of Cuyp, who was trained as a landscape painter, may be roughly sketched in three phases based on the painters who most influenced him during that time and the subsequent artistic characteristics that are apparent in his paintings. Generally, Cuyp learned tone from the exceptionally prolific Jan van Goyen, light from Jan Both and form from his father, Jacob Gerritsz Cuyp. Cuyp's "van Goyen phase" can be placed approximately in the early 1640s. Cuyp probably first encountered a painting by van Goyen in 1640 when van Goyen was, as Stephen Reiss points, out "at the height of [his] powers". This is noticeable in the comparison between two of Cuyp's landscape paintings inscribed 1639 where no properly formed style is apparent and the landscape backgrounds he painted two years later for two of his father's group portraits that are distinctly van Goyenesque. Cuyp took from van Goyen the straw yellow and light brown tones that are so apparent in his "Dunes" (1629) and the broken brush technique also very noticeable in that same work. This technique, a precursor to impressionism, is noted for the short brush strokes where the colors are not necessarily blended smoothly. In Cuyp's "River Scene, Two Men Conversing" (1641) both of these van Goyen-influenced stylistic elements are noticeable. The next phase in the development of Cuyp's increasingly amalgamated style is due to the influence of Jan Both. In the mid-1640s Both, a native and resident of Utrecht, had just returned to his hometown from a trip to Rome. It is around this same time that Cuyp's style changed fundamentally. In Rome, Both had developed a new style of composition due, at least in part, to his interaction with Claude Lorrain. This new style was focused on changing the direction of light in the painting. Instead of the light being placed at right angles in relation to the line of vision, Both started moving it to a diagonal position from the back of the picture. In this new form of lighting, the artist (and viewer of the painting) faced the sun more or less contre-jour. Both, and subsequently Cuyp, used the advantages of this new lighting style to alter the sense of depth and luminosity possible in a painting. To make notice of these new capabilities, much use was made of elongated shadows. Cuyp was one of the first Dutch painters to appreciate this new leap forward in style and while his own Both-inspired phase was quite short (limited to the mid-1640s) he did, more than any other contemporary Dutch artist, maximize the full chromatic scale for sunsets and sunrises. Cuyp's third stylistic phase (which occurred throughout his career) is based on the influence of his father. While it is assumed that the younger Cuyp did work with his father initially to develop rudimentary talents, Aelbert became more focused on landscape paintings while Jacob was a portrait painter by profession. As has been mentioned and as will be explained in depth below, there are pieces where Aelbert provided the landscape background for his father's portraits. What is meant by stating that Aelbert learned form from his father is that his eventual transition from a specifically landscape painter to the involvement of foreground figures is attributed to his interaction with his father Jacob. The evidence for Aelbert's evolution to foreground figure painter is in the production of some paintings from 1645–1650 featuring foreground animals that do not fit with Jacob's style. Adding to the confusion regarding Aelbert's stylistic development and the problem of attribution is of course the fact that Jacob's style was not stagnant either. Their converging styles make it difficult to exactly understand the influences each had on the other, although it is clear enough to say that Aelbert started representing large scale forms (something he had not done previously) and placing animals as the focus of his paintings (something that was specific to him). Sunlight in his paintings rakes across the panel, accentuating small bits of detail in the golden light. In large, atmospheric panoramas of the countryside, the highlights on a blade of meadow grass, the mane of a tranquil horse, the horn of a dairy cow reclining by a stream, or the tip of a peasant's hat are all caught in a bath of yellow ocher light. The richly varnished medium refracts the rays of light like a jewel as it dissolves into numerous glazed layers. Cuyp's landscapes were based on reality and on his own invention of what an enchanting landscape should be. Cuyp's drawings reveal him to be a draftsman of superior quality. Light-drenched washes of golden brown ink depict a distant view of the city of Dordrecht or Utrecht. A Cuyp drawing may look like he intended it to be a finished work of art, but it was most likely taken back to the studio and used as a reference for his paintings. Often the same section of a sketch can be found in several different pictures. Cuyp signed many of his works but rarely dated them, so that a chronology of his career has not been satisfactorily reassembled. A phenomenal number of paintings are ascribed to him, some of which are likely to be by other masters of the golden landscape, such as Abraham Calraet (1642–1722), whose initials "A.C." may be mistaken for Cuyp's. However, not everyone appreciates his work and "River Landscape" (1660), despite being widely regarded as amongst his best work, has been described as having "chocolate box blandness". At the Madrid's Thyssen-Bornemisza Museum most likely, the sole Cuyp's painting in Spanish public collections can be seen, a "Landscape with a sunset" ca. 1655 with animals. In addition to the scarcely documented and confirmed biography of Cuyp's life, and even more so than his amalgamated style from his three main influences, there are yet other factors that have led to the misattribution and confusion over Aelbert Cuyp's works for hundreds of years. His highly influenced style which incorporated Italianate lighting from Jan Both, broken brush technique and atonality from Jan van Goyen, and his ever-developing style from his father Jacob Gerritsz Cuyp was studied acutely by his most prominent follower, Abraham van Calraet. Calraet mimicked Cuyp's style, incorporating the same aspects, and produced similar landscapes to that of the latter. This made it quite difficult to tell whose paintings where whose. Adding to the confusion is the similar initials between the two and the inconsistent signing of paintings which were produced by Cuyp's studio. Although Aelbert Cuyp signed many of his paintings with a script "A. Cuyp" insignia, many paintings were left unsigned (not to mention undated) after being painted, and so a similar signature was added later on, presumably by collectors who inherited or discovered the works. Furthermore, many possible Cuyp paintings were not signed but rather initialed "A. C." referring to his name. However, Abraham van Calraet could also have used the same initials to denote a painting. Although this is unlikely (as Calraet would likely have signed his paintings "A. v.C."), this brings up the question of how paintings were signed to show ownership. Most original Cuyp paintings were signed by him, and in the script manner in which his name was inscribed. This would denote that the painting was done almost entirely by him. Conversely, paintings which came out of his workshop that were not necessarily physically worked on by Cuyp but merely overseen by him technically, were marked with A.C. to show that it was his instruction which saw the paintings' completion. Cuyp's pupils and assistants often worked on paintings in his studio, and so most of the work of a painting could be done without Cuyp ever touching the canvas, but merely approving its finality. Hence, the initialed inscription rather than a signature. Common among the mislabeled works are all of the reasons identified for misattributing Cuyp's works: the lack of biography and chronology of his works made it difficult to discern when paintings were created (making it difficult to pinpoint an artist); contentious signatures added to historians' confusion as to who actually painted the works; and the collaborations and influences by different painters makes it hard to justify that a painting is genuinely that of Aelbert Cuyp; and finally, accurate identification is made extremely difficult by the fact that this same style was copied (rather accurately) by his predecessor. As it turns out, even the historians and expert researchers have been fooled and forced to reassess their conclusions over "Cuyp's" paintings over the years. After he married Cornelia Boschman in 1658, the number of works produced by him declined almost to nothing. This may have been because his wife was a very religious woman and a not very big patron of the arts. It could also be that he became more active in the church under his wife's guidance. He was also active as deacon and elder of the Reformed Church. Though long lacking a modern biography, and with the chronology of his works rather unclear, his style emerged from various influences and makes his works distinctive, although his collaborations with his father and works by his imitators often make attributions uncertain. His follower Abraham van Calraet represents a particular problem, and the signatures on paintings are not to be relied on. The Rijksmuseum has reattributed many works to other painters; Abraham van Calraet does not even appear in a Museum catalogue until 1926, and even then he was not given his own entry.
https://en.wikipedia.org/wiki?curid=2758
Alkene In organic chemistry, an alkene is an unsaturated hydrocarbon that contains a carbon–carbon double bond. Unsaturated hydrocarbons containing two or more double bonds are known as alkadienes, alkatrienes, alkatetraenes, and so on. The words alkene and olefin are often used interchangeably (see nomenclature section below). Acyclic alkenes, with only one double bond and no other functional groups, known as mono-enes, form a homologous series of hydrocarbons with the general formula "n"2"n". Alkenes have two hydrogen atoms fewer than the corresponding alkane with the same number of carbon atoms. The simplest alkene, ethylene (C2H4), with the IUPAC name "ethene", is the organic compound produced on the largest scale industrially. Aromatic compounds are often drawn as cyclic alkenes, but their structure and properties are sufficiently distinct that they are not classified as alkenes. Alkenes having four or more carbon atoms can form diverse structural isomers. Some alkenes are also isomers of cycloalkanes. Acyclic alkene structural isomers with only one double bond follow: Many of these molecules exhibit cis-trans isomerism. There may also be chiral carbons particularly within the larger molecules (from C5). The number of potential isomers increases rapidly with additional carbon atoms. Like a single covalent bond, double bonds can be described in terms of overlapping atomic orbitals, except that, unlike a single bond (which consists of a single sigma bond), a carbon–carbon double bond consists of one sigma bond and one pi bond. This double bond is stronger than a single covalent bond (611 kJ/mol for C=C vs. 347 kJ/mol for C–C) and also shorter, with an average bond length of 1.33 ångströms (133 pm). Each carbon of the double bond uses its three sp2 hybrid orbitals to form sigma bonds to three atoms (the other carbon and two hydrogen atoms). The unhybridized 2p atomic orbitals, which lie perpendicular to the plane created by the axes of the three sp² hybrid orbitals, combine to form the pi bond. This bond lies outside the main C–C axis, with half of the bond on one side of the molecule and a half on the other. With a strength of 65 kcal/mol, the pi bond is significantly weaker than the sigma bond. Rotation about the carbon–carbon double bond is restricted because it incurs an energetic cost to break the alignment of the p orbitals on the two carbon atoms. As a consequence, substituted alkenes may exist as one of two isomers, called "cis" or "trans" isomers. More complex alkenes may be named with the "E"–"Z" notation for molecules with three or four different substituents (side groups). For example, of the isomers of butene, the two methyl groups of ("Z")-but-2-ene (a.k.a. "cis"-2-butene) appear on the same side of the double bond, and in ("E")-but-2-ene (a.k.a. "trans"-2-butene) the methyl groups appear on opposite sides. These two isomers of butene have distinct properties. Twisting to a 90° dihedral angle between two of the groups on the carbons requires less energy than the strength of a pi bond, and the bond still holds. The carbons of the double bond become pyramidal, which allows preserving some p orbital alignment—and hence pi bonding. The other two attached groups remain at a larger dihedral angle. This contradicts a common textbook assertion that the two carbons retain their planar nature when twisting, in which case the p orbitals would rotate enough away from each other to be unable to sustain a pi bond. In a 90°-twisted alkene, the p orbitals are only misaligned by 42° and the strain energy is only around 40 kcal/mol. In contrast, a fully broken pi bond has an energetic cost of around 65 kcal/mol. Some pyramidal alkenes are stable. For example, "trans"-cyclooctene is a stable strained alkene and the orbital misalignment is only 19°, despite having a significant dihedral angle of 137° (a planar system has a dihedral angle of 180°) and a degree of pyramidalization of 18°. Even "trans"-cycloheptene is stable at low temperatures. As predicted by the VSEPR model of electron pair repulsion, the molecular geometry of alkenes includes bond angles about each carbon in a double bond of about 120°. The angle may vary because of steric strain introduced by nonbonded interactions between functional groups attached to the carbons of the double bond. For example, the C–C–C bond angle in propylene is 123.9°. For bridged alkenes, Bredt's rule states that a double bond cannot occur at the bridgehead of a bridged ring system unless the rings are large enough. Following Fawcett and defining "S" as the total number of non-bridgehead atoms in the rings, bicyclic systems require "S" ≥ 7 for stability and tricyclic systems require "S" ≥ 11. Many of the physical properties of alkenes and alkanes are similar: they are colorless, nonpolar, and combustible. The physical state depends on molecular mass: like the corresponding saturated hydrocarbons, the simplest alkenes (ethene, propene, and butene) are gases at room temperature. Linear alkenes of approximately five to sixteen carbons are liquids, and higher alkenes are waxy solids. The melting point of the solids also increases with increase in molecular mass. Alkenes generally have stronger smells than their corresponding alkanes. Ethylene has a sweet and musty odor. The binding of cupric ion to the olefin in the mammalian olfactory receptor MOR244-3 is implicated in the smell of alkenes (as well as thiols). Strained alkenes, in particular, like norbornene and "trans"-cyclooctene are known to have strong, unpleasant odors, a fact consistent with the stronger π complexes they form with metal ions including copper. Alkenes are relatively stable compounds, but are more reactive than alkanes, either because of the reactivity of the carbon–carbon pi-bond or the presence of allylic CH centers. Most reactions of alkenes involve additions to this pi bond, forming new single bonds. Alkenes serve as a feedstock for the petrochemical industry because they can participate in a wide variety of reactions, prominently polymerization and alkylation. Alkenes react in many addition reactions, which occur by opening up the double-bond. Most of these addition reactions follow the mechanism of electrophilic addition. Examples are hydrohalogenation, halogenation, halohydrin formation, oxymercuration, hydroboration, dichlorocarbene addition, Simmons–Smith reaction, catalytic hydrogenation, epoxidation, radical polymerization and hydroxylation. Hydrogenation of alkenes produces the corresponding alkanes. The reaction is carried out under pressure at a temperature of 200 °C in the presence of a metallic catalyst. Common industrial catalysts are based on platinum, nickel or palladium. For laboratory syntheses, Raney nickel (an alloy of nickel and aluminium) is often employed. The simplest example of this reaction is the catalytic hydrogenation of ethylene to yield ethane: Hydration, the addition of water across the double bond of alkenes, yields alcohols. The reaction is catalyzed by phosphoric acid or sulfuric acid. This reaction is carried out on an industrial scale to produce synthetic ethanol. Alkenes can also be converted into alcohols via the oxymercuration–demercuration reaction , the hydroboration–oxidation reaction or by Mukaiyama hydration. In electrophilic halogenation the addition of elemental bromine or chlorine to alkenes yields vicinal dibromo- and dichloroalkanes (1,2-dihalides or ethylene dihalides), respectively. The decoloration of a solution of bromine in water is an analytical test for the presence of alkenes: Related reactions are also used as quantitative measures of unsaturation, expressed as the bromine number and iodine number of a compound or mixture. Hydrohalogenation is the addition of hydrogen halides, such as HCl or HI, to alkenes to yield the corresponding haloalkanes: If the two carbon atoms at the double bond are linked to a different number of hydrogen atoms, the halogen is found preferentially at the carbon with fewer hydrogen substituents. This patterns is known as Markovnikov's rule. The use of radical initiators or other compounds can lead to the opposite product result. Hydrobromic acid in particular is prone to forming radicals in the presence of various impurities or even atmospheric oxygen, leading to the reversal of the Markovnikov result: Alkenes react with water and halogens to form halohydrins by an addition reaction. Markovnikov regiochemistry and anti-stereochemistry occur. Alkenes are oxidized with many oxidizing agents. In the presence of oxygen, alkenes burn with a bright flame to produce carbon dioxide and water. Catalytic oxidation with oxygen or the reaction with percarboxylic acids yields epoxides. Reaction with ozone in ozonolysis leads to the breaking of the double bond, yielding two aldehydes or ketones. Reaction with concentrated, hot KMnO4 (or other oxidizing salts) in an acidic solution will yield ketones or carboxylic acids. This reaction can be used to determine the position of a double bond in an unknown alkene. The oxidation can be stopped at the vicinal diol rather than full cleavage of the alkene by using milder (dilute, lower temperature) KMnO4 or with osmium tetroxide or other oxidants. In the presence of an appropriate photosensitiser, such as methylene blue and light, alkenes can undergo reaction with reactive oxygen species generated by the photosensitiser, such as hydroxyl radicals, singlet oxygen or superoxide ion. These reactive photochemical intermediates are generated in what are known as Type I, Type II, and Type III processes, respectively. These various alternative processes and reactions can be controlled by choice of specific reaction conditions, leading to a wide range of products. A common example is the [4+2]-cycloaddition of singlet oxygen with a diene such as cyclopentadiene to yield an endoperoxide: Another example is the Schenck ene reaction, in which singlet oxygen reacts with an allylic structure to give a transposed allyl peroxide: Polymerization of alkenes is a reaction that yields polymers of high industrial value at great economy, such as the plastics polyethylene and polypropylene. Polymers from alkene monomers are referred to in a general way as "polyolefins" or in rare instances as "polyalkenes". A polymer from alpha-olefins is called a polyalphaolefin (PAO). Polymerization can proceed via either a free-radical or an ionic mechanism, converting the double to a single bond and forming single bonds to join the other monomers. Polymerization of conjugated dienes such as buta-1,3-diene or isoprene (2-methylbuta-1,3-diene) results in largely 1,4-addition with possibly some 1,2-addition of the diene monomer to a growing polymer chain. Alkenes are ligands in transition metal alkene complexes. The two carbon centres bond to the metal using the C–C pi- and pi*-orbitals. Mono- and diolefins are often used as ligands in stable complexes. Cyclooctadiene and norbornadiene are popular chelating agents, and even ethylene itself is sometimes used as a ligand, for example, in Zeise's salt. In addition, metal–alkene complexes are intermediates in many metal-catalyzed reactions including hydrogenation, hydroformylation, and polymerization. Alkenes are produced by hydrocarbon cracking. Raw materials are mostly natural gas condensate components (principally ethane and propane) in the US and Mideast and naphtha in Europe and Asia. Alkanes are broken apart at high temperatures, often in the presence of a zeolite catalyst, to produce a mixture of primarily aliphatic alkenes and lower molecular weight alkanes. The mixture is feedstock and temperature dependent, and separated by fractional distillation. This is mainly used for the manufacture of small alkenes (up to six carbons). Related to this is catalytic dehydrogenation, where an alkane loses hydrogen at high temperatures to produce a corresponding alkene. This is the reverse of the catalytic hydrogenation of alkenes. This process is also known as reforming. Both processes are endothermic and are driven towards the alkene at high temperatures by entropy. Catalytic synthesis of higher α-alkenes (of the type RCH=CH2) can also be achieved by a reaction of ethylene with the organometallic compound triethylaluminium in the presence of nickel, cobalt, or platinum. One of the principal methods for alkene synthesis in the laboratory is the room elimination of alkyl halides, alcohols, and similar compounds. Most common is the β-elimination via the E2 or E1 mechanism, but α-eliminations are also known. The E2 mechanism provides a more reliable β-elimination method than E1 for most alkene syntheses. Most E2 eliminations start with an alkyl halide or alkyl sulfonate ester (such as a tosylate or triflate). When an alkyl halide is used, the reaction is called a dehydrohalogenation. For unsymmetrical products, the more substituted alkenes (those with fewer hydrogens attached to the C=C) tend to predominate (see Zaitsev's rule). Two common methods of elimination reactions are dehydrohalogenation of alkyl halides and dehydration of alcohols. A typical example is shown below; note that if possible, the H is "anti" to the leaving group, even though this leads to the less stable "Z"-isomer. Alkenes can be synthesized from alcohols via dehydration, in which case water is lost via the E1 mechanism. For example, the dehydration of ethanol produces ethene: An alcohol may also be converted to a better leaving group (e.g., xanthate), so as to allow a milder "syn"-elimination such as the Chugaev elimination and the Grieco elimination. Related reactions include eliminations by β-haloethers (the Boord olefin synthesis) and esters (ester pyrolysis). Alkenes can be prepared indirectly from alkyl amines. The amine or ammonia is not a suitable leaving group, so the amine is first either alkylated (as in the Hofmann elimination) or oxidized to an amine oxide (the Cope reaction) to render a smooth elimination possible. The Cope reaction is a "syn"-elimination that occurs at or below 150 °C, for example: The Hofmann elimination is unusual in that the "less" substituted (non-Saytseff) alkene is usually the major product. Alkenes are generated from α-halosulfones in the Ramberg–Bäcklund reaction, via a three-membered ring sulfone intermediate. Another important method for alkene synthesis involves construction of a new carbon–carbon double bond by coupling of a carbonyl compound (such as an aldehyde or ketone) to a carbanion equivalent. Such reactions are sometimes called "olefinations". The most well-known of these methods is the Wittig reaction, but other related methods are known, including the Horner–Wadsworth–Emmons reaction. The Wittig reaction involves reaction of an aldehyde or ketone with a Wittig reagent (or phosphorane) of the type Ph3P=CHR to produce an alkene and Ph3P=O. The Wittig reagent is itself prepared easily from triphenylphosphine and an alkyl halide. The reaction is quite general and many functional groups are tolerated, even esters, as in this example: Related to the Wittig reaction is the Peterson olefination, which uses silicon-based reagents in place of the phosphorane. This reaction allows for the selection of "E"- or "Z"-products. If an "E"-product is desired, another alternative is the Julia olefination, which uses the carbanion generated from a phenyl sulfone. The Takai olefination based on an organochromium intermediate also delivers E-products. A titanium compound, Tebbe's reagent, is useful for the synthesis of methylene compounds; in this case, even esters and amides react. A pair of ketones or aldehydes can be deoxygenated to generate an alkene. Symmetrical alkenes can be prepared from a single aldehyde or ketone coupling with itself, using titanium metal reduction (the McMurry reaction). If different ketones are to be coupled, a more complicated method is required, such as the Barton–Kellogg reaction. A single ketone can also be converted to the corresponding alkene via its tosylhydrazone, using sodium methoxide (the Bamford–Stevens reaction) or an alkyllithium (the Shapiro reaction). Alkenes can be prepared by exchange with other alkenes, in a reaction known as olefin metathesis. In the implementations, formation of ethene gas is used to drive the reaction towards a desired product. In other cases, ethylene is used to convert internal alkenes into two terminal alkenes, a process referred to as ethenolysis. Olefin metathesis tolerates many functional groups. The method is particularly effective for the preparation of cyclic alkenes, as in this synthesis of muscone: Transition metal catalyzed hydrovinylation is another important alkene synthesis process starting from alkene itself. It involves the addition of a hydrogen and a vinyl group (or an alkenyl group) across a double bond. Reduction of alkynes is a useful method for the stereoselective synthesis of disubstituted alkenes. If the "cis"-alkene is desired, hydrogenation in the presence of Lindlar's catalyst (a heterogeneous catalyst that consists of palladium deposited on calcium carbonate and treated with various forms of lead) is commonly used, though hydroboration followed by hydrolysis provides an alternative approach. Reduction of the alkyne by sodium metal in liquid ammonia gives the "trans"-alkene. For the preparation multisubstituted alkenes, carbometalation of alkynes can give rise to a large variety of alkene derivatives. Alkenes can be synthesized from other alkenes via rearrangement reactions. Besides olefin metathesis (described above), many pericyclic reactions can be used such as the ene reaction and the Cope rearrangement. In the Diels–Alder reaction, a cyclohexene derivative is prepared from a diene and a reactive or electron-deficient alkene. Although the nomenclature is not followed widely, according to IUPAC, alkenes are acyclic hydrocarbons with one double bond between carbon centers. Olefins comprise a larger collection of cyclic and acyclic alkenes as well as dienes and polyenes. To form the root of the IUPAC names for alkenes, simply change the "-an-" infix of the parent to "-en-". For example, CH3-CH3 is the alkane "ethANe". The name of CH2=CH2 is therefore "ethENe". In higher alkenes, where isomers exist that differ in location of the double bond, the following numbering system is used: In the specific case of disubstituted alkenes where the two carbons have one substituent each, "cis"–"trans" notation may be used. If both substituents are on the same side of the bond, it is defined as "cis"-. If the substituents are on either side of the bond, it is defined as "trans"-. When an alkene has more than one substituent (especially necessary with 3 or 4 substituents), the double bond geometry is described using the labels "E" and "Z". These labels come from the German words "entgegen", meaning "opposite", and "zusammen", meaning "together". Alkenes with the higher priority groups (as determined by CIP rules) on the same side of the double bond have these groups together and are designated "Z". Alkenes with the higher priority groups on opposite sides are designated "E". A mnemonic to remember this: "Z" notation has the higher priority groups on "ze zame zide." IUPAC recognizes two names for hydrocarbon groups containing carbon–carbon double bonds, the vinyl group and the allyl group.
https://en.wikipedia.org/wiki?curid=2761
Allenes Allenes are organic compounds in which one carbon atom has double bonds with each of its two adjacent carbon centres. Allenes are classified as cumulated dienes. The parent compound of this class is propadiene, which is itself also called "allene". Compounds with an allene-type structure but with more than three carbon atoms are members of a larger class of compounds called cumulenes with X=C=Y bonding. Unlike polyenes, the two double bonds do not exhibit resonance. The central carbon atom of allenes forms two sigma bonds and two pi bonds. The central carbon is sp-hybridized, and the two terminal carbon atoms are sp2-hybridized. The bond angle formed by the three carbon atoms is 180°, indicating linear geometry for the central carbon atom. The two terminal carbon atoms are planar, and these planes are twisted 90° from each other. The structure can also be viewed as an "extended tetrahedral" with a similar shape to methane, an analogy that is continued into the stereochemical analysis of certain derivative molecules. The symmetry and isomerism of allenes has long fascinated organic chemists. For allenes with four identical substituents, there exist two twofold axes of rotation through the center carbon, inclined at 45° to the CH2 planes at either end of the molecule. The molecule can thus be thought of as a two-bladed propeller. A third twofold axis of rotation passes through the C=C=C bonds, and there is a mirror plane passing through both CH2 planes. Thus this class of molecules belong to the D2d point group. Because of the symmetry, an unsubstituted allene has no net dipole moment. An allene with two different substituents on each of the two carbon atoms will be chiral because there will no longer be any mirror planes. The chirality of these types of allenes was first predicted in 1875 by van 't Hof, but not proven experimentally until 1935. Where A has a greater priority than B according to the Cahn–Ingold–Prelog priority rules, the configuration of the axial chirality can be determined by considering the substituents on the front atom followed by the back atom when viewed along the allene axis. For the back atom, only the group of higher priority need be considered. Chiral allenes have been recently used as building blocks in the construction of organic materials with exceptional chiroptical properties. Although allenes often require specialized syntheses, the parent, propadiene is produced on a large scale as an equilibrium mixture with methylacetylene: This mixture, known as MAPP gas, is commercially available. Laboratory methods for the formation of allenes include: Numerous natural products contain the allene functional group. Noteworthy are the pigments fucoxanthin and peridinin. Little is known about the biosynthesis, although it is conjectured that they are often generated from alkyne precursors. Allenes serve as ligands in organometallic chemistry. A typical complex is Pt(η2-allene)(PPh3)2. Ni(0) reagents catalyze the cyclooligomerization of allene. Using a suitable catalyst (e.g. Wilkinson's catalyst), it is possible to reduce just one of the double bonds of an allene. Many rings or ring systems are known by semisystematic names that assume a maximum number of noncumulative bonds. To unambiguously specify derivatives that include cumulated bonds (and hence fewer hydrogens than would be expected from the skeleton), a lowercase delta may be used with a subscript indicating the number of cumulated double bonds from that atom, e.g. 8δ2-Benzocyclononene. This may be combined with the λ-convention for specifying nonstandard valency states, e.g. 2λ4δ2,5λ4δ2-Thieno[3,4-c]thiophene.
https://en.wikipedia.org/wiki?curid=2762
Alkyne In organic chemistry, an alkyne is an unsaturated hydrocarbon containing at least one carbon—carbon triple bond. The simplest acyclic alkynes with only one triple bond and no other functional groups form a homologous series with the general chemical formula "n"2"n"−2. Alkynes are traditionally known as acetylenes, although the name "acetylene" also refers specifically to C2H2, known formally as ethyne using IUPAC nomenclature. Like other hydrocarbons, alkynes are generally hydrophobic. In acetylene, the H–C≡C bond angles are 180°. By virtue of this bond angle, alkynes are rod-like. Correspondingly, cyclic alkynes are rare. Benzyne is highly unstable. The C≡C bond distance of 121 picometers is much shorter than the C=C distance in alkenes (134 pm) or the C–C bond in alkanes (153 pm). The triple bond is very strong with a bond strength of 839 kJ/mol. The sigma bond contributes 369 kJ/mol, the first pi bond contributes 268 kJ/mol and the second pi-bond of 202 kJ/mol bond strength. Bonding usually discussed in the context of molecular orbital theory, which recognizes the triple bond as arising from overlap of s and p orbitals. In the language of valence bond theory, the carbon atoms in an alkyne bond are sp hybridized: they each have two unhybridized p orbitals and two sp hybrid orbitals. Overlap of an sp orbital from each atom forms one sp–sp sigma bond. Each p orbital on one atom overlaps one on the other atom, forming two pi bonds, giving a total of three bonds. The remaining sp orbital on each atom can form a sigma bond to another atom, for example to hydrogen atoms in the parent acetylene. The two sp orbitals project on opposite sides of the carbon atom. Internal alkynes feature carbon substituents on each acetylenic carbon. Symmetrical examples include diphenylacetylene and 3-hexyne. Terminal alkynes have the formula RC2H. An example is methylacetylene (propyne using IUPAC nomenclature). Terminal alkynes, like acetylene itself, are mildly acidic, with p"K"a values of around 25. They are far more acidic than alkenes and alkanes, which have p"K"a values of around 40 and 50, respectively. The acidic hydrogen on terminal alkynes can be replaced by a variety of groups resulting in halo-, silyl-, and alkoxoalkynes. The carbanions generated by deprotonation of terminal alkynes are called acetylides. In systematic chemical nomenclature, alkynes are named with the Greek prefix system without any additional letters. Examples include ethyne or octyne. In parent chains with four or more carbons, it is necessary to say where the triple bond is located. For octyne, one can either write 3-octyne or oct-3-yne when the bond starts at the third carbon. The lowest number possible is given to the triple bond. When no superior functional groups are present, the parent chain must include the triple bond even if it is not the longest possible carbon chain in the molecule. Ethyne is commonly called by its trivial name acetylene. In chemistry, the suffix -yne is used to denote the presence of a triple bond. In organic chemistry, the suffix often follows IUPAC nomenclature. However, inorganic compounds featuring unsaturation in the form of triple bonds may be denoted by substitutive nomenclature with the same methods used with alkynes (i.e. the name of the corresponding saturated compound is modified by replacing the "-ane" ending with "-yne"). "-diyne" is used when there are two triple bonds, and so on. The position of unsaturation is indicated by a numerical locant immediately preceding the "-yne" suffix, or 'locants' in the case of multiple triple bonds. Locants are chosen so that the numbers are low as possible. "-yne" is also used as an infix to name substituent groups that are triply bound to the parent compound. Sometimes a number between hyphens is inserted before it to state which atoms the triple bond is between. This suffix arose as a collapsed form of the end of the word "acetylene". The final "-e" disappears if it is followed by another suffix that starts with a vowel. Commercially, the dominant alkyne is acetylene itself, which is used as a fuel and a precursor to other compounds, e.g., acrylates. Hundreds of millions of kilograms are produced annually by partial oxidation of natural gas: Propyne, also industrially useful, is also prepared by thermal cracking of hydrocarbons. Specialty alkynes are prepared by double dehydrohalogenation. The reaction provides a means to generate alkynes from alkenes, which are first halogenated and then dehydrohalogenates. For example, phenylacetylene can be generated from styrene by bromination followed by treatment of the resulting of styrene dibromide with sodium amide in ammonia: Via the Fritsch–Buttenberg–Wiechell rearrangement, alkynes are prepared from vinyl bromides. Alkynes can be prepared from aldehydes using the Corey–Fuchs reaction and from aldehydes or ketones by the Seyferth–Gilbert homologation. Featuring a reactive functional group, alkynes participate in many organic reactions. Such use was pioneered by Ralph Raphael, who in 1955 wrote the first book describing their versatility as intermediates in synthesis. Being more unsaturated than alkenes, alkynes characteristically undergo reactions that show that they are "doubly unsaturated". Alkynes are capable of adding two equivalents of H2, whereas an alkene adds only one equivalent. Depending on catalysts and conditions, alkynes add one or two equivalents of hydrogen. Partial hydrogenation, stopping after the addition of only one equivalent to give the alkene, is usually more desirable since alkanes are less useful: The largest scale application of this technology is the conversion of acetylene to ethylene in refineries (the steam cracking of alkanes yields a few percent acetylene, which is selectively hydrogenated in the presence of a palladium/silver catalyst). For more complex alkynes, the Lindlar catalyst is widely recommended to avoid formation of the alkane, for example in the conversion of phenylacetylene to styrene. Similarly, halogenation of alkynes gives the alkene dihalides or alkyl tetrahalides: The addition of one equivalent of H2 to internal alkynes gives cis-alkenes. Alkynes characteristically are capable of adding two equivalents of halogens and hydrogen halides. The addition of nonpolar E–H bonds across C≡C is general for silanes, boranes, and related hydrides. The hydroboration of alkynes gives vinylic boranes which oxidize to the corresponding aldehyde or ketone. In the thiol-yne reaction the substrate is a thiol. Addition of hydrogen halides has long been of interest. In the presence of mercuric chloride as a catalyst, acetylene and hydrogen chloride react to give vinyl chloride. While this method has been abandoned in the West, it remains the main production method in China. The hydration reaction of acetylene gives acetaldehyde. The reaction proceeds by formation of vinylalcohol, which undergoes tautomerizes to form the aldehyde. This reaction was once a major industrial process but it has been displaced by the Wacker process. This reaction occurs in nature, the catalyst being acetylene hydratase. The hydration of phenylacetylene gives acetophenone, and the (Ph3P)AuCH3-catalyzed hydration of 1,8-nonadiyne to 2,8-nonanedione: Alkynes undergo diverse cycloaddition reactions. The Diels–Alder reaction with 1,3-dienes give 1,4-cyclohexadienes. This general reaction has been extensively developed. Electrophilic alkynes are especially effective dienophiles. The "cycloadduct" derived from the addition of alkynes to 2-pyrone eliminates carbon dioxide to give the aromatic compound. Other specialized cycloadditions include multicomponent reactions such as alkyne trimerisation to give aromatic compounds and the [2+2+1]-cycloaddition of an alkyne, alkene and carbon monoxide in the Pauson–Khand reaction. Non-carbon reagents also undergo cyclization, e.g. Azide alkyne Huisgen cycloaddition to give triazoles. Cycloaddition processes involving alkynes are often catalyzed by metals, e.g. enyne metathesis and alkyne metathesis, which allows the scrambling of carbyne (RC) centers: Oxidative cleavage of alkynes proceeds via cycloaddition to metal oxides. Most famously, potassium permanganate converts alkynes to a pair of carboxylic acids. Terminal alkynes are readily converted to many derivatives, e.g. by coupling reactions and condensations. Via the condensation with formaldehyde and acetylene is produced butynediol: In the Sonogashira reaction, terminal alkynes are coupled with aryl or vinyl halides This reactivity exploits the fact that terminal alkynes are weak acids, whose typical p"K"a values around 25 place them between that of ammonia (35) and ethanol (16): The reactions of alkynes with certain metal cations, e.g. Ag+ and Cu+ also gives acetylides. Thus, few drops of diamminesilver(I) hydroxide (Ag(NH3)2OH) reacts with terminal alkynes signaled by formation of a white precipitate of the silver acetylide. This reactivity is the basis of alkyne coupling reactions, including the Cadiot–Chodkiewicz coupling, Glaser coupling, and the Eglinton coupling: In the Favorskii reaction and in alkynylations in general, terminal alkynes add to carbonyl compounds to give the hydroxyalkyne. Alkynes form complexes with transition metals. Such complexes occur also in metal catalyzed reactions of alkynes such as alkyne trimerization. Terminal alkynes, including acetylene itself, react with water to give aldehydes. The transformation typically requires metal catalysts to give this anti-Markovnikov addition result. According to Ferdinand Bohlmann, the first naturally occurring acetylenic compound, dehydromatricaria ester, was isolated from an "Artemisia" species in 1826. In the nearly two centuries that have followed, well over a thousand naturally occurring acetylenes have been discovered and reported. Polyynes, a subset of this class of natural products, have been isolated from a wide variety of plant species, cultures of higher fungi, bacteria, marine sponges, and corals. Some acids like tariric acid contain an alkyne group. Diynes and triynes, species with the linkage RC≡C–C≡CR′ and RC≡C–C≡C–C≡CR′ respectively, occur in certain plants ("Ichthyothere", "Chrysanthemum", "Cicuta", "Oenanthe" and other members of the Asteraceae and Apiaceae families). Some examples are cicutoxin, oenanthotoxin, and falcarinol . These compounds are highly bioactive, e.g. as nematocides. 1-Phenylhepta-1,3,5-triyne is illustrative of a naturally occurring triyne. Alkynes occur in some pharmaceuticals, including the contraceptive noretynodrel. A carbon–carbon triple bond is also present in marketed drugs such as the antiretroviral Efavirenz and the antifungal Terbinafine. Molecules called ene-diynes feature a ring containing an alkene ("ene") between two alkyne groups ("diyne"). These compounds, e.g. calicheamicin, are some of the most aggressive antitumor drugs known, so much so that the ene-diyne subunit is sometimes referred to as a "warhead". Ene-diynes undergo rearrangement via the Bergman cyclization, generating highly reactive radical intermediates that attack DNA within the tumor.
https://en.wikipedia.org/wiki?curid=2763
AbiWord AbiWord () is a free and open-source software word processor. It was originally written in C++ and since version 3 it is based on GTK+ 3. The name "AbiWord" is derived from the root of the Spanish word ""abierto"", meaning "open". AbiWord was originally started by SourceGear Corporation as the first part of a proposed AbiSuite but was adopted by open source developers after SourceGear changed its business focus and ceased development. It now runs on Linux, ReactOS, Solaris, AmigaOS 4.0 (through its Cygnix X11 engine), MeeGo (on the Nokia N9 smartphone), Maemo (on the Nokia N810) QNX and other operating systems. Development of a version for Microsoft Windows has ended due to lack of maintainers (the latest released versions are 2.8.6 and 2.9.4 beta ). The macOS port has remained on version 2.4 since 2005, although the current version does run non-natively on macOS through XQuartz. AbiWord is part of the AbiSource project which develops a number of office-related technologies. AbiWord supports both basic word processing features such as lists, indents and character formats, and more sophisticated features including tables, styles, page headers and footers, footnotes, templates, multiple views, page columns, spell checking, and grammar checking. Starting with version 2.8.0, AbiWord includes a collaboration plugin that allows integration with AbiCollab.net, a Web-based service that permits multiple users to work on the same document in real time, in full synchronization. The Presentation view of AbiWord, which permits easy display of presentations created in AbiWord on "screen-sized" pages, is another feature not often found in word processors. AbiWord generally works similarly to classic versions (pre-Office 2007) of Microsoft Word, as direct ease of migration was a high priority early goal. While many interface similarities remain, cloning the Word interface is no longer a top priority. The interface is intended to follow user interface guidelines for each respective platform. AbiWord comes with several import and export filters providing a partial support for such formats as HTML, Microsoft Word (.doc), Office Open XML (.docx), OpenDocument Text (.odt), Rich Text Format (.rtf), and text documents (.txt). LaTeX is supported for export only. Plug-in filters are available to deal with many other formats, notably WordPerfect documents. The native file format, .abw, uses XML, so as to mitigate vendor lock-in concerns with respect to interoperability and digital archiving. The AbiWord project includes a US English-only grammar checking plugin using Link Grammar. AbiWord had grammar checking before any other open source word processor, although a grammar checker was later added to OpenOffice.org. Link Grammar is both a theory of syntax and an open source parser which is now developed by the AbiWord project.
https://en.wikipedia.org/wiki?curid=2764
Ames test The Ames test is a widely employed method that uses bacteria to test whether a given chemical can cause mutations in the DNA of the test organism. More formally, it is a biological assay to assess the mutagenic potential of chemical compounds. A positive test indicates that the chemical is mutagenic and therefore may act as a carcinogen, because cancer is often linked to mutation. The test serves as a quick and convenient assay to estimate the carcinogenic potential of a compound because standard carcinogen assays on mice and rats are time-consuming (taking two to three years to complete) and expensive. However, false-positives and false-negatives are known. The procedure was described in a series of papers in the early 1970s by Bruce Ames and his group at the University of California, Berkeley. The Ames test uses several strains of the bacterium "Salmonella typhimurium" that carry mutations in genes involved in histidine synthesis. These strains are auxotrophic mutants, i.e. they require histidine for growth, but cannot produce it. The method tests the capability of the tested substance in creating mutations that result in a return to a "prototrophic" state, so that the cells can grow on a histidine-free medium. The tester strains are specially constructed to detect either frameshift (e.g. strains TA-1537 and TA-1538) or point (e.g. strain TA-1531) mutations in the genes required to synthesize histidine, so that mutagens acting via different mechanisms may be identified. Some compounds are quite specific, causing reversions in just one or two strains. The tester strains also carry mutations in the genes responsible for lipopolysaccharide synthesis, making the cell wall of the bacteria more permeable, and in the excision repair system to make the test more sensitive. Larger organisms like mammals have metabolic processes that could potentially turn a chemical considered not mutagenic into one that is or one that is considered mutagenic into one that is not. Therefore, to more effectively test a chemical compound's mutagenicity in relation to larger organisms, rat liver enzymes can be added in an attempt to replicate the metabolic processes' effect on the compound being tested in the Ames Test. Rat liver extract is optionally added to simulate the effect of metabolism, as some compounds, like benzo["a"]pyrene, are not mutagenic themselves but their metabolic products are. The bacteria are spread on an agar plate with small amount of histidine. This small amount of histidine in the growth medium allows the bacteria to grow for an initial time and have the opportunity to mutate. When the histidine is depleted only bacteria that have mutated to gain the ability to produce its own histidine will survive. The plate is incubated for 48 hours. The mutagenicity of a substance is proportional to the number of colonies observed. Mutagens identified via Ames test are also possible carcinogens, and early studies by Ames showed that 90% of known carcinogens may be identified via this test. Later studies however showed identification of 50–70% of known carcinogens. The test was used to identify a number of compounds previously used in commercial products as potential carcinogens. Examples include tris(2,3-dibromopropyl)phosphate, which was used as a flame retardant in plastic and textiles such as children's sleepwear, and furylfuramide which was used as an antibacterial additive in food in Japan in 1960s and 1970s. Furylfuramide in fact had previously passed animal test, but more vigorous tests after its identification in the Ames test showed it to be carcinogenic. Their positive tests resulted in those chemicals being withdrawn from use in consumer products. One interesting result from the Ames test is that the dose response curve using varying concentrations of chemical is almost always linear, indicating that there is no threshold concentration for mutagenesis. It therefore suggests that, as with radiations, there may be no safe threshold for chemical mutagens or carcinogens. However some proposed that organisms can tolerate low level of mutagens due to protective mechanisms such as DNA repair, and threshold may exist for certain chemical mutagens. Bruce Ames himself argued against linear dose-response extrapolation from the high dose used in carcinogenesis tests in animal systems to the lower dose of chemicals normally encountered in human exposure, as the results may be false positives due to mitogenic response caused by the artificially high dose of chemicals used in such tests. He also cautioned against the "hysteria over tiny traces of chemicals that may or may not cause cancer", that "completely drives out the major risks you should be aware of". The Ames test is often used as one of the initial screens for potential drugs to weed out possible carcinogens, and it is one of the eight tests required under the Pesticide Act (USA) and one of six tests required under the Toxic Substances Control Act (USA). "Salmonella typhimurium" is a prokaryote, therefore it is not a perfect model for humans. Rat liver S9 fraction is used to mimic the mammalian metabolic conditions so that the mutagenic potential of metabolites formed by a parent molecule in the hepatic system can be assessed; however, there are differences in metabolism between human and rat that can affect the mutagenicity of chemicals being tested. The test may therefore be improved by the use of human liver S9 fraction; its use was previously limited by its availability, but it is now available commercially and therefore may be more feasible. An adapted "in vitro" model has been made for eukaryotic cells, for example yeast. Mutagens identified in the Ames test need not necessarily be carcinogenic, and further tests are required for any potential carcinogen identified in the test. Drugs that contain the nitrate moiety sometimes come back positive for Ames when they are indeed safe. The nitrate compounds may generate nitric oxide, an important signal molecule that can give a false positive. Nitroglycerin is an example that gives a positive Ames yet is still used in treatment today. Nitrates in food however may be reduced by bacterial action to nitrites which are known to generate carcinogens by reacting with amines and amides. Long toxicology and outcome studies are needed with such compounds to disprove a positive Ames test. The Ames test was initially developed using agar plates (the plate incorporation technique), as described above. Since that time, an alternative to performing the Ames test has been developed, which is known as the "fluctuation method". This technique is the same in concept as the agar-based method, with bacteria being added to a reaction mixture with a small amount of histidine, which allows the bacteria to grow and mutate, returning to synthesize their own histidine. By including a pH indicator, the frequency of mutation is counted in microplates as the number of wells which have changed color (caused by a drop in pH due to metabolic processes of reproducing bacteria). As with the traditional Ames test, the sample is compared to the natural background rate of reverse mutation in order to establish the genotoxicity of a substance. The fluctuation method is performed entirely in liquid culture and is scored by counting the number of wells that turn yellow from purple in 96-well or 384-well microplates. In the 96-well plate method the frequency of mutation is counted as the number of wells out of 96 which have changed color. The plates are incubated for up to five days, with mutated (yellow) colonies being counted each day and compared to the background rate of reverse mutation using established tables of significance to determine the significant differences between the background rate of mutation and that for the tested samples. In the more scaled-down 384-well plate microfluctuation method the frequency of mutation is counted as the number of wells out of 48 which have changed color after 2 days of incubation. A test sample is assayed across 6 dose levels with concurrent zero-dose (background) and positive controls which all fit into one 384-well plate. The assay is performed in triplicates to provide statistical robustness. It uses the recommended OECD Guideline 471 tester strains (histidine auxotrophs and tryptophan auxotrophs). The fluctuation method is comparable to the traditional pour plate method in terms of sensitivity and accuracy, however, it does have a number of advantages: it needs less test sample, it has a simple colorimetric endpoint, counting the number of positive wells out of possible 96 or 48 wells is much less time-consuming than counting individual colonies on an agar plate. Several commercial kits are available. Most kits have consumable components in a ready-to-use state, including lyophilized bacteria, and tests can be performed using multichannel pipettes. The fluctuation method also allows for testing higher volumes of aqueous samples (up to 75% v/v), increasing the sensitivity and extending its application to low-level environmental mutagens.
https://en.wikipedia.org/wiki?curid=2766
ACE inhibitor Angiotensin-converting-enzyme inhibitors (ACE inhibitors) are a class of medication used primarily for the treatment of high blood pressure and heart failure. They work by causing relaxation of blood vessels as well as a decrease in blood volume, which leads to lower blood pressure and decreased oxygen demand from the heart. ACE inhibitors inhibit the activity of angiotensin-converting enzyme, an important component of the renin–angiotensin system liable to convert angiotensin I to angiotensin II, and hydrolyse bradykinin. Therefore, ACE inhibitors decrease the formation of angiotensin II, a vasoconstrictor, and increase the level of bradykinin, a peptide vasodilator. This combination is synergistic in lowering blood pressure. Frequently prescribed ACE inhibitors include benazepril, zofenopril, perindopril, trandolapril, captopril, enalapril, lisinopril, and ramipril. ACE inhibitors were initially approved for the treatment of hypertension and can be used alone or in combination with other anti-hypertensive medications. Later, they were found useful for other cardiovascular and kidney diseases including: In treating high blood pressure, ACE inhibitors are often the first drug choice, particularly when diabetes is present, but age can lead to different choices and it is common to need more than one drug to obtain the desired improvement. There are fixed-dose combination drugs, such as ACE inhibitor and thiazide combinations. ACE inhibitors have also been used in chronic kidney failure and kidney involvement in systemic sclerosis (hardening of tissues, as scleroderma renal crisis). In those with stable coronary artery disease, but no heart failure, benefits are similar to other usual treatments. In 2012, there was a meta-analysis published in the BMJ that described the protective role of ACE inhibitors in reducing the risk of pneumonia when compared to ARBs (Angiotensin II Receptor Blockers). The authors found a decreased risk in patients with previous stroke (54% risk reduction), with heart failure (37% risk reduction), and of Asian descent (43% risk reduction vs 54% risk reduction in non-Asian population). However, no reduced pneumonia related mortality was observed. ACE inhibitors may also be used to help decrease excessive water consumption in people with schizophrenia resulting in psychogenic polydipsia. A double-blind, placebo-controlled trial showed that when used for this purpose, enalapril led to decreased consumption (determined by urine output and osmolality) in 60% of people; the same effect has been demonstrated in other ACE inhibitors. Additionally ACEi are commonly used after renal transplant to manage post-transplant erythrocytosis, a condition characterised by a persistently high hematocrit greater than 51% which often develops 8-24 months after successful transplantation, as ACEi have been shown to decrease erythropoietin production. Common side effects include: low blood pressure, cough, hyperkalemia, headache, dizziness, fatigue, nausea, and kidney impairment. The main adverse effects of ACE inhibition can be understood from their pharmacological action. The other reported adverse effects are liver problems and effect on the fetus. Kidney problems may occur with all ACE inhibitors that directly follows from their mechanism of action. Patients starting on an ACE inhibitor usually have a modest reduction in glomerular filtration rate (GFR). However, the decrease may be significant in conditions of "pre-existing" decreased renal perfusion, such as renal artery stenosis, heart failure, polycystic kidney disease, or volume depletion. In these patients, the maintenance of GFR depends on angiotensin-II-dependent efferent vasomotor tone. Therefore, renal function should be closely monitored over the first few days after initiation of treatment with ACE inhibitor in patients with decreased renal perfusion. A moderate reduction in renal function, no greater than 30% rise in serum creatinine, that is stabilized after a week of treatment is deemed acceptable as part of the therapeutic effect, providing the residual renal function is sufficient. Reduced GFR is especially a problem if the patient is concomitantly taking an NSAID and a diuretic. When the three drugs are taken together, the risk of developing renal failure is significantly increased. High blood potassium is another possible complication of treatment with an ACE inhibitor due to its effect on aldosterone. Suppression of angiotensin II leads to a decrease in aldosterone levels. Since aldosterone is responsible for increasing the excretion of potassium, ACE inhibitors can cause retention of potassium. Some people, however, can continue to lose potassium while on an ACE inhibitor. Hyperkalemia may decrease the velocity of impulse conduction in the nerves and muscles, including cardiac tissues. This leads to cardiac dysfunction and neuromuscular consequences, such as muscle weakness, paresthesia, nausea, diarrhea, and others. Close monitoring of potassium levels is required in patients receiving treatment with ACE inhibitors who are at risk of hyperkalemia. Another possible adverse effect specific for ACE inhibitors, but not for other RAAS blockers, is an increase in bradykinin level. A persistent dry cough is a relatively common adverse effect believed to be associated with the increases in bradykinin levels produced by ACE inhibitors, although the role of bradykinin in producing these symptoms has been disputed. Many cases of cough in people on ACE inhibitors may not be from the medication itself, however. People who experience this cough are often switched to angiotensin II receptor antagonists. Some (0.7%) develop angioedema due to increased bradykinin levels. A genetic predisposition may exist. A severe rare allergic reaction can affect the bowel wall and secondarily cause abdominal pain. Hematologic effects, such as neutropenia, agranulocytosis and other blood dyscrasias, have occurred during therapy with ACE inhibitors, especially in people with additional risk factors. In pregnant women, ACE inhibitors taken during all the trimesters have been reported to cause congenital malformations, stillbirths, and neonatal deaths. Commonly reported fetal abnormalities include hypotension, renal dysplasia, anuria/oliguria, oligohydramnios, intrauterine growth retardation, pulmonary hypoplasia, patent ductus arteriosus, and incomplete ossification of the skull. Overall, about half of newborns exposed to ACE inhibitors are adversely affected, leading to birth defects. ACE inhibitors are ADEC pregnancy category D, and should be avoided in women who are likely to become pregnant. In the U.S., ACE inhibitors must be labeled with a boxed warning concerning the risk of birth defects when taken during the second and third trimester. Their use in the first trimester is also associated with a risk of major congenital malformations, particularly affecting the cardiovascular and central nervous systems. Symptoms and Treatment: There are few reports of ACE inhibitor overdose in the literature. The most likely manifestations are hypotension, which may be severe, hyperkalemia, hyponatremia and renal impairment with metabolic acidosis. Treatment should be mainly symptomatic and supportive, with volume expansion using normal saline to correct hypotension and improve renal function, and gastric lavage followed by activated charcoal and a cathartic to prevent further absorption of the drug. Captopril, enalapril, lisinopril and perindopril are known to be removable by hemodialysis. The ACE inhibitors are contraindicated in people with: ACE inhibitors should be used with caution in people with: A combination of ACE inhibitor with other drugs may increase effects of these drugs, but also the risk of adverse effects. The commonly reported adverse effects of drug combination with ACE are acute renal failure, hypotension, and hyperkalemia. The drugs interacting with ACE inhibitor should be prescribed with caution. Special attention should be given to combinations of ACE inhibitor with other RAAS blockers, diuretics (especially potassium-sparing diuretics), NSAIDs, anticoagulants, cyclosporine, DPP-4 inhibitors, and potassium supplements. Potassium supplementation should be used with caution and under medical supervision owing to the hyperkalemic effect of ACE inhibitors. Concomitant use with cyclooxygenase inhibitors tends to decrease ACE inhibitor's hypotensive effect. ACE inhibitors reduce the activity of the renin–angiotensin–aldosterone system (RAAS) as the primary etiologic (causal) event in the development of hypertension in people with diabetes mellitus, as part of the insulin-resistance syndrome or as a manifestation of renal disease. Renin–angiotensin–aldosterone system is a major blood pressure regulating mechanism. Markers of electrolyte and water imbalance in the body such as hypotension, low distal tubule sodium concentration, decreased blood volume and high sympathetic tone trigger the release of the enzyme renin from the cells of juxtaglomerular apparatus in the kidney. Renin activates a circulating liver derived prohormone angiotensinogen by proteolytic cleavage of all but its first ten amino acid residues known as angiotensin I. ACE (Angiotensin converting enzyme) then removes a further two residues, converting angiotensin I into angiotensin II. ACE is found in the pulmonary circulation and in the endothelium of many blood vessels. The system increases blood pressure by increasing the amount of salt and water the body retains, although angiotensin is also very good at causing the blood vessels to tighten (a potent vasoconstrictor). ACE inhibitors block the conversion of Angiotensin I (ATI) to Angiotensin II (ATII). They thereby lower arteriolar resistance and increase venous capacity; decrease cardiac output, cardiac index, stroke work, and volume; lower resistance in blood vessels in the kidneys; and lead to increased natriuresis (excretion of sodium in the urine). Renin increases in concentration in the blood as a result of negative feedback of conversion of ATI to ATII. ATI increases for the same reason; ATII and aldosterone decrease. Bradykinin increases because of less inactivation by ACE. Under normal conditions, angiotensin II has these effects: During the course of ACE inhibitor use, the production of ATII is decreased, which prevents aldosterone release from the adrenal cortex. This allows the kidney to excrete sodium ions along with obligate water, and retain potassium ions. This decreases blood volume, leading to decreased blood pressure. Epidemiological and clinical studies have shown ACE inhibitors reduce the progress of diabetic nephropathy independently from their blood pressure-lowering effect. This action of ACE inhibitors is used in the prevention of diabetic renal failure. ACE inhibitors have been shown to be effective for indications other than hypertension even in patients with normal blood pressure. The use of a maximum dose of ACE inhibitors in such patients (including for prevention of diabetic nephropathy, congestive heart failure, and prophylaxis of cardiovascular events) is justified, because it improves clinical outcomes independently of the blood pressure-lowering effect of ACE inhibitors. Such therapy, of course, requires careful and gradual titration of the dose to prevent the effects of rapidly decreasing blood pressure (dizziness, fainting, etc.). ACE inhibitors have also been shown to cause a central enhancement of parasympathetic nervous system activity in healthy volunteers and patients with heart failure. This action may reduce the prevalence of malignant cardiac arrhythmias, and the reduction in sudden death reported in large clinical trials. ACE Inhibitors also reduce plasma norepinephrine levels, and its resulting vasoconstriction effects, in heart failure patients, thus breaking the vicious circles of sympathetic and renin angiotensin system activation, which sustains the downward spiral in cardiac function in congestive heart failure The ACE inhibitor enalapril has also been shown to reduce cardiac cachexia in patients with chronic heart failure. Cachexia is a poor prognostic sign in patients with chronic heart failure. ACE inhibitors are under early investigation for the treatment of frailty and muscle wasting (sarcopenia) in elderly patients without heart failure. ACE inhibitors are easily identifiable by their common suffix, '-pril'. ACE inhibitors can be divided into three groups based on their molecular structure of the enzyme binding sites (sulfhydryl, phosphinyl, carboxyl) to the active center of ACE: These agents appear to show antioxidative properties but may be involved in adverse events such as skin eruptions. This is the largest group, including: All ACE inhibitors have similar antihypertensive efficacy when equivalent doses are administered. The main differences lie with captopril, the first ACE inhibitor. Captopril has a shorter duration of action and an increased incidence of adverse effects. It is also the only ACE inhibitor capable of passing through the blood–brain barrier, although the significance of this characteristic has not been shown to have any positive clinical effects. In a large clinical study, one of the agents in the ACE inhibitor class, ramipril (Altace), demonstrated an ability to reduce the mortality rates of patients suffering from a myocardial infarction, and to slow the subsequent development of heart failure. This finding was made after it was discovered that regular use of ramipril reduced mortality rates even in test subjects not having suffered from hypertension. Some believe ramipril's additional benefits may be shared by some or all drugs in the ACE-inhibitor class. However, ramipril currently remains the only ACE inhibitor for which such effects are actually evidence-based. A meta-analysis confirmed that ACE inhibitors are effective and certainly the first-line choice in hypertension treatment. This meta-analysis was based on 20 trials and a cohort of 158,998 patients, of whom 91% were hypertensive. ACE inhibitors were used as the active treatment in seven trials (n=76,615) and angiotensin receptor blocker (ARB) in 13 trials (n=82,383). ACE inhibitors were associated with a statistically significant 10% mortality reduction: (HR 0.90; 95% CI, 0.84–0.97; P=0.004). In contrast, no significant mortality reduction was observed with ARB treatment (HR 0.99; 95% CI, 0.94–1.04; P=0.683). Analysis of mortality reduction by different ACE inhibitors showed that perindopril-based regimens are associated with a statistically significant 13% all-cause mortality reduction. Taking into account the broad spectrum of the hypertensive population, one might expect that an effective treatment with ACE inhibitors, in particular with perindopril, would result in an important gain of lives saved. The ACE inhibitors have different strengths with different starting dosages. Dosage should be adjusted according to the clinical response. ACE inhibitors possess many common characteristics with another class of cardiovascular drugs, angiotensin II receptor antagonists, which are often used when patients are intolerant of the adverse effects produced by ACE inhibitors. ACE inhibitors do not completely prevent the formation of angiotensin II, as blockage is dose-dependent, so angiotensin II receptor antagonists may be useful because they act to prevent the action of angiotensin II at the AT1 receptor, leaving AT2 receptor unblocked; the latter may have consequences needing further study. The combination therapy of angiotensin II receptor antagonists with ACE inhibitors may be superior to either agent alone. This combination may increase levels of bradykinin while blocking the generation of angiotensin II and its activity at the AT1 receptor. This 'dual blockade' may be more effective than using an ACE inhibitor alone, because angiotensin II can be generated via non-ACE-dependent pathways. Preliminary studies suggest this combination of pharmacologic agents may be advantageous in the treatment of essential hypertension, chronic heart failure, and nephropathy. However, the more recent ONTARGET study showed no benefit of combining the agents and more adverse events. While statistically significant results have been obtained for its role in treating hypertension, clinical significance may be lacking. There are warnings about the combination of ACE inhibitors with ARBs. Patients with heart failure may benefit from the combination in terms of reducing morbidity and ventricular remodeling. The most compelling evidence for the treatment of nephropathy has been found: This combination therapy partially reversed the proteinuria and also exhibited a renoprotective effect in patients afflicted with diabetic nephropathy, and pediatric IgA nephropathy. Leonard T. Skeggs and his colleagues (including Norman Shumway) discovered ACE in plasma in 1956. It was also noted that those who worked in banana plantations in South-western Brazil collapsed after being bitten by a pit viper, leading to a search for a blood pressure lowering component in its venom. Brazilian scientist Sérgio Henrique Ferreira reported a bradykinin-potentiating factor (BPF) present in the venom of "Bothrops jararaca", a South American pit viper, in 1965. Ferreira then went to John Vane's laboratory as a postdoctoral fellow with his already-isolated BPF. The conversion of the inactive angiotensin I to the potent angiotensin II was thought to take place in the plasma. However, in 1967, Kevin K. F. Ng and John R. Vane showed plasma ACE is too slow to account for the conversion of angiotensin I to angiotensin II "in vivo". Subsequent investigation showed rapid conversion occurs during its passage through the pulmonary circulation. Bradykinin is rapidly inactivated in the circulating blood, and it disappears completely in a single pass through the pulmonary circulation. Angiotensin I also disappears in the pulmonary circulation because of its conversion to angiotensin II. Furthermore, angiotensin II passes through the lungs without any loss. The inactivation of bradykinin and the conversion of angiotensin I to angiotensin II in the lungs was thought to be caused by the same enzyme. In 1970, Ng and Vane, using BPF provided by Ferreira, showed the conversion is inhibited during its passage through the pulmonary circulation. BPFs are members of a family of peptides whose potentiating action is linked to inhibition of bradykinin by ACE. Molecular analysis of BPF yielded a nonapeptide BPF teprotide (SQ 20,881), which showed the greatest ACE inhibition potency and hypotensive effect "in vivo". Teprotide had limited clinical value as a result of its peptide nature and lack of activity when given orally. In the early 1970s, knowledge of the structure-activity relationship required for inhibition of ACE was growing. David Cushman, Miguel Ondetti and colleagues used peptide analogues to study the structure of ACE, using carboxypeptidase A as a model. Their discoveries led to the development of captopril, the first orally-active ACE inhibitor, in 1975. Captopril was approved by the United States Food and Drug Administration in 1981. The first nonsulfhydryl-containing ACE inhibitor, enalapril, was marketed two years later. At least 12 other ACE inhibitors have since been marketed. In 1991, Japanese scientists created the first milk-based ACE inhibitor, in the form of a fermented milk drink, using specific cultures to liberate the tripeptide isoleucine-proline-proline (IPP) from the dairy protein. Valine-proline-proline (VPP) is also liberated in this process—another milk tripeptide with a very similar chemical structure to IPP. Together, these peptides are now often referred to as lactotripeptides. In 1996, the first human study confirmed the blood pressure-lowering effect of IPP in fermented milk. Although twice the amount of VPP is needed to achieve the same ACE-inhibiting activity as the originally discovered IPP, VPP also is assumed to add to the total blood pressure lowering effect. Since the first lactotripeptides discovery, more than 20 human clinical trials have been conducted in many different countries.
https://en.wikipedia.org/wiki?curid=2767
Anatomical Therapeutic Chemical Classification System The Anatomical Therapeutic Chemical (ATC) Classification System is a drug classification system that classifies the active ingredients of drugs according to the organ or system on which they act and their therapeutic, pharmacological and chemical properties. It is controlled by the World Health Organization Collaborating Centre for Drug Statistics Methodology (WHOCC), and was first published in 1976. This pharmaceutical coding system divides drugs into different groups according to the organ or system on which they act, their therapeutic intent or nature, and the drug's chemical characteristics. Different brands share the same code if they have the same active substance and indications. Each bottom-level ATC code stands for a pharmaceutically used substance, or a combination of substances, in a single indication (or use). This means that one drug can have more than one code, for example acetylsalicylic acid (aspirin) has as a drug for local oral treatment, as a platelet inhibitor, and as an analgesic and antipyretic; as well as one code can represent more than one active ingredient, for example is the combination of perindopril with amlodipine, two active ingredients that have their own codes ( and respectively) when prescribed alone. The ATC classification system is a strict hierarchy, meaning that each code necessarily has one and only one parent code, except for the 14 codes at the topmost level which have no parents. The codes are semantic identifiers, meaning they depict information by themselves beyond serving as identifiers (namely, the codes depict themselves the complete lineage of parenthood). As of 7 May 2020, there are 6,331 codes in ATC; the table below gives the count per level. The ATC system is based on the earlier Anatomical Classification System, which is intended as a tool for the pharmaceutical industry to classify pharmaceutical products (as opposed to their active ingredients). This system, confusingly also called ATC, was initiated in 1971 by the European Pharmaceutical Market Research Association (EphMRA) and is being maintained by the EphMRA and Intellus. Its codes are organised into four levels. The WHO's system, having five levels, is an extension and modification of the EphMRA's. It was first published in 1976. In this system, drugs are classified into groups at five different levels: The first level of the code indicates the anatomical main group and consists of one letter. There are 14 main groups: "Example": C Cardiovascular system The second level of the code indicates the therapeutic subgroup and consists of two digits. "Example": C03 Diuretics The third level of the code indicates the therapeutic/pharmacological subgroup and consists of one letter. "Example": C03C High-ceiling diuretics The fourth level of the code indicates the chemical/therapeutic/pharmacological subgroup and consists of one letter. "Example": C03CA Sulfonamides The fifth level of the code indicates the chemical substance and consists of two digits. "Example": C03CA01 furosemide The "Anatomical Therapeutic Chemical Classification System for veterinary medicinal products" (ATCvet) is used to classify veterinary drugs. ATCvet codes can be created by placing the letter Q in front of the ATC code of most human medications. For example, furosemide for veterinary use has the code QC03CA01. Some codes are used exclusively for veterinary drugs, such as "QI Immunologicals", "QJ51 Antibacterials for intramammary use" or "QN05AX90 amperozide". The ATC system also includes defined daily doses (DDDs) for many drugs. This is a measurement of drug consumption based on the usual daily dose for a given drug. According to the definition, "[t]he DDD is the assumed average maintenance dose per day for a drug used for its main indication in adults." National issues of the ATC classification, such as the German "Anatomisch-therapeutisch-chemische Klassifikation mit Tagesdosen", may include additional codes and DDDs not present in the WHO version. ATC follows guidelines in creating new codes for newly approved drugs. In order to create a new ATC code, an application has to be sent to the WHO. New ATC codes are published twice annually. Revisions of the ATC are formally released every year.
https://en.wikipedia.org/wiki?curid=2770
Parallel ATA Parallel ATA (PATA), originally ', is an interface standard for the connection of storage devices such as hard disk drives, floppy disk drives, and optical disc drives in computers. The standard is maintained by the X3/INCITS committee. It uses the underlying ' (ATA) and Packet Interface (ATAPI) standards. The Parallel ATA standard is the result of a long history of incremental technical development, which began with the original AT Attachment interface, developed for use in early PC AT equipment. The ATA interface itself evolved in several stages from Western Digital's original Integrated Drive Electronics (IDE) interface. As a result, many near-synonyms for ATA/ATAPI and its previous incarnations are still in common informal use, in particular Extended IDE (EIDE) and Ultra ATA (UATA). After the introduction of Serial ATA (SATA) in 2003, the original ATA was renamed to Parallel ATA, or PATA for short. Parallel ATA cables have a maximum allowable length of . Because of this limit, the technology normally appears as an internal computer storage interface. For many years, ATA provided the most common and the least expensive interface for this application. It has largely been replaced by SATA in newer systems. The standard was originally conceived as the "AT Bus Attachment," officially called "AT Attachment" and abbreviated "ATA" because its primary feature was a direct connection to the 16-bit ISA bus introduced with the IBM PC/AT. The original ATA specifications published by the standards committees use the name "AT Attachment". The "AT" in the IBM PC/AT referred to "Advanced Technology" so ATA has also been referred to as "Advanced Technology Attachment". When a newer Serial ATA (SATA) was introduced in 2003, the original ATA was renamed to Parallel ATA, or PATA for short. Physical ATA interfaces became a standard component in all PCs, initially on host bus adapters, sometimes on a sound card but ultimately as two physical interfaces embedded in a Southbridge chip on a motherboard. Called the "primary" and "secondary" ATA interfaces, they were assigned to base addresses 0x1F0 and 0x170 on ISA bus systems. They were replaced by SATA interfaces. The first version of what is now called the ATA/ATAPI interface was developed by Western Digital under the name "Integrated Drive Electronics" (IDE). Together with Control Data Corporation (the hard drive manufacturer) and Compaq Computer (the initial customer), they developed the connector, the signaling protocols and so on, with the goal of remaining software compatible with the existing ST-506 hard drive interface. The first such drives appeared internally in Compaq PCs in 1986. and were first separately offered by Conner Peripherals as the CP342 in June 1987. The term "Integrated Drive Electronics" refers not just to the connector and interface definition, but also to the fact that the drive controller is integrated into the drive, as opposed to a separate controller on or connected to the motherboard. The interface cards used to connect a parallel ATA drive to, for example, a PCI slot are not drive controllers: they are merely bridges between the host bus and the ATA interface. Since the original ATA interface is essentially just a 16-bit ISA bus in disguise, the bridge was especially simple in case of an ATA connector being located on an ISA interface card. The integrated controller presented the drive to the host computer as an array of 512-byte blocks with a relatively simple command interface. This relieved the mainboard and interface cards in the host computer of the chores of stepping the disk head arm, moving the head arm in and out, and so on, as had to be done with earlier ST-506 and ESDI hard drives. All of these low-level details of the mechanical operation of the drive were now handled by the controller on the drive itself. This also eliminated the need to design a single controller that could handle many different types of drives, since the controller could be unique for the drive. The host need only to ask for a particular sector, or block, to be read or written, and either accept the data from the drive or send the data to it. The interface used by these drives was standardized in 1994 as ANSI standard X3.221-1994, "AT Attachment Interface for Disk Drives". After later versions of the standard were developed, this became known as "ATA-1". A short-lived, seldom-used implementation of ATA was created for the IBM XT and similar machines that used the 8-bit version of the ISA bus. It has been referred to as "XT-IDE", "XTA" or "XT Attachment". In 1994, about the same time that the ATA-1 standard was adopted, Western Digital introduced drives under a newer name, Enhanced IDE (EIDE). These included most of the features of the forthcoming ATA-2 specification and several additional enhancements. Other manufacturers introduced their own variations of ATA-1 such as "Fast ATA" and "Fast ATA-2". The new version of the ANSI standard, "AT Attachment Interface with Extensions ATA-2" (X3.279-1996), was approved in 1996. It included most of the features of the manufacturer-specific variants. ATA-2 also was the first to note that devices other than hard drives could be attached to the interface: As mentioned in the previous sections, ATA was originally designed for, and worked only with hard disk drives and devices that could emulate them. The introduction of ATAPI (ATA Packet Interface) by a group called the Small Form Factor committee (SFF) allowed ATA to be used for a variety of other devices that require functions beyond those necessary for hard disk drives. For example, any removable media device needs a "media eject" command, and a way for the host to determine whether the media is present, and these were not provided in the ATA protocol. The Small Form Factor committee approached this problem by defining ATAPI, the "ATA Packet Interface". ATAPI is actually a protocol allowing the ATA interface to carry SCSI commands and responses; therefore, all ATAPI devices are actually "speaking SCSI" other than at the electrical interface. In fact, some early ATAPI devices were simply SCSI devices with an ATA/ATAPI to SCSI protocol converter added on. The SCSI commands and responses are embedded in "packets" (hence "ATA Packet Interface") for transmission on the ATA cable. This allows any device class for which a SCSI command set has been defined to be interfaced via ATA/ATAPI. ATAPI devices are also "speaking ATA", as the ATA physical interface and protocol are still being used to send the packets. On the other hand, ATA hard drives and solid state drives do not use ATAPI. ATAPI devices include CD-ROM and DVD-ROM drives, tape drives, and large-capacity floppy drives such as the Zip drive and SuperDisk drive. The SCSI commands and responses used by each class of ATAPI device (CD-ROM, tape, etc.) are described in other documents or specifications specific to those device classes and are not within ATA/ATAPI or the T13 committee's purview. One commonly used set is defined in the MMC SCSI command set. ATAPI was adopted as part of ATA in INCITS 317-1998, "AT Attachment with Packet Interface Extension (ATA/ATAPI-4)". The ATA/ATAPI-4 standard also introduced several "Ultra DMA" transfer modes. These initially supported speeds from 16 MByte/s to 33 MByte/second. In later versions, faster Ultra DMA modes were added, requiring new 80-wire cables to reduce crosstalk. The latest versions of Parallel ATA support up to 133 MByte/s. Ultra ATA, abbreviated UATA, is a designation that has been primarily used by Western Digital for different speed enhancements to the ATA/ATAPI standards. For example, in 2000 Western Digital published a document describing "Ultra ATA/100", which brought performance improvements for the then-current ATA/ATAPI-5 standard by improving maximum speed of the Parallel ATA interface from 66 to 100 MB/s. Most of Western Digital's changes, along with others, were included in the ATA/ATAPI-6 standard (2002). The terms "integrated drive electronics" (IDE), "enhanced IDE" and "EIDE" have come to be used interchangeably with ATA (now Parallel ATA, or PATA). In addition, there have been several generations of "EIDE" drives marketed, compliant with various versions of the ATA specification. An early "EIDE" drive might be compatible with ATA-2, while a later one with ATA-6. Nevertheless, a request for an "IDE" or "EIDE" drive from a computer parts vendor will almost always yield a drive that will work with most Parallel ATA interfaces. Another common usage is to refer to the specification version by the fastest mode supported. For example, ATA-4 supported Ultra DMA modes 0 through 2, the latter providing a maximum transfer rate of 33 megabytes per second. ATA-4 drives are thus sometimes called "UDMA-33" drives, and sometimes "ATA-33" drives. Similarly, ATA-6 introduced a maximum transfer speed of 100 megabytes per second, and some drives complying to this version of the standard are marketed as "PATA/100" drives. Initially, the size of an ATA drive was stored in the system x86 BIOS using a type number (1 through 45) that predefined the C/H/S parameters and also often the landing zone, in which the drive heads are parked while not in use. Later, a "user definable" format called C/H/S or cylinders, heads, sectors was made available. These numbers were important for the earlier ST-506 interface, but were generally meaningless for ATA—the CHS parameters for later ATA large drives often specified impossibly high numbers of heads or sectors that did not actually define the internal physical layout of the drive at all. From the start, and up to ATA-2, every user had to specify explicitly how large every attached drive was. From ATA-2 on, an "identify drive" command was implemented that can be sent and which will return all drive parameters. Owing to a lack of foresight by motherboard manufacturers, the system BIOS was often hobbled by artificial C/H/S size limitations due to the manufacturer assuming certain values would never exceed a particular numerical maximum. The first of these BIOS limits occurred when ATA drives reached sizes in excess of 504 MiB, because some motherboard BIOSes would not allow C/H/S values above 1024 cylinders, 16 heads, and 63 sectors. Multiplied by 512 bytes per sector, this totals bytes which, divided by bytes per MiB, equals 504 MiB (528 MB). The second of these BIOS limitations occurred at 1024 cylinders, 256 heads, and 63 sectors, and a bug in MS-DOS and MS-Windows 95 limited the number of heads to 255. This totals to bytes (8032.5 MiB), commonly referred to as the 8.4 gigabyte barrier. This is again a limit imposed by x86 BIOSes, and not a limit imposed by the ATA interface. It was eventually determined that these size limitations could be overridden with a tiny program loaded at startup from a hard drive's boot sector. Some hard drive manufacturers, such as Western Digital, started including these override utilities with new large hard drives to help overcome these problems. However, if the computer was booted in some other manner without loading the special utility, the invalid BIOS settings would be used and the drive could either be inaccessible or appear to the operating system to be damaged. Later, an extension to the x86 BIOS disk services called the "Enhanced Disk Drive" (EDD) was made available, which makes it possible to address drives as large as 264 sectors. The first drive interface used 22-bit addressing mode which resulted in a maximum drive capacity of two gigabytes. Later, the first formalized ATA specification used a 28-bit addressing mode through LBA28, allowing for the addressing of 228 () sectors (blocks) of 512 bytes each, resulting in a maximum capacity of 128 GiB (137 GB). ATA-6 introduced 48-bit addressing, increasing the limit to 128 PiB (144 PB). As a consequence, any ATA drive of capacity larger than about 137 GB must be an ATA-6 or later drive. Connecting such a drive to a host with an ATA-5 or earlier interface will limit the usable capacity to the maximum of the interface. Some operating systems, including Windows XP pre-SP1, and Windows 2000 pre-SP3, disable LBA48 by default, requiring the user to take extra steps to use the entire capacity of an ATA drive larger than about 137 gigabytes. Older operating systems, such as Windows 98, do not support 48-bit LBA at all. However, members of the third-party group MSFN have modified the Windows 98 disk drivers to add unofficial support for 48-bit LBA to Windows 95 OSR2, Windows 98, Windows 98 SE and Windows ME. Some 16-bit and 32-bit operating systems supporting LBA48 may still not support disks larger than 2 TiB due to using 32-bit arithmetics only; a limitation also applying to many boot sectors. Parallel ATA (then simply called ATA or IDE) became the primary storage device interface for PCs soon after its introduction. In some systems, a third and fourth motherboard interface was provided, allowing up to eight ATA devices to be attached to the motherboard. Often, these additional connectors were implemented by inexpensive RAID controllers. Soon after the introduction of Serial ATA (SATA) in 2003, use of Parallel ATA declined. The first motherboards with built-in SATA interfaces usually had only a single PATA connector (for up to two PATA devices), along with multiple SATA connectors. As of 2007, some PC chipsets, for example the Intel ICH10, had removed support for PATA. Motherboard vendors still wishing to offer Parallel ATA with those chipsets must include an additional interface chip. In more recent computers, the Parallel ATA interface is rarely used even if present, as four or more Serial ATA connectors are usually provided on the motherboard and SATA devices of all types are common. With Western Digital's withdrawal from the PATA market, hard disk drives with the PATA interface were no longer in production after December 2013 for other than specialty applications. Parallel ATA cables transfer data 16 bits at a time. The traditional cable uses 40-socket connectors attached to a 40- or 80-conductor ribbon cable. Each cable has two or three connectors, one of which plugs into a host adapter interfacing with the rest of the computer system. The remaining connector(s) plug into storage devices, most commonly hard disk drives or optical drives. Each connector has 39 physical pins arranged into two rows, with a gap or key at pin 20. Round parallel ATA cables (as opposed to ribbon cables) were eventually made available for 'case modders' for cosmetic reasons, as well as claims of improved computer cooling and were easier to handle; however, only ribbon cables are supported by the ATA specifications. In the ATA standard, pin 20 is defined as a mechanical key and is not used. This pin's socket on the female connector is often obstructed, requiring pin 20 to be omitted from the male cable or drive connector; it is thus impossible to plug it in the wrong way round. However, some flash memory drives can use pin 20 as VCC_in to power the drive without requiring a special power cable; this feature can only be used if the equipment supports this use of pin 20. Pin 28 of the gray (slave/middle) connector of an 80-conductor cable is not attached to any conductor of the cable. It is attached normally on the black (master drive end) and blue (motherboard end) connectors. Pin 34 is connected to ground inside the blue connector of an 80-conductor cable but not attached to any conductor of the cable. It is attached normally on the gray and black connectors. A 44-pin variant PATA connector is used for 2.5 inch drives inside laptops. The pins are closer together and the connector is physically smaller than the 40-pin connector. The extra pins carry power. ATA's cables have had 40 wires for most of its history (44 conductors for the smaller form-factor version used for 2.5" drives—the extra four for power), but an 80-wire version appeared with the introduction of the "UDMA/66" mode. All of the additional wires in the new cable are ground wires, interleaved with the previously defined wires to reduce the effects of capacitive coupling between neighboring signal wires, reducing crosstalk. Capacitive coupling is more of a problem at higher transfer rates, and this change was necessary to enable the 66 megabytes per second (MB/s) transfer rate of "UDMA4" to work reliably. The faster "UDMA5" and "UDMA6" modes also require 80-conductor cables. Though the number of wires doubled, the number of connector pins and the pinout remain the same as 40-conductor cables, and the external appearance of the connectors is identical. Internally, the connectors are different; the connectors for the 80-wire cable connect a larger number of ground wires to the ground pins, while the connectors for the 40-wire cable connect ground wires to ground pins one-for-one. 80-wire cables usually come with three differently colored connectors (blue, black, and gray for controller, master drive, and slave drive respectively) as opposed to uniformly colored 40-wire cable's connectors (commonly all gray). The gray connector on 80-conductor cables has pin 28 CSEL not connected, making it the slave position for drives configured cable select. The image on the right shows PATA connectors after removal of strain relief, cover, and cable. Pin one is at bottom left of the connectors, pin 2 is top left, etc., except that the lower image of the blue connector shows the view from the opposite side, and pin one is at top right. The connector is an insulation-displacement connector—in other words, each contact comprises a pair of points which together pierce the insulation of the ribbon cable with such precision that they make a connection to the desired conductor without harming the insulation on the neighboring wires. The center row of contacts are all connected to the common ground bus and attached to the odd numbered conductors of the cable. The top row of contacts are the even-numbered sockets of the connector (mating with the even-numbered pins of the receptacle) and attach to every other even-numbered conductor of the cable. The bottom row of contacts are the odd-numbered sockets of the connector (mating with the odd-numbered pins of the receptacle) and attach to the remaining even-numbered conductors of the cable. Note the connections to the common ground bus from sockets 2 (top left), 19 (center bottom row), 22, 24, 26, 30, and 40 on all connectors. Also note (enlarged detail, bottom, looking from the opposite side of the connector) that socket 34 of the blue connector does not contact any conductor but unlike socket 34 of the other two connectors, it does connect to the common ground bus. On the gray connector, note that socket 28 is completely missing, so that pin 28 of the drive attached to the gray connector will be open. On the black connector, sockets 28 and 34 are completely normal, so that pins 28 and 34 of the drive attached to the black connector will be connected to the cable. Pin 28 of the black drive reaches pin 28 of the host receptacle but not pin 28 of the gray drive, while pin 34 of the black drive reaches pin 34 of the gray drive but not pin 34 of the host. Instead, pin 34 of the host is grounded. The standard dictates color-coded connectors for easy identification by both installer and cable maker. All three connectors are different from one another. The blue (host) connector has the socket for pin 34 connected to ground inside the connector but not attached to any conductor of the cable. Since the old 40 conductor cables do not ground pin 34, the presence of a ground connection indicates that an 80 conductor cable is installed. The wire for pin 34 is attached normally on the other types and is not grounded. Installing the cable backwards (with the black connector on the system board, the blue connector on the remote device and the gray connector on the center device) will ground pin 34 of the remote device and connect host pin 34 through to pin 34 of the center device. The gray center connector omits the connection to pin 28 but connects pin 34 normally, while the black end connector connects both pins 28 and 34 normally. If two devices are attached to a single cable, one must be designated as "Device 0" (in the past, commonly designated "master") and the other as "Device 1" (in the past, incorrectly designated as "slave"). This distinction is necessary to allow both drives to share the cable without conflict. The "Device 0" drive is the drive that usually appears "first" to the computer's BIOS and/or operating system. On Intel 486 PCs era and older, the drives are often referred to by the BIOS as "C" for the "master" and "D" for the "slave" following the way DOS would refer to the active primary partitions on each. The mode that a drive must use is often set by a jumper setting on the drive itself, which must be manually set to "Device 0" ("Primary"/"Master") or "Device 1" ("Secondary"/"Slave"). If there is a single device on a cable, it should be configured as "Device 0". However, some certain era drives have a special setting called "Single" for this configuration (Western Digital, in particular). Also, depending on the hardware and software available, a "Single" drive on a cable will often work reliably even though configured as the "Device 1" drive (most often seen where an optical drive is the only device on the secondary ATA interface). Although they were in common use, the terms "master" and "slave" do not appear anymore in current versions of the ATA specifications, or any current documentation. Since ATA-2 the two devices are referred to as "Device 0" and "Device 1", respectively. This is more appropriate since the two devices have always operated, since the earliest ATA specification, as equal peers on the cable, with neither having control or priority over the other. It is a common myth that the controller on the master drive assumes control over the slave drive, or that the master drive may claim priority of communication over the other device on the same ATA interface. In fact, the drivers in the host operating system perform the necessary arbitration and serialization, and each drive's onboard controller operates independently of the other. While it may have remained in colloquial use, the PC industry has not used ATA master/slave terminology in many years. A drive mode called "cable select" was described as optional in ATA-1 and has come into fairly widespread use with ATA-5 and later. A drive set to "cable select" automatically configures itself as "Device 0" or "Device 1", according to its position on the cable. Cable select is controlled by pin 28. The host adapter grounds this pin; if a device sees that the pin is grounded, it becomes the "Device 0" device; if it sees that pin 28 is open, the device becomes the "Device 1" device. This setting is usually chosen by a jumper setting on the drive called "cable select", usually marked "CS", which is separate from the "Device 0/1" setting. Note that if two drives are configured as "Device 0" and "Device 1" manually, this configuration does not need to correspond to their position on the cable. Pin 28 is only used to let the drives know their position on the cable; it is not used by the host when communicating with the drives. With the 40-wire cable, it was very common to implement cable select by simply cutting the pin 28 wire between the two device connectors; putting the "Device 1" device at the end of the cable, and the "Device 0" on the middle connector. This arrangement eventually was standardized in later versions. If there is just one device on the cable, this results in an unused stub of cable, which is undesirable for physical convenience and electrical reasons. The stub causes signal reflections, particularly at higher transfer rates. Starting with the 80-wire cable defined for use in ATAPI5/UDMA4, the "Device 0" device goes at the end of the cable—the black connector—and the slave "Device 1" goes on the middle connector—the gray one—and the blue connector goes onto the motherboard. So, if there is only one ("Device 0") device on the cable, there is no cable stub to cause reflections. Also, cable select is now implemented in the "device 1" device connector, usually simply by omitting the contact from the connector body The parallel ATA protocols up through ATA-3 require that once a command has been given on an ATA interface, it must complete before any subsequent command may be given. Operations on the devices must be serializedwith only one operation in progress at a timewith respect to the ATA host interface. A useful mental model is that the host ATA interface is busy with the first request for its entire duration, and therefore can not be told about another request until the first one is complete. The function of serializing requests to the interface is usually performed by a device driver in the host operating system. The ATA-4 and subsequent versions of the specification have included an "overlapped feature set" and a "queued feature set" as optional features, both being given the name "Tagged Command Queuing" (TCQ), a reference to a set of features from SCSI which the ATA version attempts to emulate. However, support for these is extremely rare in actual parallel ATA products and device drivers because these feature sets were implemented in such a way as to maintain software compatibility with its heritage as originally an extension of the ISA bus. This implementation resulted in excessive CPU utilization which largely negated the advantages of command queuing. By contrast, overlapped and queued operations have been common in other storage buses; in particular, SCSI's version of tagged command queuing had no need to be compatible with APIs designed for ISA, allowing it to attain high performance with low overhead on buses which supported first party DMA like PCI. This has long been seen as a major advantage of SCSI. The Serial ATA standard has supported native command queueing (NCQ) since its first release, but it is an optional feature for both host adapters and target devices. Many obsolete PC motherboards do not support NCQ, but modern SATA hard disk drives and SATA solid-state drives usually support NCQ, which is not the case for removable (CD/DVD) drives because the ATAPI command set used to control them prohibits queued operations. There are many debates about how much a slow device can impact the performance of a faster device on the same cable. There is an effect, but the debate is confused by the blurring of two quite different causes, called here "Lowest speed" and "One operation at a time". On early ATA host adapters, both devices' data transfers can be constrained to the speed of the slower device, if two devices of different speed capabilities are on the same cable. For all modern ATA host adapters, this is not true, as modern ATA host adapters support "independent device timing". This allows each device on the cable to transfer data at its own best speed. Even with earlier adapters, without independent timing, this effect applies only to the data transfer phase of a read or write operation. This is caused by the omission of both overlapped and queued feature sets from most parallel ATA products. Only one device on a cable can perform a read or write operation at one time; therefore, a fast device on the same cable as a slow device under heavy use will find it has to wait for the slow device to complete its task first. However, most modern devices will report write operations as complete once the data is stored in their onboard cache memory, before the data is written to the (slow) magnetic storage. This allows commands to be sent to the other device on the cable, reducing the impact of the "one operation at a time" limit. The impact of this on a system's performance depends on the application. For example, when copying data from an optical drive to a hard drive (such as during software installation), this effect probably will not matter. Such jobs are necessarily limited by the speed of the optical drive no matter where it is. But if the hard drive in question is also expected to provide good throughput for other tasks at the same time, it probably should not be on the same cable as the optical drive. ATA devices may support an optional security feature which is defined in an ATA specification, and thus not specific to any brand or device. The security feature can be enabled and disabled by sending special ATA commands to the drive. If a device is locked, it will refuse all access until it is unlocked. A device can have two passwords: A User Password and a Master Password; either or both may be set. There is a Master Password identifier feature which, if supported and used, can identify without disclosing the current Master Password. A device can be locked in two modes: High security mode or Maximum security mode. Bit 8 in word 128 of the IDENTIFY response shows which mode the disk is in: 0 = High, 1 = Maximum. In High security mode, the device can be unlocked with either the User or Master password, using the "SECURITY UNLOCK DEVICE" ATA command. There is an attempt limit, normally set to 5, after which the disk must be power cycled or hard-reset before unlocking can be attempted again. Also in High security mode, the SECURITY ERASE UNIT command can be used with either the User or Master password. In Maximum security mode, the device can be unlocked only with the User password. If the User password is not available, the only remaining way to get at least the bare hardware back to a usable state is to issue the SECURITY ERASE PREPARE command, immediately followed by SECURITY ERASE UNIT. In Maximum security mode, the SECURITY ERASE UNIT command requires the Master password and will completely erase all data on the disk. Word 89 in the IDENTIFY response indicates how long the operation will take. While the ATA lock is intended to be impossible to defeat without a valid password, there are purported workarounds to unlock a device. Due to a short cable length specification and shielding issues it is extremely uncommon to find external PATA devices that directly use PATA for connection to a computer. A device connected externally needs additional cable length to form a U-shaped bend so that the external device may be placed alongside, or on top of the computer case, and the standard cable length is too short to permit this. For ease of reach from motherboard to device, the connectors tend to be positioned towards the front edge of motherboards, for connection to devices protruding from the front of the computer case. This front-edge position makes extension out the back to an external device even more difficult. Ribbon cables are poorly shielded, and the standard relies upon the cabling to be installed inside a shielded computer case to meet RF emissions limits. External hard disk drives or optical disk drives that have an internal PATA interface, use some other interface technology to bridge the distance between the external device and the computer. USB is the most common external interface, followed by Firewire. A bridge chip inside the external devices converts from the USB interface to PATA, and typically only supports a single external device without cable select or master/slave. Compact Flash in its "IDE mode" is essentially a miniaturized ATA interface, intended for use on devices that use flash memory storage. No interfacing chips or circuitry are required, other than to directly adapt the smaller CF socket onto the larger ATA connector. (Although most CF cards only support IDE mode up to PIO4, making them much slower in IDE mode than their CF capable speed) The ATA connector specification does not include pins for supplying power to a CF device, so power is inserted into the connector from a separate source. The exception to this is when the CF device is connected to a 44-pin ATA bus designed for 2.5-inch hard disk drives, commonly found in notebook computers, as this bus implementation must provide power to a standard hard disk drive. CF devices can be designated as master or slave on an ATA interface, though since most CF devices offer only a single socket, it is not necessary to offer this selection to end users. Although CF can be hot-pluggable with additional design methods, by default when wired directly to an ATA interface, it is not intended to be hot-pluggable. The following table shows the names of the versions of the ATA standards and the transfer modes and rates supported by each. Note that the transfer rate for each mode (for example, 66.7 MB/s for UDMA4, commonly called "Ultra-DMA 66", defined by ATA-5) gives its maximum theoretical transfer rate on the cable. This is simply two bytes multiplied by the effective clock rate, and presumes that every clock cycle is used to transfer end-user data. In practice, of course, protocol overhead reduces this value. Congestion on the host bus to which the ATA adapter is attached may also limit the maximum burst transfer rate. For example, the maximum data transfer rate for conventional PCI bus is 133 MB/s, and this is shared among all active devices on the bus. In addition, no ATA hard drives existed in 2005 that were capable of measured sustained transfer rates of above 80 MB/s. Furthermore, sustained transfer rate tests do not give realistic throughput expectations for most workloads: They use I/O loads specifically designed to encounter almost no delays from seek time or rotational latency. Hard drive performance under most workloads is limited first and second by those two factors; the transfer rate on the bus is a distant third in importance. Therefore, transfer speed limits above 66 MB/s really affect performance only when the hard drive can satisfy all I/O requests by reading from its internal cache—a very unusual situation, especially considering that such data is usually already buffered by the operating system. , mechanical hard disk drives can transfer data at up to 157 MB/s, which is beyond the capabilities of the PATA/133 specification. High-performance solid state drives can transfer data at up to 308 MB/s. Only the Ultra DMA modes use CRC to detect errors in data transfer between the controller and drive. This is a 16-bit CRC, and it is used for data blocks only. Transmission of command and status blocks do not use the fast signaling methods that would necessitate CRC. For comparison, in Serial ATA, 32-bit CRC is used for both commands and data. ATAPI devices with removable media, other than CD and DVD drives, are classified as ARMD (ATAPI Removable Media Device) and can appear as either a super-floppy (non-partitioned media) or a hard drive (partitioned media) to the operating system. These can be supported as bootable devices by a BIOS complying with the ATAPI Removable Media Device BIOS Specification, originally developed by Compaq Computer Corporation and Phoenix Technologies. It specifies provisions in the BIOS of a personal computer to allow the computer to be bootstrapped from devices such as Zip drives, Jaz drives, SuperDisk (LS-120) drives, and similar devices. These devices have removable media like floppy disk drives, but capacities more commensurate with hard drives, and programming requirements unlike either. Due to limitations in the floppy controller interface most of these devices were ATAPI devices, connected to one of the host computer's ATA interfaces, similarly to a hard drive or CD-ROM device. However, existing BIOS standards did not support these devices. An ARMD-compliant BIOS allows these devices to be booted from and used under the operating system without requiring device-specific code in the OS. A BIOS implementing ARMD allows the user to include ARMD devices in the boot search order. Usually an ARMD device is configured earlier in the boot order than the hard drive. Similarly to a floppy drive, if bootable media is present in the ARMD drive, the BIOS will boot from it; if not, the BIOS will continue in the search order, usually with the hard drive last. There are two variants of ARMD, ARMD-FDD and ARMD-HDD. Originally ARMD caused the devices to appear as a sort of very large floppy drive, either the primary floppy drive device 00h or the secondary device 01h. Some operating systems required code changes to support floppy disks with capacities far larger than any standard floppy disk drive. Also, standard-floppy disk drive emulation proved to be unsuitable for certain high-capacity floppy disk drives such as Iomega Zip drives. Later the ARMD-HDD, ARMD-"Hard disk device", variant was developed to address these issues. Under ARMD-HDD, an ARMD device appears to the BIOS and the operating system as a hard drive. In August 2004, Sam Hopkins and Brantley Coile of Coraid specified a lightweight ATA over Ethernet protocol to carry ATA commands over Ethernet instead of directly connecting them to a PATA host adapter. This permitted the established block protocol to be reused in storage area network (SAN) applications.
https://en.wikipedia.org/wiki?curid=2778
Atari 2600 The Atari 2600, originally branded as the Atari Video Computer System or Atari VCS for short until November 1982, is a home video game console from Atari, Inc. Released on September 11, 1977, it is credited with popularizing the use of microprocessor-based hardware and games stored on ROM cartridges (a format first used with the Fairchild Channel F in 1976) instead of dedicated hardware with games physically built into the unit. The 2600 was bundled with two joystick controllers, a conjoined pair of paddle controllers, and a game cartridge: initially "Combat", and later "Pac-Man". The Atari VCS launched with nine simple, low-resolution games in 2 KiB cartridges. The system found its killer game with its version of Taito's "Space Invaders" in 1980 and became widely successful, leading to the creation of Activision and other third-party game developers as well as competition from home console manufacturers Mattel and Coleco. By the end of its primary lifecycle in 1983–84, games for the 2600 were using more than four times the ROM of the launch titles with significantly more advanced visuals and gameplay than the system was designed for, such as "Pitfall!" and its scrolling sequel "". Atari invested heavily in two games for the 2600, "Pac-Man," and "E.T. the Extra-Terrestrial", both of them being commercial failures that contributed to the video game crash of 1983, which ended the market relevance of the 2600. Warner sold off the home console division of Atari to Commodore CEO Jack Tramiel. In 1986, the new Atari Corporation under Tramiel released a lower-cost version of the 2600 and the backwards-compatible Atari 7800, but these were not enough to turn things around, and ultimately it was Nintendo that led the recovery of the industry. Atari finally ended production of the Atari 2600 on January 1, 1992. Across the system's lifetime, an estimated 30 million units were sold. Atari was founded by Nolan Bushnell and Ted Dabney in 1972. Their first major product was "Pong", one of the first successful arcade games. By 1975, Atari had released a "Pong" home console, competing against Magnavox, the only other major producer of home consoles at the time. Bushnell recognized, however, the limitation of custom logic burned onto the circuit board; this allowed only one game (and its variants), meaning that any new game would require a completely different console. Because development of a console cost at least and time to complete, and had only about a three-month shelf life before becoming outdated, it was a risky business model. By 1974, Atari had acquired Cyan Engineering, an electronics company founded by Steve Mayer and Larry Emmons (both former colleagues of Bushnell and Dabney from Ampex), and started Atari's Grass Valley Think Tank, where they developed new ideas for arcade games. Due to Bushnell's concern about single-game consoles, the Grass Valley team started working on a home console with multi-game support. Mayer and Emmons determined that a home console would require newly invented microprocessors to support multiple games, but such microprocessors cost at the time, far outside the range that their market would support. In September 1975, Chuck Peddle of MOS Technology created a low-cost replacement for the Motorola 6800, the MOS Technology 6502, and introduced it at the 1975 Wescon trade show in San Francisco. Mayer and Ron Milner attended the show, met with Peddle, and later invited Peddle to Cyan's headquarters to discuss using MOS's microprocessors for a game console. Mayer and Milner negotiated a deal for the 6502 chips at each, which was sufficient to begin development of a console; however, further discussions determined that the better solution would be the MOS Technology 6507, a more restrictive but lower-cost version of the 6502. Cyan and MOS also enlisted Synertek, a semiconductor manufacturer whose co-founder, Bob Schreiner, was good friends with Peddle, to act as a second source for the 6507. By December 1975, Atari hired Joe Decuir to help design the first prototype, codenamed "Stella" (the name of Decuir's bicycle). A second prototype was completed by March 1976 with the help of Jay Miner, who had managed to fit the entire Television Interface Adaptor (TIA), to send graphics and audio to the television display, into a single chip. The second prototype included the 6507, the TIA, and a ROM cartridge slot and adapter, each cartridge holding a ROM game image. Believing that Stella would be a success, Bushnell acquired the entire Grass Valley Think Tank and relocated them to Atari's new headquarters in Sunnyvale, California by mid-1976, putting Steve Mayer in charge of the project. Bushnell feared that once this unit was released, competitors would try to copy it, and preemptively arranged with all integrated chip manufacturers who were interested in the games market to deny sales to his competitors. Fairchild Semiconductor introduced its Fairchild Channel F home console in November 1976, beating Atari to the market with ROM cartridge technology. This pressured Atari to finish Stella more quickly, but the company lacked the funds to do so. Bushnell considered taking Atari public but instead sold the company to Warner Communications for ; subsequently, Warner provided approximately in new funding, allowing Stella to be prioritized and fast-tracked. By 1977, the product had advanced far enough to brand it as the "Atari Video Computer System" (VCS) and engage Atari's programmers to develop games for it. The unit was showcased in mid-1977 at the Summer Consumer Electronics Show with plans for retail release in October. However, Atari encountered production problems during its first batch, and its testing was complicated by the use of cartridges. Ultimately, the consoles were shipped to retailers in November 1977. At release in September 1977, the unit was originally priced at ($ adjusted for inflation), and shipped with two joysticks and a "Combat" cartridge (eight additional games were available at launch and sold separately). Atari sold between 350,000 and 400,000 Atari VCS units during 1977, attributed to the delay in shipping the units and consumers' unfamiliarity with non-dedicated consoles. 1978 was a challenging year. About 800,000 units were manufactured, but only 550,000 were sold by year's end. This required further financial support from Warner to cover losses and caused Bushnell to leave the company. Although Warren Robinett invented in this year the first action-adventure game, "Adventure", changing the fundamentals of gaming with the first game featuring a "virtual space bigger than the screen", Atari had suffered the loss of programmers David Crane, Bob Whitehead, Larry Kaplan, and Alan Miller, the company's "Fantastic Four", who had programmed most of the successful VCS games to that point. The four left Atari, disgruntled over Warner's oversight of the company and treatment of programmers, and formed Activision, which introduced third-party software to the VCS. The VCS was the best-selling console during the 1979 holiday season, with more than 1 million units sold, but Atari saw new competition from the Mattel Intellivision and Magnavox Odyssey², which also used ROM cartridges. Atari licensed the arcade hit "Space Invaders" by Taito. Its release in January 1980 doubled console sales to more than 2 million units. The VCS and its cartridges helped Atari earn more than $2 billion in 1980. Sales then doubled again for the next two years; by 1982, the console had sold 10 million units, while its best-selling game "Pac-Man" sold 7 million copies. The console also sold 450,000 units in West Germany by 1984. In 1982, Atari launched its second home console, which it named the Atari 5200; to standardize its naming, the VCS was renamed to the "Atari 2600 Video Computer System", or "Atari 2600" for short, derived from the manufacture part number CX2600. By 1982 the 2600 console cost Atari about $40 to make and sold for an average of $125. The company spent $4.50 to $6 to manufacture its cartridges, plus $1 to $2 for advertising, and sold them for $18.95 wholesale. Activision, formed by four former Atari VCS programmers, started developing third-party games to be used on cartridge systems, including the VCS, in 1979. Atari attempted to block the sale of the Activision cartridges, but lost in court, allowing other third-party VCS game developers to flourish. One such developer was Rob Fulop, who had created the VCS version of "Missile Command" while at Atari. More than two million copies of "Missile Command" were sold and, for this milestone, Atari rewarded him with a gift certificate for a free turkey. Fulop promptly left Atari. Fulop co-founded Imagic, and his first game for the company, "Demon Attack", became a hit in 1982. Other VCS-focused game development companies that sprang up in the early 1980s include US Games, Telesys, Games by Apollo, Data Age, Zimag, Mystique, and CommaVid. Mattel and Coleco, each already producing its own more advanced console, created simplified versions of their existing titles for the 2600. Mattel used the M Network brand name for its cartridges. Third-party titles competed with Atari's share of VCS games, accounting for half of VCS game sales by 1982. With third parties competing for market share, Atari worked to maintain dominance in the market by acquiring licenses for popular arcade games and other properties to make games from. "Pac-Man", despite having numerous technical flaws, still sold more than 10 million copies. Heading into the 1982 holiday season, Atari had placed high sales expectations on "E.T. the Extra-Terrestrial", a game programmed in about six weeks, as to capture the strong interest in the film. Atari produced an estimated four million cartridges, expecting the title to sell well. However, the game was poorly reviewed, leading to only about 1.5 million units sold. Atari, through Warner Communications, reported weaker results than expected in December 1982 to their shareholders, having expected a 50% year-to-year growth but only obtaining 10–15%. Coupled with the oversaturated market due to hardware and software competition, Atari's weakened position led investors to start pulling their funds out of the video game market, beginning a cascade of effects known as the video game crash of 1983 that devastated the North American industry. In September 1983, Atari sent 14 truckloads of unsold Atari 2600 cartridges and other equipment to be buried in a landfill in the New Mexico desert, known as Atari video game burial. While the burial had long been an urban legend, with reportedly millions of cartridges buried there, the site was excavated in 2014, with Atari representatives affirming that only about 700,000 cartridges had been deposited there. The burial remains recognized as a reminder of the end of Atari's dominance of the market. Warner, tired of supporting Atari, started looking for buyers in 1984. By mid-1984, software development for the 2600 had essentially stopped except for that of Atari and Activision. The 2600 was de-prioritized after Warner's 1984 sale of Atari, Inc.'s Consumer Division to Commodore founder Jack Tramiel; Tramiel shifted the business' focus to home computers and ended all development of console games. The North American video game market did not recover until about 1986, after Nintendo had introduced the Nintendo Entertainment System in North America. Atari Corporation released a redesigned model of the 2600 in 1986, supported by an ad campaign touting a price of "under 50 bucks". With a large library of games and a low price point, the 2600 continued to sell into the late 1980s. Atari released a final batch of titles in 1989–90 including "Secret Quest" and "Fatal Run". The final Atari-licensed release is the PAL-only port of the arcade game "KLAX" in 1990. After more than 14 years on the market, the 2600 line was formally discontinued on January 1, 1992, along with the Atari 7800 and Atari 8-bit family of home computers. The Atari 2600's CPU is the MOS Technology 6507, a version of the 6502, running at 1.19 MHz in the 2600. Though their internal silicon was identical, the 6507 was cheaper than the 6502 because its package included fewer memory-address pins—13 instead of 16. The designers of the Atari 2600 selected an inexpensive cartridge interface that has one fewer address than the 13 allowed by the 6507, further reducing the already limited addressable memory to 4 kiB (212 = 4096). This was believed to be sufficient as "Combat" is itself only 2 kiB. Later games circumvented this limitation with bank switching. The console has 128 bytes of RAM for scratch space, the call stack, and the state of the game environment. The top bezel of the console originally included six switches: power, TV type selection (color or black-and-white), game selection, player difficulty, and game reset. The difficulty switches were later moved to the back of the top bezel. The back bezel also included the controller ports, television output, and power adapter outlets. The 2600 does not have a frame buffer. Instead the video device provides two 8-pixel bitmapped sprites, two 1-pixel "missile" sprites, a 1-pixel "ball", and a 40-pixel "playfield" that is drawn by writing a bit pattern for each line into a register just before the television scans that line. As each line is scanned, a game must identify the non-sprite objects that overlap the next line, assemble the appropriate bit patterns to draw for those objects, and write the pattern into the register. Similar to its predecessor "Pong", the right side of the screen is a mirrored duplicate of the left; to control it separately, the software may modify the patterns as the scan line is drawn. After the controller scans the last active line, a slower vertical blanking interval begins, during which the game can process inputs and update the positions and states of objects in the game world. Any mistake in timing produces visual artifacts, a problem that programmers call "racing the beam". The 2600's video hardware is therefore highly flexible, but also challenging to program. One advantage the 2600 has over more powerful contemporary competitors such as the ColecoVision is that the 2600 has no protection against altering settings in mid-line. For example, although each sprite nominally has only one color, it is possible to color the rows differently by changing the sprite's color as it is drawn. If the two hardware sprites are not enough for a game, a developer may share one sprite among several objects (as with the ghosts in "Pac-Man") or draw software sprites, which is only a little more difficult than drawing a fixed playfield. The "Pitfall!" screenshot demonstrates some of these tricks: the player is a multi-color sprite, one sprite is multiplexed for the logs and the scorpion, and the swinging vine is drawn by shifting the position of the "ball" on each scan line. Warren Robinett, the programmer for "Adventure", described numerous tricks that he had to do to complete "Adventure" to both fit it within the memory limitations of the Atari 2600, as well as creatively use the graphics capability to create a multi-screen maze for players to navigate, using the "missile" sprite to create the maze's walls. The Atari 2600 was designed to be compatible with the cathode-ray tube television sets produced in the late 1970s and early 1980s, and uses different color palettes depending on the television signal format. With the NTSC format, a 128-color palette was available; while in PAL, only 104 colors; in SECAM, only 8. A side effect of how graphics work on the 2600 is that all games are required to run at either 60 frames per second (NTSC) or 50 frames per second (PAL). The VCS originally shipped with two types of controllers: a joystick (part number CX10) and pair of rotary paddle controllers (CX30). Driving controllers, which are similar to paddle controllers but can be continuously rotated, shipped with the "Indy 500" launch game. After less than a year, the CX10 joystick was replaced with the CX40 model designed by James C. Asher. Because the Atari joystick port and CX40 joystick became industry standards, 2600 joysticks and some other peripherals work with later systems, including the MSX, Commodore 64, Amiga, Atari 8-bit family, and Atari ST. The CX40 joystick can be used with the Master System and Sega Genesis, but does not provide all the buttons of a native controller. Third-party controllers include Wico's Command Control joystick. Later, the CX42 Remote Control Joysticks, similar in appearance but using wireless technology, were released, together with a receiver whose wires could be inserted in the controller jacks. Atari introduced the CX50 Keyboard Controller in June 1978 along with two games that required it: "Codebreaker" and "Hunt & Score". The similar, but simpler, CX23 Kid's Controller was released later for a series of games aimed at a younger audience. The CX22 Trak-Ball controller was announced in January 1983 and was compatible with the Atari 8-bit family. There were two attempts to turn the Atari 2600 into a keyboard-equipped home computer: Atari's never-released CX3000 "Graduate" keyboard, and the CompuMate keyboard by Spectravideo which was released in 1983. The initial production of the VCS was made in Sunnyvale, using thick plastic for the casing, and with all six console switches on the front panel. Production of the unit was moved to Hong Kong in 1979, where a thinner plastic was used for the casing, reducing the system's weight. These two versions are commonly referred to as "Heavy Sixers" and "Light Sixers" respectively, representing the six front switches. In 1980, the difficulty switches were moved to the back of the console, leaving four switches on the front. Other than this, these four-switch consoles looked nearly identical to the earlier six-switch models. In 1982 Atari rebranded the console as the "Atari 2600", a name first used on a version of the four-switch model without woodgrain, giving it an all-black appearance. Atari continued their OEM relationship with Sears under the latter's Tele-Games brand, which started in 1975 with the original "Pong". (The company Telegames, which later produced cartridges for the 2600, is unrelated.) Sears released several models of the VCS as the Sears Video Arcade series starting in 1977. In 1983, the previously Japan-only Atari 2800 was rebranded as the Sears Video Arcade II. Sears released versions of Atari's games with Tele-Games branding, usually with different titles. Three games were produced by Atari for Sears as exclusive releases: "Steeplechase", "Stellar Track", and "Submarine Commander". The Atari 2800 is the Japanese version of the 2600 released in October 1983. It was the first release of a 2600 designed specifically for the Japanese market, though companies like Epoch had distributed the 2600 in Japan previously. It was released a short time after Nintendo's Family Computer, which became the dominant console in Japan, and the 2800 did not gain a significant share of the market. Sears released the 2800 in the US in 1983 as the Sears Video Arcade II packaged with two controllers and "Space Invaders". Around 30 specially branded games were released for the 2800. Designed by engineer Joe Tilly, the 2800 has four controller ports instead of the two of the 2600. The controllers are an all-in one design using a combination of an 8-direction digital joystick and a 270-degree paddle, designed by John Amber. The 2800's case design departed from the 2600, using a wedge shape with non-protruding switches. The case style was used as the basis for the Atari 7800's case designed by Barney Huang. The 1986 model, unofficially referred to as "the 2600 Jr.", features a smaller, cost-reduced form factor with a modernized, Atari 7800-like appearance. The redesigned 2600 was advertised as a budget gaming system (under US$49.99) with the ability to run a large collection of classic games. Though released after the video game crash of 1983, and after the North American launch of the Nintendo Entertainment System, the 2600 was supported with new games and television commercials promoting "The fun is back!" Later European versions of the 2600 Jr. include a joypad, which is also featured with the European 7800. Atari released several minor stylistic variations of the 2600 Jr. design: the "large rainbow" (shown), "short rainbow", and an all-black version sold only in Ireland. In 1977, nine games were released on cartridge to accompany the launch of the machine, including "Air-Sea Battle", "Star Ship", and "Street Racer". Atari, Inc. was the only developer for the first few years of the VCS, releasing "Adventure" (often credited the first action-adventure game), "Breakout", and dozens of other games. Rick Maurer's port of Taito's "Space Invaders", released in 1980, is the first VCS title to have more than one million copies sold—eventually doubling that. It became the killer app to drive console sales. Versions of Atari's own "Asteroids" and "Missile Command" arcade games, released the following year, were also major hits. Each early VCS game is in a 2K ROM. Later games like "Space Invaders", and even "River Raid" from 1982, increased this capacity to 4K. The VCS port of "Asteroids" (1981) is the first game for the system to have used 8K via a bank switching technique. Some later releases, including Atari's ports of "Dig Dug" and "Crystal Castles", are 16K cartridges. Two Atari published games, both from 1982, "E.T. the Extra-Terrestrial" and "Pac-Man", are frequently blamed for contributing to the video game crash of 1983. To clear inventory and save money in wake of the crash, Atari shipped about 700,000 Atari 2600 games, including unsold lots of "Pac-Man" and "E.T.", to a landfill in Alamogordo, New Mexico in 1983. Until 2014, there was skepticism of how much Atari had buried there, in some cases estimates of millions of cartridges, creating an urban legend of an Atari video game burial. The legend was verified in 2014 when interested parties were able to dig up a portion of the landfill and confirmed the presence of unsold Atari games. A company named Mystique produced a number of pornographic games for the 2600. The most notorious of these, "Custer's Revenge", was protested by women's and Native American groups because it depicted General George Armstrong Custer raping a bound Native American woman. Atari sued Mystique in court over the release of the game. The Atari VCS—and later as the 2600—was so successful in the late 1970s and early 1980s that "Atari" was a synonym for the system in mainstream media and, by extension, for video games in general. Jay Miner directed the creation of the successors to the 2600's TIA chip—CTIA and ANTIC—which were central to the Atari 8-bit computers released in 1979 and later the Atari 5200 console. The Atari 2600 was inducted into the National Toy Hall of Fame at The Strong in Rochester, New York, in 2007. In 2009, the Atari 2600 was named the number two console of all time by IGN, who cited its remarkable role as the console behind both the first video game boom and the video game crash of 1983, and called it "the console that our entire industry is built upon". Modern Atari 2600 clones remain on the market. The Atari Classics 10-in-1 TV Game, manufactured by Jakks Pacific, emulates the 2600 console and includes converted versions of 10 games into a single Atari-brand-lookalike joystick with composite-video outputs for connecting directly to modern televisions or VCRs. The TV Boy includes 127 games in an enlarged joypad. The Atari Flashback 2 console, released in 2005, contains 40 games (with four additional programs unlocked by a cheat code). The console implements the original 2600 architecture and can be modified to play original 2600 cartridges by adding a cartridge port, and is also compatible with original 2600 controllers. In 2017, Hyperkin announced the RetroN 77, a clone of the Atari 2600 that plays original cartridges instead of pre-installed games. The Atari VCS microconsole, planned for release from Atari Interactive by 2019, is designed on a miniaturization of the Atari 2600, including the option for its faux wood paneling. This microconsole is designed to play several Atari 2600 games under emulation on a Linux operating system, as well as potentially other games for computers and other gaming systems. The Atari 2700 is a version of the 2600 which uses wireless controllers. An Atari 2600 variant, known by its production code "CX2000" and nickname "Val", was found as two 1982 prototypes at the New York and Sunnyvale Atari facilities, respectively. It was an attempt to make a new version of the aging 2600. Its design, with two integrated joystick controllers, was the result of human factor analysis by Henry Dreyfuss Associates. The project never saw market production. Atari started work on a 2600 successor called the "Atari 3200". It was to be compatible with 2600 cartridges, and was rumored to be based on a 10-bit processor, although design documents show it was to be based on the 8-bit 6502. It was unfinished when preliminary game developers discovered that it was difficult to program. Atari cloned the Atari 3200 into the Sears Super Arcade II, but this was never released.
https://en.wikipedia.org/wiki?curid=2779