id
stringlengths
2
8
url
stringlengths
31
117
title
stringlengths
1
71
text
stringlengths
153
118k
topic
stringclasses
4 values
section
stringlengths
4
49
sublist
stringclasses
9 values
2669531
https://en.wikipedia.org/wiki/Cosmic%20neutrino%20background
Cosmic neutrino background
The cosmic neutrino background (CNB or CB) is the universe's background particle radiation composed of neutrinos. They are sometimes known as relic neutrinos. The CB is a relic of the Big Bang; while the cosmic microwave background radiation (CMB) dates from when the universe was 379,000 years old, the CB decoupled (separated) from matter when the universe was just one second old. It is estimated that today, the CB has a temperature of roughly . As neutrinos rarely interact with matter, these neutrinos still exist today. They have a very low energy, around 10 to 10 eV. Even high energy neutrinos are notoriously difficult to detect, and the CB has energies around 1010 times smaller, so the CB may not be directly observed in detail for many years, if at all. However, Big Bang cosmology makes many predictions about the CB, and there is very strong indirect evidence that the CB exists. Temperature estimation Given the temperature of the cosmic microwave background (CMB) the temperature of the cosmic neutrino background (CB) can be estimated. It involves a change between two regimes: Regime 1 The original state of the universe is a thermal equilibrium, the final stage of which has photons and leptons freely creating each other through annihilation (leptons create photons) and pair production (photons create leptons). This was the very brief state, right after the Big Bang. Its last stage involves only the lowest-mass possible fermions that interact with photons: electrons and positrons. Regime 2 Once the universe has expanded enough that the photon+lepton plasma has cooled to the point that Big Bang photons no longer have enough energy for pair production of the lowest mass/energy leptons, the remaining electron–positron pairs annihilate. The photons they create are cool, and are then unable to create new particle pairs. This is the current state of most of the universe. At very high temperatures, before neutrinos decoupled from the rest of matter, the universe primarily consisted of neutrinos, electrons, positrons, and photons, all in thermal equilibrium with each other. Once the temperature dropped to approximately ( K), the neutrinos decoupled from the rest of matter, and for practical purposes, all lepton and photon interactions with these neutrinos stopped. Despite this decoupling, neutrinos and photons remained at the same temperature as the universe expanded as a "fossil" of the prior Regime 1, since both are cooled in the same way by the same process of cosmic expansion, from the same starting temperature. However, when the temperature dropped below double the mass of the electron, most electrons and positrons annihilated, transferring their heat and entropy to photons, thus increasing the temperature of the photons. So the ratio of the temperature of the photons before and after the electron–positron annihilation is the same as the ratio of the temperature of the neutrinos and the photons in the current Regime 2. To find this ratio, we assume that the entropy    of the universe was approximately conserved by the electron–positron annihilation. Then using where    is the effective number of degrees of freedom and is the plasma or photon temperature. Once reactions cease, the entropy    should remain approximately "stuck" for all temperatures below the cut-off temperature, and we find that Here denotes the lowest temperature where pair production and annihilation were in equilibrium; and denotes the temperature after the temperature fell below the regime-shift temperature , after the remaining, but no longer refreshed, electron–positron pairs had annihilated and contributed to the total photon energy. The related temperatures and are the simultaneous temperatures of the photons () and neutrinos () respectively, whose ratio stays "stuck" at the same value indefinitely, after The factor is determined by a sum, based on the particle species engaged in the original equilibrium reaction: +   2   for each photon (or other massless bosons, if any). +     for each electron, positron, or other fermion. Whereas the factor is simply   2,   since the present regime only concerns photons, in thermal equilibrium with at most themselves. So Since the cosmic photon background temperature at present has cooled to it follows that the neutrino background temperature is currently The above discussion is technically valid for massless neutrinos, which are always relativistic. For neutrinos with a non-zero rest mass, at low temperature where the neutrinos become non-relativistic, a description in terms of a temperature is not appropriate. In other words, when the neutrinos' thermal energy ( is the Boltzmann constant) falls below the rest mass energy in a low-temperature case one should instead speak of the neutrinos' collective energy density, which remains both relevant and well-defined. Indirect evidence Relativistic neutrinos contribute to the radiation energy density of the universe , typically parameterized in terms of the effective number of neutrino species : where denotes the redshift. The first term in the square brackets is due to the CMB, the second comes from the CB. The Standard Model with its three neutrino species predicts a value of , including a small correction caused by a non-thermal distortion of the spectra during e×e annihilation. The radiation density had a major impact on various physical processes in the early universe, leaving potentially detectable imprints on measurable quantities, thus allowing us to infer the value of from observations. Big Bang nucleosynthesis Due to its effect on the expansion rate of the universe during Big Bang nucleosynthesis (BBN), the theoretical expectations for the primordial abundances of light elements depend on Astrophysical measurements of the primordial and abundances lead to a value of = at 68% c.l., in very good agreement with the Standard Model expectation. From the cosmic microwave background Anisotropies and structure formation The presence of the CB affects the evolution of CMB anisotropies as well as the growth of matter perturbations in two ways: Due to its contribution to the radiation density of the universe (which determines for instance the time of matter–radiation equality), and due to the neutrinos' anisotropic stress which dampens the acoustic oscillations of the spectra. Additionally, free-streaming massive neutrinos suppress the growth of structure on small scales. The WMAP spacecraft's five-year data combined with type Ia supernova data and information about the baryon acoustic oscillation scale yielded = at 68% c.l., providing an independent confirmation of the BBN constraints. The Planck spacecraft collaboration has published the tightest bound to date on the effective number of neutrino species, at = . Phase changes Big Bang cosmology makes many predictions about the CB, and there is very strong indirect evidence that the cosmic neutrino background exists, both from Big Bang nucleosynthesis predictions of the helium abundance, and from anisotropies in the cosmic microwave background. One of these predictions is that neutrinos will have left a subtle imprint on the cosmic microwave background (CMB). It is well known that the CMB has irregularities. Some of the CMB fluctuations were roughly regularly spaced, because of the effect of baryon acoustic oscillation. In theory, the decoupled neutrinos should have had a very slight effect on the phase of the various CMB fluctuations. In 2015, it was reported that such shifts had been detected in the CMB. Moreover, the fluctuations corresponded to neutrinos of almost exactly the temperature predicted by Big Bang theory ( compared to a prediction of 1.95 K), and exactly three types of neutrino, the same number of neutrino flavours currently predicted by the Standard Model. Prospects for the direct detection Confirmation of the existence of these relic neutrinos may only be possible by directly detecting them using experiments on Earth. This will be difficult as the neutrinos which make up the CB are non-relativistic, in addition to interacting only weakly with normal matter, and so any effect they have in a detector will be hard to identify. One proposed method of direct detection of the CB is to use the capture of cosmic relic neutrinos on tritium i.e. 3H, leading to an induced form of beta decay. The neutrinos of the CB would lead to the production of electrons via the reaction while the main background comes from electrons produced via natural beta decay These electrons would be detected by the experimental apparatus in order to measure the size of the CB. The latter source of electrons is far more numerous, however, their maximum energy is smaller than the average energy of the CB-electrons by twice the average neutrino mass. Since this mass is tiny, of the order of a few eVs or less, such a detector must have an excellent energy resolution in order to separate the signal from the background. One such proposed experiment is called PTOLEMY, which will be made up of 100 g of tritium target. The detector demonstrator (with about 0.2 g of tritium) should be ready by 2025.
Physical sciences
Physical cosmology
Astronomy
2674010
https://en.wikipedia.org/wiki/Salt%20metathesis%20reaction
Salt metathesis reaction
A salt metathesis reaction is a chemical process involving the exchange of bonds between two reacting chemical species which results in the creation of products with similar or identical bonding affiliations. This reaction is represented by the general scheme: AB + CD -> AD + CB Typical examples are the reactions between oxysalts and binary compounds such as salts, hydrohalic acids and metal hydroxides: \mathit{ab}A_\mathit{x}(BO_\mathit{y})_\mathit{z}{} + \mathit{xz}C_\mathit{a}D_\mathit{b}{} -> \mathit{bx}A_\mathit{a}D_\mathit{z}{} + \mathit{az}C_\mathit{x}(BO_\mathit{y})_\mathit{b}{} Another classical example are the reactions between oxysalts in solution: \mathit{ab}A_\mathit{x}(BO_\mathit{p})_\mathit{y}{} + \mathit{xy}C_\mathit{a}(DO_\mathit{q})_\mathit{b}{} -> \mathit{bx}A_\mathit{a}(DO_\mathit{q})_\mathit{y}{} + \mathit{ay}C_\mathit{x}(BO_\mathit{p})_\mathit{b}{} The bond between the reacting species can be either ionic or covalent. Classically, these reactions result in the precipitation of one product. In older literature, the term double decomposition is common. The term double decomposition is more specifically used when at least one of the substances does not dissolve in the solvent, as the ligand or ion exchange takes place in the solid state of the reactant. For example: AX(aq) + BY(s) → AY(aq) + BX(s). Types of reactions Counterion exchange Salt metathesis is a common technique for exchanging counterions. The choice of reactants is guided by a solubility chart or lattice energy. HSAB theory can also be used to predict the products of a metathesis reaction. Salt metathesis is often employed to obtain salts that are soluble in organic solvents. Illustrative is the conversion of sodium perrhenate to the tetrabutylammonium salt: NaReO4 + N(C4H9)4Cl → N(C4H9)4[ReO4] + NaCl The tetrabutylammonium salt precipitates from the aqueous solution. It is soluble in dichloromethane. Salt metathesis can be conducted in nonaqueous solution, illustrated by the conversion of ferrocenium tetrafluoroborate to a more lipophilic salt containing the tetrakis(pentafluorophenyl)borate anion: [Fe(C5H5)2]BF4 + NaB(C6F5)4 → [Fe(C5H5)2]B(C6F5)4 + NaBF4 When the reaction is conducted in dichloromethane, the salt NaBF4 precipitates and the B(C6F5)4- salt remains in solution. Metathesis reactions can occur between two inorganic salts when one product is insoluble in water. For example, the precipitation of silver chloride from a mixture of silver nitrate and cobalt hexammine chloride delivers the nitrate salt of the cobalt complex: 3 + [Co(NH3)6]Cl3 → 3 AgCl + [Co(NH3)6](NO3)3 The reactants need not be highly soluble for metathesis reactions to take place. For example barium thiocyanate forms when boiling a slurry of copper(I) thiocyanate and barium hydroxide in water: + 2 → + 2CuOH Alkylation Metal complexes are alkylated via salt metathesis reactions. Illustrative is the methylation of titanocene dichloride to give the Petasis reagent: (C5H5)2TiCl2 + 2 ClMgCH3 → (C5H5)2Ti(CH3)2 + 2 MgCl2 The salt product typically precipitates from the reaction solvent. Neutralization reaction A neutralization reaction is a type of double replacement reaction. A neutralization reaction occurs when an acid reacts with an equal amount of a base. This reaction usually produces a salt. One example, hydrochloric acid reacts with disodium iron tetracarbonyl to produce the iron dihydride: Reaction between an acid and a carbonate or bicarbonate salt yields carbonic acid, which spontaneously decomposes into carbon dioxide and water. The release of carbon dioxide gas from the reaction mixture drives the reaction to completion. For example, a common, science-fair "volcano" reaction involves the reaction of hydrochloric acid with sodium carbonate: Salt-free metathesis reaction In contrast to salt metathesis reactions, which are driven by the precipitation of solid salts, are salt-free reductions, which are driven by formation of silyl halides, Salt-free metathesis reactions proceed homogeneously.
Physical sciences
Other reactions
Chemistry
3628220
https://en.wikipedia.org/wiki/Tyrannosauroidea
Tyrannosauroidea
Tyrannosauroidea (meaning 'tyrant lizard forms') is a superfamily (or clade) of coelurosaurian theropod dinosaurs that includes the family Tyrannosauridae as well as more basal relatives. Tyrannosauroids lived on the Laurasian supercontinent beginning in the Jurassic Period. By the end of the Cretaceous Period, tyrannosauroids were the dominant large predators in the Northern Hemisphere, culminating in the gigantic Tyrannosaurus. Fossils of tyrannosauroids have been recovered on what are now the continents of North America, Europe and Asia, with fragmentary remains possibly attributable to tyrannosaurs also known from South America and Australia. Tyrannosauroids were bipedal carnivores, as were most theropods, and were characterized by numerous skeletal features, especially of the skull and pelvis. Early in their existence, tyrannosauroids were small predators with long, three-fingered forelimbs. Late Cretaceous genera became much larger, including some of the largest land-based predators ever to exist, but most of these later genera had proportionately small forelimbs with only two digits. Primitive feathers have been identified in fossils of two species and may have been present in other tyrannosauroids as well. Prominent bony crests in a variety of shapes and sizes on the skulls of many tyrannosauroids may have served display functions. Description Tyrannosauroids varied widely in size, although there was a general trend towards increasing size over time. Early tyrannosauroids were small animals. One specimen of Dilong, almost fully grown, measured in length, and a fully grown Guanlong measured long. Teeth from Lower Cretaceous rocks (140 to 136 million years old) of Hyogo, Japan, appear to have come from an approximately long animal, possibly indicating an early size increase in the lineage. An immature Eotyrannus was over in length, and a subadult Appalachiosaurus was estimated at more than long, indicating that both genera reached larger sizes. The Late Cretaceous tyrannosaurids ranged from the Albertosaurus and Gorgosaurus to Tyrannosaurus, which exceeded in length and may have weighed more than 6,400 kilograms (7 short tons). A 2010 review of the literature concluded that tyrannosaurs were "small- to mid-sized" for their first 80 million years but were "some of the largest terrestrial carnivores to ever live" in their last 20 million years. Skulls of early tyrannosauroids were long, low and lightly constructed, similar to other coelurosaurs, while later forms had taller and more massive skulls. Despite the differences in form, certain skull features are found in all known tyrannosauroids. The premaxillary bone is very tall, blunting the front of the snout, a feature which evolved convergently in abelisaurids. The nasal bones are characteristically fused, arched slightly upwards and often very roughly textured on their upper surface. The premaxillary teeth at the front of the upper jaw are shaped differently from the rest of the teeth, smaller in size and with a D-shaped cross section. In the lower jaw, a prominent ridge on the surangular bone extends sideways from just below the jaw joint, except in the basal Guanlong. Tyrannosauroids had S-shaped necks and long tails, as did most other theropods. Early genera had long forelimbs, about 60% the length of the hindlimb in Guanlong, with the typical three digits of coelurosaurs. The long forelimb persisted at least through the Early Cretaceous Eotyrannus, but is unknown in Appalachiosaurus. Derived tyrannosaurids have forelimbs strongly reduced in size, the most extreme example being Tarbosaurus from Mongolia, where the humerus was only one-quarter the length of the femur. The third digit of the forelimb was also reduced over time. This digit was unreduced in the basal Guanlong, while in Dilong it was more slender than the other two digits. Eotyrannus also had three functional digits on each hand. Tyrannosaurids had only two, although the vestigial metacarpal of the third are preserved in some well-preserved specimens. As in most coelurosaurs, the second digit of the hand is the largest, even when the third digit is not present. Characteristic features of the tyrannosauroid pelvis include a concave notch at the upper front end of the ilium, a sharply defined vertical ridge on the outside surface of the ilium, extending upwards from the acetabulum (hip socket), and a huge "boot" on the end of the pubis, more than half as long as the shaft of the pubis itself. These features are found in all known tyrannosauroids, including basal members Guanlong and Dilong. The pubis is not known in Aviatyrannis or Stokesosaurus but both show typical tyrannosauroid characters in the ilium. The hindlimbs of all tyrannosauroids, like most theropods, had four toes, although the first toe (the hallux) did not contact the ground. Tyrannosauroid hindlimbs are longer relative to body size than almost any other theropods, and show proportions characteristic of fast-running animals, including elongated tibiae and metatarsals. These proportions persist even in the largest adult Tyrannosaurus, despite its probable inability to run. The third metatarsal of tyrannosaurids was pinched at the top between the second and fourth, forming a structure known as the arctometatarsus. The arctometatarsus was also present in Appalachiosaurus but it is unclear whether it was found in Eotyrannus or Dryptosaurus. This structure was shared by derived ornithomimids, troodontids and caenagnathids, but was not present in basal tyrannosauroids like Dilong paradoxus, indicating convergent evolution. Classification Tyrannosaurus was named by Henry Fairfield Osborn in 1905, along with the family Tyrannosauridae. The name is derived from the Ancient Greek words ('tyrant') and ('lizard'). The superfamily name Tyrannosauroidea was first published in a 1964 paper by the British paleontologist Alick Walker. The suffix -oidea, commonly used in the name of animal superfamilies, is derived from the Greek ειδος ('form'). Scientists have commonly understood Tyrannosauroidea to include the tyrannosaurids and their immediate ancestors. With the advent of phylogenetic taxonomy in vertebrate paleontology, however, the clade has received several more explicit definitions. The first was by Paul Sereno in 1998, where Tyrannosauroidea was defined as a stem-based taxon including all species sharing a more recent common ancestor with Tyrannosaurus rex than with neornithean birds. To make the family more exclusive, Thomas Holtz redefined it in 2004 to include all species more closely related to Tyrannosaurus rex than to Ornithomimus velox, Deinonychus antirrhopus or Allosaurus fragilis. Sereno published a new definition in 2005, using Ornithomimus edmontonicus, Velociraptor mongoliensis and Troodon formosus as external specifiers. The Sereno definition was adopted in a 2010 review. Some studies have suggested that the clade Megaraptora, usually considered to be allosauroids, are basal tyrannosauroids. However, other authors disputed the placement of megaraptorans within Tyrannosauroidea, and a study of megaraptoran hand anatomy published in 2016 caused even the original scientists suggesting their tyrannosauroid relationships to at least partly reject their prior conclusion. Phylogeny While paleontologists have long recognized the family Tyrannosauridae, its ancestry has been the subject of much debate. For most of the twentieth century, tyrannosaurids were commonly accepted as members of the Carnosauria, which included almost all large theropods. Within this group, the allosaurids were often considered to be ancestral to tyrannosaurids. In the early 1990s, cladistic analyses instead began to place tyrannosaurids into the Coelurosauria, echoing suggestions first published in the 1920s. Tyrannosaurids are now universally considered to be large coelurosaurs. In 1994, Holtz grouped tyrannosauroids with elmisaurids, ornithomimosaurs and troodonts into a coelurosaurian clade called Arctometatarsalia based on a common ankle structure where the second and fourth metatarsals meet near the tarsal bones, covering the third metatarsal when viewed from the front. Basal tyrannosauroids like Dilong, however, were found with non-arctometatarsalian ankles, indicating that this feature evolved convergently. Arctometatarsalia has been dismantled and is no longer used by most paleontologists, with tyrannosauroids usually considered to be basal coelurosaurs outside Maniraptoriformes. While many place tyrannosauroids as basal coelurosaurs, Paul Sereno in his 1990s analysis of theropods would find the Tyrannosaurs to be sister taxa to the Maniraptora with them being closer to birds than Ornithomimosaurs were. He called this group Tyrannoraptora (which in the absence of papers that recover a Tyrannosaur-maniraptoran clade), is a clade which contains most Coelurosaurs. A 2007 analysis found the family Coeluridae, including the Late Jurassic North American genera Coelurus and Tanycolagreus, to be the sister group of Tyrannosauroidea. The most basal tyrannosauroid known from complete skeletal remains is Guanlong, a representative of the family Proceratosauridae. Other early taxa include Stokesosaurus and Aviatyrannis, known from far less complete material. The better-known Dilong is considered slightly more derived than Guanlong and Stokesosaurus. Dryptosaurus, long a difficult genus to classify, has turned up in several recent analyses as a basal tyrannosauroid as well, slightly more distantly related to Tyrannosauridae than Eotyrannus and Appalachiosaurus. Alectrosaurus, a poorly known genus from Mongolia, is definitely a tyrannosauroid but its exact relationships are unclear. Other taxa have been considered possible tyrannosauroids by various authors, including Bagaraatan and Labocania. Siamotyrannus from the Early Cretaceous of Thailand was originally described as an early tyrannosaurid, but is usually considered a carnosaur today. Iliosuchus has a vertical ridge on the ilium reminiscent of tyrannosauroids and may in fact be the earliest known member of the superfamily, but not enough material is known to be sure. Below on the left is a cladogram of Tyrannosauroidea from a 2022 study by Darren Naish and Andrea Cau on the genus Eotyrannus, and on the right is a cladogram of Eutyrannosauria from a 2020 study by Jared T. Voris and colleagues on the genus Thanatotheristes: Phylogeography In 2018 authors Rafael Delcourt and Orlando Nelson Grillo published a phylogenetic analysis of Tyrannosauroidea which incorporated taxa from the ancient continent of Gondwana (which today consists of the southern hemisphere), such as Santanaraptor and Timimus, whose placement in the group has been controversial. They have found that not only Santanaraptor and Timimus were placed as tyrannosaurs more derived than Dilong, but they have found in their analysis that tyrannosauroids were widespread in Laurasia and Gondwana since the Middle Jurassic. They have proposed new subclade names for Tyrannosaurioidea. The first is Pantyrannosauria referring to all non-proceratosaurid members of the group, while Eutyrannosauria for larger tyrannosaur taxa found in the northern hemisphere such as Dryptosaurus, Appalachiosaurus, Bistahieversor, and Tyrannosauridae. Below is their phylogeographic tree they have recovered, in which displays the phylogenetic relationships of the taxa as well as the continents those taxa have been found. In 2021, Chase Brownstein published a research article based on more thorough descriptions of tyrannosauroid metatarsals and vertebra from the Merchantville Formation in Delaware. This reanalysis of phylogenetic relationships of tyrannosauroids in Appalachia has brought the rediscovery of the clade Dryptosauridae due to the similarities of metatarsals II and IV with the same bones in the Dryptosaurus holotype. However. the Merchantville taxon was found to still be different enough to separate it on the genus level from Dryptosaurus. In the phylogentic tree constructed Dryptosauridae is found to be a valid family of non tyrannosaurid eutyrannosaur. It currently sits in a polytomy with the Iren Dabasu taxon and more basal eutryannosaurs. Distribution The tyrannosauroids lived on the supercontinent Laurasia, which split from Gondwana in the Middle Jurassic. The earliest recognized tyrannosauroids lived in the Middle Jurassic, represented by the proceratosaurids Kileskus from the Western Siberia and Proceratosaurus from Great Britain. Upper Jurassic tyrannosauroids include Guanlong from China, Stokesosaurus from the western United States and Aviatyrannis and Juratyrant from Europe. Early Cretaceous tyrannosauroids are known from Laurasia, being represented by Eotyrannus from England and Dilong, Sinotyrannus, and Yutyrannus from northeastern China. Early Cretaceous tyrannosauroid premaxillary teeth are known from the Cedar Mountain Formation in Utah and the Tetori Group of Japan. The Middle Cretaceous record of Tyrannosauroidea is rather patchy. Teeth and indeterminate postcrania of this interval are known from the Cenomanian-age Dakota Formation of western North America and Potomac Formation of New Jersey, as well as formations in Kazakhstan and Tajikistan; two genera, Timurlengia and Xiongguanlong, have been found in Asia, while the Brazilian Santanaraptor may belong to this group. Suskityrannus has been found in the Moreno Hill Formation of the Zuni Basin of western New Mexico. The first unquestionable remains of tyrannosaurids occur in the Campanian stage of the Late Cretaceous in North America and Asia. Two subfamilies are recognized. The albertosaurines are only known from North America, while the tyrannosaurines are found on both continents. Tyrannosaurid fossils have been found in Alaska, which may have served as a land bridge allowing dispersal between the two continents. Non-tyrannosaurid tyrannosauroids like Alectrosaurus and possibly Bagaraatan were contemporaneous with tyrannosaurids in Asia, while they are absent from western North America. Eastern North America was divided by the Western Interior Seaway in the middle of the Cretaceous and isolated from the western portion of the continent. The absence of tyrannosaurids from the eastern part of the continent suggests that the family evolved after the appearance of the seaway, allowing basal tyrannosauroids like Dryptosaurus and Appalachiosaurus to survive in the east as a relict population until the end of the Cretaceous. Basal tyrannosauroids have also been suggested to be present in Australia and South America during the Early Cretaceous. NMV P186069, a partial pubis (a hip bone) with a supposed distinctive tyrannosauroid-like form, was discovered in Dinosaur Cove in Victoria. However, a response suggested that critical tyrannosauroid characters were absent from the fossil. The Australian taxon Timimus, known from a femur, and the Brazilian Santanaraptor, known from a partial juvenile skeleton, have also been suggested to be tyrannosaurs. However, these placements have been considered questionable, with the supposed tyrannosauroid characters of Santanaraptor being widely distributed within Coelurosauria, in other characters having similarities to noasaurids. Paleobiology Facial tissue A conference paper by Tracy Ford states that there was rough bone texture on the skulls of theropods and higher foramina frequency than lepidosaurs and mammals which would be evidential for a sensitive snout for theropods. A study in 2017 study on a new tyrannosaurid named Daspletosaurus horneri was published in the journal Scientific Reports, where paleontologist Thomas Carr analyzed the craniofacial texture of Daspletosaurus horneri and observed a hummocky rugosity which compared to crocodilian skulls, suggesting Daspletosaurus horneri and with it all tyrannosaurids have flat sensory scales. The subordinate regions were analyzed to have cornified epidermis. However, a 2018 presentation has an alternative interpretation. Crocodilians do not have flat sensory scales, but rather cracked, cornified epidermis due to growth. The hummocky rugosity in the skulls of lepidosaurs have correlation with scales which this bone texture is also present in tyrannosaurid skulls. The foramina frequency in theropod skulls does not exceed 50 foramina, which shows that theropods had lips. It's been proposed that lips are a primitive trait in tetrapods and the soft tissue present in crocodilians are a derived trait because of aquatic or semiaquatic adaptations. Body integument Long filamentous structures have been preserved along with skeletal remains of numerous coelurosaurs from the Early Cretaceous Yixian Formation and other nearby geological formations from Liaoning, China. These filaments have usually been interpreted as "protofeathers," homologous with the branched feathers found in birds and some non-avian theropods, although other hypotheses have been proposed. A skeleton of Dilong paradoxus was described in 2004 that included the first example of feathers in a tyrannosauroid. Similarly to down feathers of modern birds, the feathers found in Dilong were branched but not pennaceous, and may have been used for insulation. Even large tyrannosauroids have been found with evidence of feathers. Yutyrannus huali, also from the Yixian Formation, is known from three specimens, each preserving traces of feathers on various parts of the body. While not all areas of the body preserve impressions across all three specimens, these fossils demonstrate that even in this medium-sized species, most of the body was covered in feathers. The presence of feathers in basal tyrannosauroids is not surprising since they are now known to be characteristic of coelurosaurs, found in other basal genera like Sinosauropteryx, as well as all more derived groups. Rare fossilized skin impressions of some Late Cretaceous tyrannosaurids lack feathers, however, instead showing skin covered in fine, non-overlapping scales. Possibly, feathers were present on other areas of the body: preserved skin impressions are very small and come primarily from the legs, pelvic region, and underside of the tail, which either lack feathers or only covered in a light down in some modern large ground-dwelling birds. Alternatively, secondary loss of feathers in large tyrannosaurids may be analogous with the similar loss of hair in the largest modern mammals like elephants, where a low surface area-to-volume ratio slows down heat transfer, making insulation by a coat of hair unnecessary or even detrimental. A scientific publication by Phil Bell and colleagues in 2017 show that tyrannosaurids such as Gorgosaurus, Tarbosaurus, Albertosaurus, Daspletosaurus, and Tyrannosaurus had scales. The Bell et al. 2017 paper notes that the scale-like integument on bird feet were actually secondarily derived feathers according to paleontological and evolutionary developmental evidence so they hypothesize that the scaly skin preserved on some tyrannosaurid specimens might be secondarily derived from filamentous appendages like on Yutyrannus although strong evidence is needed to support this hypothesis. However, other paleontologists argue that taphonomy is the possible cause of the lack of filamentous structures in tyrannosaurid fossils. Head crests Bony crests are found on the skulls of many theropods, including numerous tyrannosauroids. The most elaborate is found in Guanlong, where the nasal bones support a single, large crest which runs along the midline of the skull from front to back. This crest was penetrated by several large foramina (openings) which reduced its weight. A less prominent crest is found in Dilong, where low, parallel ridges run along each side of the skull, supported by the nasal and lacrimal bones. These ridges curve inwards and meet just behind the nostrils, making the crest Y-shaped. The fused nasals of tyrannosaurids are often very rough-textured. Alioramus, a possible tyrannosaurid from Mongolia, bears a single row of five prominent bony bumps on the nasal bones; a similar row of much lower bumps is present on the skull of Appalachiosaurus, as well as some specimens of Daspletosaurus, Albertosaurus, and Tarbosaurus. In Albertosaurus, Gorgosaurus and Daspletosaurus, there is a prominent horn in front of each eye on the lacrimal bone. The lacrimal horn is absent in Tarbosaurus and Tyrannosaurus, which instead have a crescent-shaped crest behind each eye on the postorbital bone. These head crests may have been used for display, perhaps for species recognition or courtship behavior. An example of the handicap principle may be the case of Guanlong, where the large, delicate crest may have been a hindrance to hunting in what was presumably an active predator. If an individual was healthy and successful at hunting despite the fragile crest, it would indicate the superior quality of the individual over others with smaller crests. Similarly to the unwieldy tail of a male peacock or the outsized antlers of an Irish elk, the crest of Guanlong may have evolved via sexual selection, providing an advantage in courtship that outweighed any decrease in hunting ability. Reproduction Neonate sized tyrannosaur fossils have been documented in the scientific literature.
Biology and health sciences
Theropods
Animals
3628996
https://en.wikipedia.org/wiki/Stethacanthus
Stethacanthus
Stethacanthus is an extinct genus of shark-like cartilaginous fish which lived from the Late Devonian to Late Carboniferous epoch, dying out around 298.9 million years ago. Fossils have been found in Australia, Asia, Europe and North America. Etymology Stethacanthus comes from the Greek στῆθος (stēthos), meaning "chest", and ἄκανθος (akanthos), meaning "spine" or "thorn". The name refers to the distinctive anvil-shaped first dorsal fin and spine displayed by mature males of the genus. Description Stethacanthus had different sizes depending on species, S. altonensis had length about , while S. productus reached . In many respects, it had a shark-like appearance. However, it is best known for its unusually shaped dorsal fin, which resembled an anvil or ironing board. Small spikes (enlarged versions of the dermal denticles commonly covering shark skin) covered this crest, and the ratfish's head as well. The crest may have played a role in mating rituals, aided in clamping to the belly of larger marine animals, or been used to frighten potential predators. Like other members of Stethacanthidae, Stethacanthus had unique pelvic girdles, single-crowned and non-growing scales, a pectoral fin composed of metapterygium with an accompanying ‘whip’ attached and a distinctive first dorsal fin and spine, termed the spine-brush complex. The neurocranium had a narrow suborbital shelf, a broad supraorbital shelf, a short otico-occipital division, large orbits, and cladodontic teeth that aligned precisely. In addition to these features, Stethacanthus also had male pelvic claspers with non-prismatic calcified cartilage at the distal ends. Spine-brush complex The spine-brush complex occupies the same site as the first dorsal fin on other ratfish and contains a basal plate extending inside a usually posterior-pointing dorsal spine composed of trabecular dentine. The spines resemble those of modern sharks and rays but curiously lack any enamel-like surface tissue. The trabecular dentine contains patches of fibers suggesting attachments to the epaxial musculature. The way these muscles would have been positioned implies that the spine could have been moved in anterio-posterior direction. The so-called "brush" is not fibrous as was originally believed, but consists of a number of parallel, membranous tubules made of globular calcified cartilage. The brush base and basal plate are covered in a thin, acellular bone layer. Zangerl asserts that these tubules are similar to erectile tissues in humans, and thus the complex may have been inflatable. The complex itself is covered in up to nine rows of large denticles pointing anteriorly. The dorsal side of the head has its own collection of denticles which point posteriorly. The presence of these large denticles has led to theories that the spine-brush complex in combination with the denticles on the head was used to scare away predators by simulating the mouth of a larger fish. The complex has been affirmed only in males, and only in those males that have reached sexual maturity. Whether the complex was present in females of the species is still unknown. Another theory for the spine-brush complex is that it was involved either in attracting a mate or in the mating process itself. Pectoral fin whip The pectoral fins of Stethacanthus were composed of the triangular-shaped metapterygium observed in modern-day sharks, but had an additional long, metapterygial structure called a fin whip. These fin whips contain at least 22 axial cartilages and extended past the pelvic fins. The three most anterior axials are shorter than the more posterior axials. The purpose of the fin whips is unknown but it has been suggested that they were used during mating. Teeth and denticles The tooth files are whorl-shaped and the palatoquadrate is scalloped with 6-7 recesses to allow for the tooth families. The individual teeth are widely separated from each other in the tooth whorls. The teeth themselves are of the cladodont variety; the bases of the teeth are broadest on the lingual side, and each support a single large cusp and two pairs of smaller accessory cusps for a total of five cusps. The medial and most lateral cusps are the most fluted. The teeth appear to be mostly orthodentine, but when viewed in cross-section, change abruptly to osteodentine. The enameloid is single-layered, overlaying the thick mantle of orthodentine. In addition to the dentition teeth, there are also a number of buccopharyngeal denticles lining the oropharynx. The denticles lining the top of the head and the top of the spine-brush complex are larger than the dentition teeth, and they appear as elongate monocuspid denticles. Pelvic girdles and claspers In Stethacanthus, the pelvic girdles consist of sheets of prismatic cartilage, each in the shape of a subtriangular, rounded plate. The anterior edge of each girdle is slightly concave while the posterior is convex. There appears to be no union of the two plates. There are two types of pelvic girdles found in stethacanthids: the primitive condition and the derived condition. In the primitive condition, the pelvic girdles have a metapterygial element supporting only one or two radials and most of the fin radials are attached directly to the pelvic plate. The derived condition differs in that there is a much higher number of radials supported by the pelvic plate. This feature, accompanied with a broadening of the pelvic girdle in order to accommodate the increased number of radials is a characteristic of Stethacanthus and other symmorriids. The males had claspers that were club-shaped at the distal ends and composed of non-prismatic globular calcified cartilage. Caudal fin There was some caudal fin variety among Stethacanthus species; while some had low angle heterocercal tails, some had tails approaching homocercal. The broad hypochordal lobe was supported by long, splayed fin radials. Paleobiology It is certain that Stethacanthus was a carnivore, and considering its small size probably fed on small fish, brachiopods, and crinoid ossicles like other sharks of its time. Additionally, as the spine-brush complex is rather a large structure, it seems likely that, in combination with the forward-facing denticles on the structure, it would have produced a drag force during fast locomotion. Therefore, Stethacanthus was probably a slow-moving shark. The fins of Stethacanthus were also smaller than in other sharks of the same size, and their teeth were also on the small side relative to other small Paleozoic sharks, suggesting that Stethacanthus may have been a bottom-dweller. Considering that most of the Stethacanthus specimens were recovered in the Bear Gulch Limestone in Montana, it is possible that this area was not only a breeding ground for other sharks but also for Stethacanthus, suggesting that they were migratory. History The several species of Stethacanthus discovered in the late 1800s were established based solely upon isolated spines, which initially confused paleontologist John Strong Newberry into thinking the spines constituted a new kind of fin. He originally believed that the spines were part of the pectoral fins and that they were not bilaterally symmetrical. Meanwhile, the first associated skeletal remains found in the Mississippian of Montana and the Devonian and Mississippian of Ohio remained undescribed for nearly a century. Since complete skeletons were extremely rare, Stethacanthus classification was vague and based on few characteristics. It was not until 1974 that the family Stethacanthidae was defined by Richard Lund because Stethacanthus differed so greatly from other elasmobranchs of the time. Relative classifications of symmoriids compared to stethacanthids are still debated. More Stethacanthus specimens have been discovered, expanding their range from the Midwestern United States to the Lower Carboniferous of Bearsden, Scotland and the Lower Tournaisian of the Tula Region of Central Russia and China. Stethacanthus teeth have been recovered from the Frasnian-Famennian Napier Formation and the Tournaisian Laurel Formation and Moogooree Limestone in Western Australia. A partial palate and jawbone referred to a Stethacanthus sp. has also been recovered from the Bonaparte Basin, Western Australia. Classification The presence of globular calcified cartilage in both the spine-brush complex base plate and brush and in the claspers is interesting because it is the first record of such a large mass of globular calcified cartilage in chondrichthyan. The high presence of globular calcified cartilage raises several questions about the evolution of sharks. It is possible that prismatic cartilage, a defining feature of chondrichthyans, is an evolutionary derivative of globular calcified cartilage. If this were the case, primitive chondrichthyans would have appeared with shark-like scales based instead on globular calcified cartilage. Another feature of note is the thin, acellular bone layer coating the brush and baseplate of the spine-brush complex. It is possible that the coating on the spine-brush complex is the first record of endoskeletal bone in primitive chondrichthyans, and that these endoskeletal features were lost in extant chondrichthyans. It is also possible that the fin spine could be a unique distribution of dermal skeleton and thus derived from neural crest. Following this assumption, the brush would be a fin-baseplate extension. The endoskeletal location and absence of fin radials supports the latter hypothesis. Taxonomic relationships are hard to define for Stethacanthus as there is much variability in the characteristics of the discovered specimens. Chondrichthyes is a monophyletic group characterized by the development of endoskeletal tesserae (mineralized blocks of cartilage) and internal fertilization. Chondrichthyes is further divided into two subclasses: Elasmobranchii and Holocephali. Stethacanthids have been classified as a member of the group Paleoselachii, which is a subdivision of Elasmobranchii. Stethacanthus has been further classified as part of the order Symmoriida, a classification that has sparked a controversy. There are two main hypotheses regarding this classification. One hypothesis states that the order Symmoriida consists of the families Symmoriidae, Stethacanthidae and Falcatidae and thus are a monophyletic group. Another is that symmoriids are actually the females of stethacanthids or are derived from stethacanthids. This hypothesis is due to the fact that stethacanthids and symmoriids are poorly defined; symmoriids are thought to lack a spine-brush complex but are otherwise identical to Stethacanthidae. Stethacanthids are identified by the presence of a spine-brush complex, which is in some cases non-existent (e.g. juvenile males), making the certain classification of stethacanthids and symmoriids difficult. More recently, Symmoriiformes as a whole has been reclassified as part of Holocephali, meaning that Stethacanthus may have been more closely related to modern chimaeras than to sharks.
Biology and health sciences
Prehistoric chondrichthyans
Animals
498255
https://en.wikipedia.org/wiki/Aldol%20condensation
Aldol condensation
An aldol condensation is a condensation reaction in organic chemistry in which two carbonyl moieties (of aldehydes or ketones) react to form a β-hydroxyaldehyde or β-hydroxyketone (an aldol reaction), and this is then followed by dehydration to give a conjugated enone. The overall reaction equation is as follows (where the Rs can be H) Aldol condensations are important in organic synthesis and biochemistry as ways to form carbon–carbon bonds. In its usual form, it involves the nucleophilic addition of a ketone enolate to an aldehyde to form a β-hydroxy ketone, or aldol (aldehyde + alcohol), a structural unit found in many naturally occurring molecules and pharmaceuticals. The term aldol condensation is also commonly used, especially in biochemistry, to refer to just the first (addition) stage of the process—the aldol reaction itself—as catalyzed by aldolases. However, the first step is formally an addition reaction rather than a condensation reaction because it does not involve the loss of a small molecule. Mechanism The first part of this reaction is an Aldol reaction, the second part a dehydration—an elimination reaction (Involves removal of a water molecule or an alcohol molecule). Dehydration may be accompanied by decarboxylation when an activated carboxyl group is present. The aldol addition product can be dehydrated via two mechanisms; a strong base like potassium t-butoxide, potassium hydroxide or sodium hydride deprotonates the product to an enolate, which eliminates via the E1cB mechanism, while dehydration in acid proceeds via an E1 reaction mechanism. Depending on the nature of the desired product, the aldol condensation may be carried out under two broad types of conditions: kinetic control or thermodynamic control. Both ketones and aldehydes are suitable for aldol condensation reactions. In the examples below, aldehydes are used. Base-catalyzed aldol condensation The mechanism for base-catalyzed aldol condensation can be seen in the image below. The process begins when a free hydroxide (strong base) strips the highly acidic proton at the alpha carbon of the aldehyde. This deprotonation causes the electrons from the C–H bond to shift and create a new C–C pi bond. The new pi bond then acts as a nucleophile and attacks the remaining aldehyde in the solution, resulting in the formation of a new C–C bond and regeneration of the base catalyst. In the second part of the reaction, the presence of base leads to elimination of water and formation of a new C–C pi bond. The product is referred to as the aldol condensation product. Acid-catalyzed aldol condensation The mechanism for acid-catalyzed aldol condensation can be seen in the image below. Crossed aldol condensation A crossed aldol condensation is a result of two dissimilar carbonyl compounds containing α-hydrogen(s) undergoing aldol condensation. Ordinarily, this leads to four possible products as either carbonyl compound can act as the nucleophile and self-condensation is possible, which makes a synthetically useless mixture. However, this problem can be avoided if one of the compounds does not contain an α-hydrogen, rendering it non-enolizable. In an aldol condensation between an aldehyde and a ketone, the ketone acts as the nucleophile, as its carbonyl carbon does not possess high electrophilic character due to the +I effect and steric hindrance. Usually, the crossed product is the major one. Any traces of the self-aldol product from the aldehyde may be disallowed by first preparing a mixture of a suitable base and the ketone and then adding the aldehyde slowly to the said reaction mixture. Using too concentrated base could lead to a competing Cannizzaro reaction. Examples The Aldox process, developed by Royal Dutch Shell and Exxon, converts propene and syngas to 2-ethylhexanol via hydroformylation to butyraldehyde, aldol condensation to 2-ethylhexanal and finally hydrogenation. Pentaerythritol is produced on a large scale beginning with crossed aldol condensation of acetaldehyde and three equivalents of formaldehyde to give pentaerythrose, which is further reduced in a Cannizzaro reaction. Scope Ethyl 2-methylacetoacetate and campholenic aldehyde react in an Aldol condensation. The synthetic procedure is typical for this type of reaction. In the process, in addition to water, an equivalent of ethanol and carbon dioxide are lost in decarboxylation. Ethyl glyoxylate 2 and glutaconate (diethyl-2-methylpent-2-enedioate) 1 react to isoprenetricarboxylic acid 3 (isoprene (2-methylbuta-1,3-diene) skeleton) with sodium ethoxide. This reaction product is very unstable with initial loss of carbon dioxide and followed by many secondary reactions. This is believed to be due to steric strain resulting from the methyl group and the carboxylic group in the cis-dienoid structure. Occasionally, an aldol condensation is buried in a multistep reaction or in catalytic cycle as in the following example: In this reaction an alkynal 1 is converted into a cycloalkene 7 with a ruthenium catalyst and the actual condensation takes place with intermediate 3 through 5. Support for the reaction mechanism is based on isotope labeling. The reaction between menthone ((2S,5R)-2-isopropyl-5-methylcyclohexanone) and anisaldehyde (4-methoxybenzaldehyde) is complicated due to steric shielding of the ketone group. This obstacle is overcome by using a strong base such as potassium hydroxide and a very polar solvent such as DMSO in the reaction below: The product can epimerize by way of a common intermediate—enolate A—to convert between the original (S,R) and the (R,R) epimers. The (R,R) product is insoluble in the reaction solvent whereas the (S,R) is soluble. The precipitation of the (R,R) product drives the epimerization equilibrium reaction to form this as the major product. Other condensation reactions There are other reactions of carbonyl compounds similar to aldol condensation: When the base is an amine and the active hydrogen compound is sufficiently activated the reaction is called a Knoevenagel condensation. In a Perkin reaction the aldehyde is aromatic and the enolate generated from an anhydride. Claisen-Schmidt condensation between an aldehyde or ketone having an α-hydrogen with an aromatic carbonyl compound lacking an α-hydrogen. A Claisen condensation involves two ester compounds. A Dieckmann condensation involves two ester groups in the same molecule and yields a cyclic molecule In the Japp–Maitland condensation water is removed not by an elimination reaction but by a nucleophilic displacement A Robinson annulation involves an α,β-unsaturated ketone and a carbonyl group, which first engage in a Michael reaction prior to the aldol condensation. In the Guerbet reaction, an aldehyde, formed in situ from an alcohol, self-condenses to the dimerized alcohol.
Physical sciences
Organic reactions
Chemistry
498427
https://en.wikipedia.org/wiki/Shooting%20method
Shooting method
In numerical analysis, the shooting method is a method for solving a boundary value problem by reducing it to an initial value problem. It involves finding solutions to the initial value problem for different initial conditions until one finds the solution that also satisfies the boundary conditions of the boundary value problem. In layman's terms, one "shoots" out trajectories in different directions from one boundary until one finds the trajectory that "hits" the other boundary condition. Mathematical description Suppose one wants to solve the boundary-value problemLet solve the initial-value problemIf , then is also a solution of the boundary-value problem. The shooting method is the process of solving the initial value problem for many different values of until one finds the solution that satisfies the desired boundary conditions. Typically, one does so numerically. The solution(s) correspond to root(s) of To systematically vary the shooting parameter and find the root, one can employ standard root-finding algorithms like the bisection method or Newton's method. Roots of and solutions to the boundary value problem are equivalent. If is a root of , then is a solution of the boundary value problem. Conversely, if the boundary value problem has a solution , it is also the unique solution of the initial value problem where , so is a root of . Etymology and intuition The term "shooting method" has its origin in artillery. An analogy for the shooting method is to place a cannon at the position , then vary the angle of the cannon, then fire the cannon until it hits the boundary value . Between each shot, the direction of the cannon is adjusted based on the previous shot, so every shot hits closer than the previous one. The trajectory that "hits" the desired boundary value is the solution to the boundary value problem — hence the name "shooting method". Linear shooting method The boundary value problem is linear if f has the form In this case, the solution to the boundary value problem is usually given by: where is the solution to the initial value problem: and is the solution to the initial value problem: See the proof for the precise condition under which this result holds. Examples Standard boundary value problem A boundary value problem is given as follows by Stoer and Bulirsch (Section 7.3.1). The initial value problem was solved for s = −1, −2, −3, ..., −100, and F(s) = w(1;s) − 1 plotted in the Figure 2. Inspecting the plot of F, we see that there are roots near −8 and −36. Some trajectories of w(t;s) are shown in the Figure 1. Stoer and Bulirsch state that there are two solutions, which can be found by algebraic methods. These correspond to the initial conditions w′(0) = −8 and w′(0) = −35.9 (approximately). Eigenvalue problem The shooting method can also be used to solve eigenvalue problems. Consider the time-independent Schrödinger equation for the quantum harmonic oscillator In quantum mechanics, one seeks normalizable wavefunctions and their corresponding energies subject to the boundary conditions The problem can be solved analytically to find the energies for , but also serves as an excellent illustration of the shooting method. To apply it, first note some general properties of the Schrödinger equation: If is an eigenfunction, so is for any nonzero constant . The -th excited state has roots where . For even , the -th excited state is symmetric and nonzero at the origin. For odd , the -th excited state is antisymmetric and thus zero at the origin. To find the -th excited state and its energy , the shooting method is then to: Guess some energy . Integrate the Schrödinger equation. For example, use the central finite difference If is even, set to some arbitrary number (say — the wavefunction can be normalized after integration anyway) and use the symmetric property to find all remaining . If is odd, set and to some arbitrary number (say — the wavefunction can be normalized after integration anyway) and find all remaining . Count the roots of and refine the guess for the energy . If there are or less roots, the guessed energy is too low, so increase it and repeat the process. If there are more than roots, the guessed energy is too high, so decrease it and repeat the process. The energy-guessing can be done with the bisection method, and the process can be terminated when the energy difference is sufficiently small. Then one can take any energy in the interval to be the correct energy.
Mathematics
Differential equations
null
498773
https://en.wikipedia.org/wiki/Benzalkonium%20chloride
Benzalkonium chloride
Benzalkonium chloride (BZK, BKC, BAK, BAC), also known as alkyldimethylbenzylammonium chloride (ADBAC) and by the trade name Zephiran, is a type of cationic surfactant. It is an organic salt classified as a quaternary ammonium compound. ADBACs have three main categories of use: as a biocide, a cationic surfactant, and a phase transfer agent. ADBACs are a mixture of alkylbenzyldimethylammonium chlorides, in which the alkyl group has various even-numbered alkyl chain lengths. Solubility and physical properties Depending on purity, benzalkonium chloride ranges from colourless to a pale yellow (impure). Benzalkonium chloride is readily soluble in ethanol and acetone. Dissolution in water is ready, upon agitation. Aqueous solutions should be neutral to slightly alkaline. Solutions foam when shaken. Concentrated solutions have a bitter taste and a faint almond-like odour. Standard concentrates are manufactured as 50% and 80% w/w solutions, and sold under trade names such as BC50, BC80, BAC50, BAC80, etc. The 50% solution is purely aqueous, while more concentrated solutions require incorporation of rheology modifiers (alcohols, polyethylene glycols, etc.) to prevent increases in viscosity or gel formation under low temperature conditions. Cationic surfactant Benzalkonium chloride possesses surfactant properties, dissolving the lipid phase of the tear film and increasing drug penetration, making it a useful excipient, but at the risk of causing damage to the surface of the eye. Laundry detergents and treatments. Softeners for textiles. Phase transfer agent Benzalkonium chloride is a mainstay of phase-transfer catalysis, an important technology in the synthesis of organic compounds, including drugs. Bioactive agents Especially for its antimicrobial activity, benzalkonium chloride is an active ingredient in many consumer products: Pharmaceutical products such as eye, ear and nasal drops or sprays, as a preservative. Personal care products such as hand sanitizers, wet wipes, shampoos, soaps, deodorants and cosmetics. Skin antiseptics and wound wash sprays, such as Bactine. Throat lozenges and mouthwashes, as a biocide Spermicidal creams. Cleaners for floor and hard surfaces as a disinfectant, such as Lysol and Dettol antibacterial spray and wipes. Algaecides for clearing of algae, moss, lichens from paths, roof tiles, swimming pools, masonry, etc. Benzalkonium chloride is also used in many non-consumer processes and products, including as an active ingredient in surgical disinfection. A comprehensive list of uses includes industrial applications. During the course of the COVID-19 pandemic, from time to time there have been shortages of hand cleaner containing ethanol or isopropanol as active ingredients. The FDA has stated that benzalkonium chloride is eligible as an alternative for use in the formulation of healthcare personnel hand rubs. However, in reference to the FDA rule, the CDC states that it does not have a recommended alternative to ethanol or isopropanol as active ingredients, and adds that "available evidence indicates benzalkonium chloride has less reliable activity against certain bacteria and viruses than either of the alcohols." In November 2020 the Journal of Hospital Infection published a study on benzalkonium chloride formulations; it was found that laboratory and commercial disinfectants with as little as 0.13% benzalkonium chloride inactivated the SARS-CoV-2 virus within 15 seconds of contact, even in the presence of a soil or hard water. This resulted in a growing consensus that BZK sanitizers are just as effective as alcohol-based sanitizers despite the CDC guidelines. As a hand sanitizer, use of BZK may be advantageous over ethanol in some situations because it has significantly more residual antibacterial action on the skin after initial application. Benzalkonium chloride has demonstrated persistent antimicrobial activity for up to four hours after contact whereas ethanol-based sanitizer demonstrate skin protection for only 10 minutes post-application. Medicine Benzalkonium chloride is a frequently used preservative in eye drops; typical concentrations range from 0.004% to 0.01%. Stronger concentrations can be caustic and cause irreversible damage to the corneal endothelium. Avoiding the use of benzalkonium chloride solutions while contact lenses are in place is discussed in the literature. Due to its antimicrobial activity when applied to skin, some topical medications for acne vulgaris have benzalkonium chloride added to increase the products' efficiency or shelf-life. Benzalkonium chloride has also been shown to be a spermicide. In Russia and China, benzalkonium chloride is used as a contraceptive. Tablets are inserted vaginally, or a gel is applied, resulting in local spermicidal contraception. It is not a wholly reliable method, and can cause irritation. Beekeeping This chemical is used in beekeeping for the treatment of rotten diseases of the brood. Adverse effects Although historically benzalkonium chloride has been ubiquitous as a preservative in ophthalmic preparations, its ocular toxicity and irritant properties, in conjunction with consumer demand, have led pharmaceutical companies to increase production of preservative-free preparations, or to replace benzalkonium chloride with preservatives which are less harmful. Many mass-marketed inhaler and nasal spray formulations contain benzalkonium chloride as a preservative, despite substantial evidence that it can adversely affect ciliary motion, mucociliary clearance, nasal mucosal histology, human neutrophil function, and leukocyte response to local inflammation. Although some studies have found no correlation between use of benzalkonium chloride in concentrations at or below 0.1% in nasal sprays and drug-induced rhinitis, others have recommended that benzalkonium chloride in nasal sprays be avoided. In the United States, nasal steroid preparations that are free of benzalkonium chloride include budesonide, triamcinolone acetonide, dexamethasone, and Beconase and Vancenase aerosol inhalers. Benzalkonium chloride is an irritant to middle ear tissues at typically used concentrations. Inner ear toxicity has been demonstrated. Occupational exposure to benzalkonium chloride has been linked to the development of asthma. In 2011, a large clinical trial designed to evaluate the efficacy of hand sanitizers based on different active ingredients in preventing virus transmission amongst schoolchildren was re-designed to exclude sanitizers based on benzalkonium chloride due to safety concerns. Benzalkonium chloride has been in common use as a pharmaceutical preservative and antimicrobial since the 1940s. While early studies confirmed the corrosive and irritant properties of benzalkonium chloride, investigations into the adverse effects of, and disease states linked to, benzalkonium chloride have only surfaced during the past 30 years. Toxicology RTECS lists the following acute toxicity data: Benzalkonium chloride is a human skin and severe eye irritant. It is a respiratory toxicant, immunotoxicant, gastrointestinal toxicant, and neurotoxicant. Benzalkonium chloride formulations for consumer use are dilute solutions. Concentrated solutions are toxic to humans, causing corrosion/irritation to the skin and mucosa, and death if taken internally in sufficient volumes. 0.1% is the maximum concentration of benzalkonium chloride that does not produce primary irritation on intact skin or act as a sensitizer. Poisoning by benzalkonium chloride is recognised in the literature. A 2014 case study detailing the fatal ingestion of up to 8.1 oz (240 ml) of 10% benzalkonium chloride in a 78-year-old male also includes a summary of the currently published case reports of benzalkonium chloride ingestion. While the majority of cases were caused by confusion about the contents of containers, one case cites incorrect pharmacy dilution of benzalkonium chloride as the cause of poisoning of two infants. In 2018 a Japanese nurse was arrested and admitted to having murdered approximately 20 patients at a hospital in Yokohama by injecting benzalkonium chloride into their intravenous drip bags. Benzalkonium chloride poisoning of domestic pets has been recognised as a result of direct contact with surfaces cleaned with disinfectants using benzalkonium chloride as an active ingredient. Biological activity The antimicrobial activity is dependent on the chain length. For example, yeast and fungi are most affected by C12, gram positive by C14, and gram negative by C16. The greatest biocidal activity is associated with the C12 dodecyl and C14 myristyl alkyl derivatives. The mechanism of bactericidal/microbicidal action is thought to be due to disruption of intermolecular interactions. This can cause dissociation of cellular membrane lipid bilayers, which compromises cellular permeability controls and induces leakage of cellular contents. Other biomolecular complexes within the bacterial cell can also undergo dissociation. Enzymes, which finely control a wide range of respiratory and metabolic cellular activities, are particularly susceptible to deactivation. Critical intermolecular interactions and tertiary structures in such highly specific biochemical systems can be readily disrupted by cationic surfactants. Benzalkonium chloride solutions are fast-acting biocidal agents with a moderately long duration of action. They are active against bacteria and some viruses, fungi, and protozoa. Bacterial spores are considered to be resistant. Solutions are bacteriostatic or bactericidal according to their concentration. Gram-positive bacteria are generally more susceptible than gram-negative bacteria. Its activity depends on the surfactant concentration and also on the bacterial concentration (inoculum) at the moment of the treatment. Activity is not greatly affected by pH, but increases substantially at higher temperatures and prolonged exposure times. In a 1998 study using the FDA protocol, a non-alcohol sanitizer with benzalkonium chloride as the active ingredient met the FDA performance standards, while Purell, a popular alcohol-based sanitizer, did not. The study, which was undertaken and reported by a leading US developer, manufacturer and marketer of topical antimicrobial pharmaceuticals based on quaternary ammonium compounds, found that their own benzalkonium chloride-based sanitizer performed better than alcohol-based hand sanitizer after repeated use. Newer formulations using benzalkonium blended with various quaternary ammonium derivatives can be used to extend the biocidal spectrum and enhance the efficacy of benzalkonium based disinfection products. Formulation techniques have been used to great effect in enhancing the virucidal activity of quaternary ammonium-based disinfectants such as Virucide 100 to typical healthcare infection hazards such as hepatitis and HIV. The use of appropriate excipients can also greatly enhance the spectrum, performance and detergency, and prevent deactivation under use conditions. Formulation can also help minimise deactivation of benzalkonium solutions in the presence of organic and inorganic contamination.. However, recent studies have demonstrated the capacity of environmental microorganisms to develop reduced susceptibility to benzalkonium chloride by employing strategies such as modifying bacterial membranes: increasing pump activity, and reducing the expression of certain porins. Degradation Benzalkonium chloride degradation follows consecutive debenzylation, dealkylation, and demethylation steps producing benzyl chloride, an alkyl dimethyl amine, dimethylamine, a long chain alkane, and ammonia. The intermediates, major, and minor products can then be broken down into CO2, H2O, NH3, and Cl–. The first step to the biodegradation of BAC is the fission or splitting of the alkyl chain from the quaternary nitrogen as shown in the diagram. This is done by abstracting the hydrogen from the alkyl chain by using a hydroxyl radical leading to a carbon centered radical. This results in dimethylbenzylamine as the first intermediate and dodecanal as the major product. From here, dimethylbenzylamine can be oxidized to benzoic acid using the Fenton process. The trimethyl amine group in dimethylbenzylamine can be cleaved to form a benzyl that can be further oxidized to benzoic acid. Benzoic acid uses hydroxylation (adding a hydroxyl group) to form p-hydroxybenzoic acid. Dimethylbenzylamine can then be converted into ammonia by performing demethylation twice, which removes both methyl groups, followed by debenzylation, removing the benzyl group using hydrogenation. The diagram represents suggested pathways of the biodegradation of BAC for both the hydrophobic and the hydrophilic regions of the surfactant. Since stearalkonium chloride is a type of BAC, the biodegradation process should happen in the same manner. Regulation Benzalkonium chloride is classed as a Category III antiseptic active ingredient by the United States Food and Drug Administration (FDA). Ingredients are categorized as Category III when "available data are insufficient to classify as safe and effective, and further testing is required”. In September 2016, the FDA announced a ban on nineteen ingredients in consumer antibacterial soaps citing a lack of evidence for safety and effectiveness. A ban on three additional ingredients, including benzalkonium chloride, was deferred at that time to allow ongoing studies to be completed. Benzalkonium chloride was deferred from further rulemaking in the 2019 FDA Final Rule on safety and effectiveness of consumer hand sanitizers, "to allow for the ongoing study and submission of additional safety and effectiveness data necessary to make a determination" on whether it met these criteria for use in OTC hand sanitizers, but the agency indicated it did not intend to take action to remove benzalkonium chloride-based hand sanitizers from the market. There is acknowledgement that more data are required on its safety, efficacy, and effectiveness, especially with relation to: Human pharmacokinetic studies, including information on its metabolites Studies on animal absorption, distribution, metabolism, and excretion Data to help define the effect of formulation on dermal absorption Carcinogenicity Studies on developmental and reproductive toxicology Potential hormonal effects Assessment of the potential for development of bacterial resistance Risks of using it as a contraceptive method. However, recent studies have demonstrated the capacity of environmental microorganisms to develop reduced susceptibility to benzalkonium chloride by employing strategies such as modifying bacterial membranes: increasing pump activity, and reducing the expression of certain porins.
Physical sciences
Organic salts
Chemistry
499294
https://en.wikipedia.org/wiki/Inflatable%20boat
Inflatable boat
An inflatable boat is a lightweight boat constructed with its sides and bow made of flexible tubes containing pressurised gas. For smaller boats, the floor and hull are often flexible, while for boats longer than , the floor typically consists of three to five rigid plywood or aluminium sheets fixed between the tubes, but not joined rigidly together. Often the transom is rigid, providing a location and structure for mounting an outboard motor. Some inflatable boats can be disassembled and packed into a small volume, so that they can be easily stored and transported. The boat, when inflated, is kept rigid cross-ways by a foldable removable thwart. This feature makes these boats suitable for liferafts for larger boats or aircraft, and for travel or recreational purposes. History Early attempts There are ancient carved images of animal skins filled with air being used as one-man floats to cross rivers. These floats were inflated by mouth. The discovery of the process to vulcanize rubber was made by Charles Goodyear in 1838, and was granted a US patent in 1844. Vulcanization stabilized the rubber, making it durable and flexible. In late 1843, Thomas Hancock filed for a UK patent, which was also granted in 1844, after the Goodyear Tire and Rubber Company patent had been granted. In 1852, while traveling in England, Charles Goodyear discovered that Thomas Hancock's company was producing vulcanized rubber and sued. Thomas Hancock had been shown a sample of Goodyear's rubber in 1842, but had not been told the process that made it—and Hancock said he had developed his process independently. The last of the suits were settled in 1855. Shortly thereafter, several people expanded on experimentation of rubber coated fabrics. In 1839 the Duke of Wellington tested the first inflatable pontoons. In 1840, the English scientist Thomas Hancock designed inflatable craft using his new methods of rubber vulcanization and described his achievements in The Origin and Progress of India Rubber Manufacture in England published a few years later. In 1844–1845, British naval officer Lieutenant Peter Halkett developed two types of inflatable boats intended for use by Arctic explorers. Both were made of rubber-impregnated "Mackintosh cloth." In the Halkett boat, the "boat cloak" served as a waterproof poncho or cloak until inflated, when it became a one-man boat. A special pocket held bellows for inflation, and a blade to turn a walking stick into a paddle. A special umbrella could double as a sail. Halkett later developed a two-man boat carried in a knapsack. When inflated, it could carry two men paddling on either side, and when deflated it served as a waterproof blanket for camping on wet ground. The Admiralty was sceptical about potential uses for Halkett's designs; on 8 May 1845, Lord Herbert, First Secretary to the Admiralty, wrote to Halkett that "My Lords are of an opinion that your invention is extremely clever and ingenious, and that it might be useful in Exploring and Surveying Expeditions, but they do not consider that it would be made applicable for general purposes in the Naval Service". The Admiralty saw no use for Halkett's designs in general naval service, but explorers liked this larger design. John Franklin bought one for the ill-fated 1845 expedition, in which the entire expedition party of 129 men and two ships vanished. In his explorations along the Oregon Trail, and the tributaries and forks of the Platte River in 1842 and 1843, John C. Frémont recorded what may have been the first use of an inflatable rubber boat for travel down rivers and rapids in the Rocky Mountains. In his account of the expedition he described his boat: Among the useful things which formed a portion of our equipage, was an India-rubber boat, 18 feet long, made somewhat in the form of a bark canoe of the northern lakes. The sides were formed by two airtight cylinders, eighteen inches in diameter, connected with others forming the bow and stern. To lessen the danger of accidents to the boat, these were divided into four different compartments, and the interior was sufficiently large to contain five or six persons, and a considerable weight of baggage. In 1848, General George Cullum, the US Army Corps of Engineers, introduced a rubber coated fabric inflatable bridge pontoon, which was used in the Mexican–American War and later on to a limited extent during the American Civil War. In 1866, four men crossed the Atlantic Ocean from New York to Britain on a three-tube raft called Nonpareil. From 1900 to 1910, the development of rubber manufacturing enabled attempts at producing circular rubber inflatable boats, similar to modern coracles. These were only usable as rafts, and could only be propelled by paddling. In addition, they tended to crack at seams and folds due to the imperfect manufacturing process of the rubber. Modern inflatable boat In the early 20th century, independent production of inflatable boats began with the airship manufacturing company RFD in England and the Zodiac company in France. This was brought about by the development of rubber-coated fabrics for the airship industry. Reginald Foster Dagnall, English designer and founder of RFD, switched in 1919 to the development of inflatable boats, using the coated fabric from hydrogen airships. The Air Ministry was impressed with trials of his boat on a lake near Guildford and began to give his firm contracts for the production of life-saving equipment. Meanwhile, in France a similar pattern emerged. The airship company Zodiac began to develop inflatable rubber boats, and in 1934, invented the inflatable kayak and catamaran. These led to the modern Zodiac inflatable boat. The company became Zodiac Nautic in 2015. Development continued after World War II with the discovery of new synthetic materials, such as neoprene and new adhesives, which allowed the boats to become sturdier and less prone to damage. World War II Submarine warfare in the Battle of the Atlantic led to casualties among warships and merchant ships. In the military, inflatable boats were used to transport torpedoes and other cargo. They also helped troops land in shallow water, and their compact size made overland transport possible. The US had two standard boats the LCRL and LCRS. The Marine Raiders were originally trained to carry out raids and landings from Landing Craft Rubber Large (LCRL) inflatable boats carried by high speed transports. In August 1942 the submarines and carried elements of the 2nd Raider Battalion who carried out the Makin Island raid from LCRL inflatable boats. Invasions of the Battle of Arawe by the 112th Cavalry Regiment and parts of the Battle of Tarawa involved amphibious landings in inflatable boats against heavy enemy resistance. One of the models, the Zodiac brand inflatable boat, became popular with the military, and contributed significantly to the rise of the civilian inflatable boat industry in Europe and in the United States. After World War II, governments sold surplus inflatable boats to the public. Post-war inflatables Inflatable liferafts were also used successfully to save crews of aircraft that ditched in the sea; bombing, naval and anti-submarine aircraft flying long distances over water being much more common from the start of WWII. In the 1950s, the French Navy officer and biologist Alain Bombard was the first to combine the outboard engine, a rigid floor and a boat shaped inflatable. The former airplane-manufacturer Zodiac built that boat and a friend of Bombard, the diver Jacques-Yves Cousteau began to use it, after Bombard sailed across the Atlantic Ocean with his inflatable in 1952. Cousteau was convinced by the shallow draught and good performance of this type of boat and used it as tenders on his expeditions. The inflatable boat was so successful that Zodiac lacked the manufacturing capacity to satisfy demand. In the early 1960s, Zodiac licensed production to a dozen companies in other countries. In the 1960s, the British company Humber was the first to build Zodiac brand inflatable boats in the United Kingdom. Some inflatables have inflated keels whose V-shape help the hull move through waves reducing the slamming effect caused by the flat hull landing back on the water surface after passing over the top of a wave at speed. Types Contemporary inflatable boats are manufactured using supported fabric. They are made of rubberized synthetic fabrics, PVC and polyurethane, providing light-weight and airtight sponsons. Depending on fabric choice, the fabric panels are assembled using either hot or cold manufacturing processes. Different styles of one-way valves are used to add or remove air, and some brands include inter-communicating valves that reduce the effect of a puncture. Inflatable boats with transoms have an inflatable keel that creates a slight V-bottom along the line of the hull to improve the hull's seakeeping and directional stability. These vessels are very light, so if powered with an engine, it is best to put weight in the bow area to keep the bow from rising while the boat is going up on plane. People increasingly use inflatables for personal recreational use on lakes, rivers, and oceans—and for white water rafting and kayaking, and for scuba divers to reach dive sites. Users can deflate, fold, and store fabric bottom inflatable boats in compact bags, making them ideal for limited storage and quick, easy access. Sail rigs are available for inflatable dinghies, kayaks, and catamarans. In keeping with the portability of the inflatable hull, sail attachments fold or disassemble to fit in a compact bundle. Leeboards on the sides perform the same function as a centerboard, so users can tack these boats into the wind. Rigid inflatable boat The modern rigid inflatable boat (RIB) is a development of the inflatable boat, which has a rigid floor and solid hull. The external shape of the hull lets it cut through waves more easily giving a more comfortable ride when traveling fast in rough conditions. The structure of the hull is capable of supporting a more powerful transom mounted outboard engine or even an inboard engine. Soft inflatable boat A soft inflatable boat (SIB) lacks the solid hull of a RIB and often has a removable slatted floor, so the boat can be deflated and transported in a car or other vehicle. Such boats have a low draft and are therefore useful for traveling across shallow water and beaching in places without landing facilities. Some SIBs have a rigid transom that can support an outboard engine. Inflatable boats with transoms have an inflatable keel that creates a slight V-bottom along the line of the hull to improve the hull's seakeeping and directional stability. These vessels are very light, so if powered with an engine, it is best to put weight in the bow area to keep the bow from rising while the boat is going up on plane. Soft inflatable boats are available with several floor choices: Roll up slat floor Hard floor made of fiberglass, aluminum or wood panels Ribbed air floor (on inflatable rafts) High pressure air floor Uses Inflatables are commonly between long and are propelled by outboard motors of . Due to their speed, portability, and weight, inflatable boats are used in diverse roles: Inflatable and rigid-hulled inflatable boats are often used for short scuba diving excursions. The International Convention for the Safety of Life at Sea publishes recommended regulations for inflatable boats used in rescue operations. Some life rafts also contain additional inflatable sections to ensure that the raft self-rights in heavy seas. Inflatable life rafts have also been used since the 1930s on military aircraft that operate over water. These boats are often used by special-operations units of the armed forces of several nations, for such purposes as landing on beaches. Because inflatable craft can be stored compactly they can also be transported on midget submarines such as those operated by the Advanced SEAL Delivery System. They have also been used by other forces without government sponsorship, such as guerrillas and pirates. Lifeguards use inflatable boats or jet skis to reduce the time to reach a swimmer in distress. Inflatables are also used in conjunction with larger rescue craft, such as the Y class lifeboat used with the Tamar and Severn class lifeboats. They are used in a number of sporting events and for recreational purposes, such as whitewater rafting, inflatable rescue boat racing, water skiing and fishing. Transportation An inflatable boat can be transported in various ways: Deflated and Packed: The most significant advantage of inflatable boats is their ability to be deflated, folded, and packed into a compact size. Once deflated, they can be stored in a carry bag and transported in the trunk of a car or the bed of a truck. On a Trailer: If the inflatable boat is larger or if it has a hard bottom (like a RIB), it might be more convenient to transport it on a trailer. This is especially useful if the boat has an outboard motor attached. On Roof Racks: Some individuals use roof racks on their vehicles to transport inflatable boats. The boat is either deflated and packed or partially inflated and secured on top of the car. Using a Boat Dolly: For short distances, like moving the boat from a parking area to the water’s edge, a boat dolly or hand cart can be useful. These are especially handy for heavier inflatables or those with motors.
Technology
Naval transport
null
499429
https://en.wikipedia.org/wiki/D%27Alembert%27s%20principle
D'Alembert's principle
D'Alembert's principle, also known as the Lagrange–d'Alembert principle, is a statement of the fundamental classical laws of motion. It is named after its discoverer, the French physicist and mathematician Jean le Rond d'Alembert, and Italian-French mathematician Joseph Louis Lagrange. D'Alembert's principle generalizes the principle of virtual work from static to dynamical systems by introducing forces of inertia which, when added to the applied forces in a system, result in dynamic equilibrium. D'Alembert's principle can be applied in cases of kinematic constraints that depend on velocities. The principle does not apply for irreversible displacements, such as sliding friction, and more general specification of the irreversibility is required. Statement of the principle The principle states that the sum of the differences between the forces acting on a system of massive particles and the time derivatives of the momenta of the system itself projected onto any virtual displacement consistent with the constraints of the system is zero. Thus, in mathematical notation, d'Alembert's principle is written as follows, where: is an integer used to indicate (via subscript) a variable corresponding to a particular particle in the system, is the total applied force (excluding constraint forces) on the -th particle, is the mass of the -th particle, is the velocity of the -th particle, is the virtual displacement of the -th particle, consistent with the constraints. Newton's dot notation is used to represent the derivative with respect to time. The above equation is often called d'Alembert's principle, but it was first written in this variational form by Joseph Louis Lagrange. D'Alembert's contribution was to demonstrate that in the totality of a dynamic system the forces of constraint vanish. That is to say that the generalized forces need not include constraint forces. It is equivalent to the somewhat more cumbersome Gauss's principle of least constraint. Derivations General case with variable mass The general statement of d'Alembert's principle mentions "the time derivatives of the momenta of the system." By Newton's second law, the first time derivative of momentum is the force. The momentum of the -th mass is the product of its mass and velocity: and its time derivative is In many applications, the masses are constant and this equation reduces to However, some applications involve changing masses (for example, chains being rolled up or being unrolled) and in those cases both terms and have to remain present, giving Special case with constant mass Consider Newton's law for a system of particles of constant mass, . The total force on each particle is where are the total forces acting on the system's particles, are the inertial forces that result from the total forces. Moving the inertial forces to the left gives an expression that can be considered to represent quasi-static equilibrium, but which is really just a small algebraic manipulation of Newton's law: Considering the virtual work, , done by the total and inertial forces together through an arbitrary virtual displacement, , of the system leads to a zero identity, since the forces involved sum to zero for each particle. The original vector equation could be recovered by recognizing that the work expression must hold for arbitrary displacements. Separating the total forces into applied forces, , and constraint forces, , yields If arbitrary virtual displacements are assumed to be in directions that are orthogonal to the constraint forces (which is not usually the case, so this derivation works only for special cases), the constraint forces don't do any work, . Such displacements are said to be consistent with the constraints. This leads to the formulation of d'Alembert's principle, which states that the difference of applied forces and inertial forces for a dynamic system does no virtual work: There is also a corresponding principle for static systems called the principle of virtual work for applied forces. D'Alembert's principle of inertial forces D'Alembert showed that one can transform an accelerating rigid body into an equivalent static system by adding the so-called "inertial force" and "inertial torque" or moment. The inertial force must act through the center of mass and the inertial torque can act anywhere. The system can then be analyzed exactly as a static system subjected to this "inertial force and moment" and the external forces. The advantage is that in the equivalent static system one can take moments about any point (not just the center of mass). This often leads to simpler calculations because any force (in turn) can be eliminated from the moment equations by choosing the appropriate point about which to apply the moment equation (sum of moments = zero). Even in the course of Fundamentals of Dynamics and Kinematics of machines, this principle helps in analyzing the forces that act on a link of a mechanism when it is in motion. In textbooks of engineering dynamics, this is sometimes referred to as d'Alembert's principle. Some educators caution that attempts to use d'Alembert inertial mechanics lead students to make frequent sign errors. A potential cause for these errors is the sign of the inertial forces. Inertial forces can be used to describe an apparent force in a non-inertial reference frame that has an acceleration with respect to an inertial reference frame. In such a non-inertial reference frame, a mass that is at rest and has zero acceleration in an inertial reference system, because no forces are acting on it, will still have an acceleration and an apparent inertial, or pseudo or fictitious force will seem to act on it: in this situation the inertial force has a minus sign. Dynamic equilibrium D'Alembert's form of the principle of virtual work states that a system of rigid bodies is in dynamic equilibrium when the virtual work of the sum of the applied forces and the inertial forces is zero for any virtual displacement of the system. Thus, dynamic equilibrium of a system of rigid bodies with generalized coordinates requires for any set of virtual displacements with being a generalized applied force and being a generalized inertia force. This condition yields equations, which can also be written as The result is a set of m equations of motion that define the dynamics of the rigid body system. Formulation using the Lagrangian D'Alembert's principle can be rewritten in terms of the Lagrangian of the system as a generalized version of Hamilton's principle for the case of point particles, as follows, where: are the applied forces is the virtual displacement of the -th particle, consistent with the constraints the critical curve satisfies the constraints With the Lagrangian the previous statement of d'Alembert principle is recovered. Generalization for thermodynamics An extension of d'Alembert's principle can be used in thermodynamics. For instance, for an adiabatically closed thermodynamic system described by a Lagrangian depending on a single entropy S and with constant masses , such as it is written as follows where the previous constraints and are generalized to involve the entropy as: Here is the temperature of the system, are the external forces, are the internal dissipative forces. It results in the mechanical and thermal balance equations: Typical applications of the principle include thermo-mechanical systems, membrane transport, and chemical reactions. For the classical d'Alembert principle and equations are recovered.
Physical sciences
Classical mechanics
Physics
499532
https://en.wikipedia.org/wiki/Zinc%20sulfate
Zinc sulfate
Zinc sulfate is an inorganic compound with the formula ZnSO4. It forms hydrates ZnSO4·nH2O, where n can range from 0 to 7. All are colorless solids. The most common form includes water of crystallization as the heptahydrate, with the formula . As early as the 16th century it was prepared on a large scale, and was historically known as "white vitriol" (the name was used, for example, in 1620s by the collective writing under the pseudonym of Basil Valentine). Zinc sulfate and its hydrates are colourless solids. Uses Manufacturing The main application of the heptahydrate is as a coagulant in the production of rayon. It is also a precursor to the pigment lithopone. It is also used as an electrolyte for zinc electroplating, as a mordant in dyeing, and as a preservative for skins and leather. Nutrition Zinc sulfate is used to supply zinc in animal feeds, fertilizers, toothpaste, and agricultural sprays. Zinc sulfate, like many zinc compounds, can be used to control moss growth on roofs. Zinc sulfate can be used to supplement zinc in the brewing process. Zinc is a necessary nutrient for optimal yeast health and performance, although it is not a necessary supplement for low-gravity beers, as the grains commonly used in brewing already provide adequate zinc. It is a more common practice when pushing yeast to their limit by increasing alcohol content beyond their comfort zone. Before modern stainless steel, brew Kettles, fermenting vessels and after wood, zinc was slowly leached by the use of copper kettles. A modern copper immersion chiller is speculated to provide trace amounts of zinc; thus care must be taken when adding supplemental zinc so as not to cause excess. Side effects include "...increased acetaldehyde and fusel alcohol production due to high yeast growth when zinc concentrations exceed 5 ppm. Excess zinc can also cause soapy or goaty flavors." Zinc sulfate is a potent inhibitor of sweetness perception for most sweet-tasting substances. Medicine It is used as a dietary supplement to treat zinc deficiency and to prevent the condition in those at high risk. Side effects of excess supplementation may include abdominal pain, vomiting, headache, and tiredness. it is also used together with oral rehydration therapy (ORT) and an astringent. Production, reactions, structure Zinc sulfate is produced by treating virtually any zinc-containing material (metal, minerals, oxides) with sulfuric acid. Specific reactions include the reaction of the metal with aqueous sulfuric acid: Pharmaceutical-grade zinc sulfate is produced by treating high-purity zinc oxide with sulfuric acid: In aqueous solution, all forms of zinc sulfate behave identically. These aqueous solutions consist of the metal aquo complex and ions. Barium sulfate forms when these solutions are treated with solutions of barium ions: With a reduction potential of −0.76 V, zinc(II) reduces only with difficulty. When heated above 680 °C, zinc sulfate decomposes into sulfur dioxide gas and zinc oxide fume, both of which are hazardous. The heptahydrate is isostructural with ferrous sulfate heptahydrate. The solid consists of [Zn(H2O)6]2+ ions interacting with sulfate and one water of crystallization by hydrogen bonds. Anhydrous zinc sulfate is isomorphous with anhydrous copper(II) sulfate. It exists as the mineral zincosite. A monohydrate is known. The hexahydrate is also recognized. Minerals As a mineral, ZnSO4•7H2O is known as goslarite. Zinc sulfate occurs as several other minor minerals, such as zincmelanterite, (structurally different from goslarite). Lower hydrates of zinc sulfate are rarely found in nature: (bianchite), (boyleite), and (gunningite). Safety Zinc sulfate powder is an eye irritant. Ingestion of trace amounts is considered safe, and zinc sulfate is added to animal feed as a source of essential zinc, at rates of up to several hundred milligrams per kilogram of feed. Excess ingestion results in acute stomach distress, with nausea and vomiting appearing at 2–8 mg/kg of body weight. Nasal irrigation with a solution of zinc sulfate has been found to be able to damage the olfactory sense nerves and induce anosmia in a number of different species, including humans.
Physical sciences
Sulfuric oxyanions
Chemistry
499660
https://en.wikipedia.org/wiki/Adhesive%20bandage
Adhesive bandage
An adhesive bandage, also called a sticking plaster, sticky plaster, medical plaster, or simply plaster in British English, is a small medical dressing used for injuries not serious enough to require a full-size bandage. They are also known by the genericized trademarks of Band-Aid (as "band-aid" or "band aid" in Australia, Canada, India and the US) or Elastoplast (in the UK). Function The adhesive bandage protects the wound and scab from friction, bacteria, damage, and dirt. Thus, the healing process of the body is less disturbed. Some of the dressings have antiseptic properties. An additional function is to hold the two cut edges of the skin together to make the healing process faster. Design An adhesive bandage is a small, flexible sheet of material which is sticky on one side, with a smaller, non-sticky, absorbent pad stuck to the sticky side. The pad is placed against the wound, and overlapping edges of the sticky material are smoothed down so they stick to the surrounding skin. Adhesive bandages are generally packaged in a sealed, sterile bag, with a backing covering the sticky side; the backing is removed as the bandage is applied. They come in a variety of sizes and shapes. Materials The backing and bag are often made of coated paper, but may be made of plastic. The adhesive sheet is usually a woven fabric, plastic (PVC, polyethylene or polyurethane), or latex strip. It may or may not be waterproof; if it is airtight, the bandage is an occlusive dressing. The adhesive is commonly an acrylate, including methacrylates and epoxy diacrylates (which are also known as vinyl resins). Some people have allergies to some of these materials, particularly latex and some adhesives. Colors Due to being widely available only in a standard color, some people with skin tones darker than the standard bandage color have expressed frustration at having to use bandages that looked less conspicuous on the skin of lighter-skinned people. This has led to greater support for pharmaceutical companies that manufacture these bandages in a variety of skin tones. Some bandages, especially those designed for children, may come in a wide variety of colors or may feature cartoon characters. Special bandages are used by food preparation workers. These are waterproof, have strong adhesive so they are less likely to fall off, and are usually blue so that they are more clearly visible in food. Some include a metal strip detectable by machines used in food manufacturing to ensure that food is free from foreign objects. Variants Transdermal patches are adhesive bandages with the function to distribute medication through the skin, rather than protecting a wound. Butterfly closures, also known as butterfly stitches, are generally thin adhesive strips which can be used to close small wounds. They are applied perpendicular to the laceration in a manner which pulls the skin on either side of the wound together. They are not true sutures, but can often be used in addition to, or in place of actual sutures for small wounds. Butterfly stitches can be advantageous in that they do not need a medical professional to be placed or removed, and are thus a common item in first aid kits. Notable brands Band-Aid Curad Elastoplast Nexcare
Technology
Equipment
null
500338
https://en.wikipedia.org/wiki/Outrigger
Outrigger
An outrigger is a projecting structure on a boat, with specific meaning depending on types of vessel. Outriggers may also refer to legs on a wheeled vehicle that are folded out when it needs stabilization, for example on a crane that lifts heavy loads. Powered vessels and sailboats An outrigger describes any contraposing float rigging beyond the side (gunwale) of a boat to improve the vessel's stability. If a single outrigger is used it is usually but not always windward. The technology was originally developed by the Austronesian people. There are two main types of boats with outriggers: double outriggers (prevalent in maritime Southeast Asia) and single outriggers (prevalent in Madagascar, Melanesia, Micronesia and Polynesia). Multihull ships are also derived from outrigger boats. In an outrigger canoe and in sailboats such as the proa, an outrigger is a thin, long, solid, hull used to stabilise an inherently unstable main hull. The outrigger is positioned rigidly and parallel to the main hull so that the main hull is less likely to capsize. If only one outrigger is used on a vessel, its weight reduces the tendency to capsize in one direction and its buoyancy reduces the tendency in the other direction. On a keelboat, "outrigger" refers to a variety of structures by which the running rigging (such as a sheet) may be attached outboard (outside the lateral limits) of the boat's hull. The Racing Rules of Sailing generally prohibit such outriggers, though they are explicitly permitted on specific classes, such as the IMOCA Open 60 used in several major offshore races. Fishing In fishing from vessels, an outrigger is a pole or series of poles that allow boats to trawl more lines in the water without tangling and simulates a school of fish. Rowing In a rowing boat or galley, an outrigger (or rigger) is a triangular frame that holds the rowlock (into which the oar is slotted) away from the saxboard (or gunwale in gig rowing) to optimize leverage. Wooden outriggers appear on the new trireme around the 7th or 6th centuries BC and later on Italian galleys around AD 1300, while Harry Clasper (1812–1870), a British professional rower, popularised the use of the modern tubular-metal version and the top rowing events accepted the physiological and ergonomic advantages so acceded to its use in competitions. In recent decades, some manufacturers of racing shells have developed wing-riggers which are reinforced arcs or flattened tubular projections akin to aircraft wings, instead of conventional triangular structures.
Technology
Naval transport
null
500475
https://en.wikipedia.org/wiki/Hypotension
Hypotension
Hypotension, also known as low blood pressure, is a cardiovascular condition characterized by abnormally reduced blood pressure. Blood pressure is the force of blood pushing against the walls of the arteries as the heart pumps out blood and is indicated by two numbers, the systolic blood pressure (the top number) and the diastolic blood pressure (the bottom number), which are the maximum and minimum blood pressures within the cardiac cycle, respectively. A systolic blood pressure of less than 90 millimeters of mercury (mmHg) or diastolic of less than 60 mmHg is generally considered to be hypotension. Different numbers apply to children. However, in practice, blood pressure is considered too low only if noticeable symptoms are present. Symptoms may include dizziness, lightheadedness, confusion, feeling tired, weakness, headache, blurred vision, nausea, neck or back pain, an irregular heartbeat or feeling that the heart is skipping beats or fluttering, sweating, and fainting. Hypotension is the opposite of hypertension, which is high blood pressure. It is best understood as a physiological state rather than a disease. Severely low blood pressure can deprive the brain and other vital organs of oxygen and nutrients, leading to a life-threatening condition called shock. Shock is classified based on the underlying cause, including hypovolemic shock, cardiogenic shock, distributive shock, and obstructive shock. Hypotension can be caused by strenuous exercise, excessive heat, low blood volume (hypovolemia), hormonal changes, widening of blood vessels, anemia, vitamin B12 deficiency, anaphylaxis, heart problems, or endocrine problems. Some medications can also lead to hypotension. There are also syndromes that can cause hypotension in patients including orthostatic hypotension, vasovagal syncope, and other rarer conditions. For many people, excessively low blood pressure can cause dizziness and fainting or indicate serious heart, endocrine or neurological disorders. For some people who exercise and are in top physical condition, low blood pressure could be normal. A single session of exercise can induce hypotension and water-based exercise can induce a hypotensive response. Treatment depends on what causes low blood pressure. Treatment of hypotension may include the use of intravenous fluids or vasopressors. When using vasopressors, trying to achieve a mean arterial pressure (MAP) of greater than 70 mmHg does not appear to result in better outcomes than trying to achieve an MAP of greater than 65 mmHg in adults. Signs and symptoms For many people, low blood pressure goes unnoticed. For some people, low blood pressure may be a sign of an underlying health condition, especially when it drops suddenly or occurs with symptoms. Older adults also have a higher risk of symptoms of low blood pressure, such as falls, fainting, or dizziness when standing or after a meal. If the blood pressure is sufficiently low, fainting (syncope) may occur. Low blood pressure is sometimes associated with certain symptoms, many of which are related to causes rather than effects of hypotension: confusion dizziness or lightheadedness feeling tired or weak shortness of breath irregular heartbeat, feeling that the heart is skipping beats, or fluttering chest pain fever headache stiff neck severe back or neck pain cough with sputum prolonged diarrhea or vomiting chills loss of appetite nausea dyspepsia (indigestion) dysuria (painful urination) acute, life-threatening allergic reaction seizures loss of consciousness temporary blurring or loss of vision black tarry stools Causes Low blood pressure can be caused by low blood volume, hormonal changes, pregnancy, widening of blood vessels, medicine side effects, severe dehydration, anemia, vitamin B12 deficiency, anaphylaxis, heart problems or endocrine problems. Reduced blood volume, hypovolemia, is the most common cause of hypotension. This can result from hemorrhage; insufficient fluid intake, as in starvation; or excessive fluid losses from diarrhea or vomiting. Hypovolemia can be induced by excessive use of diuretics. Low blood pressure may also be attributed to heat stroke which can be indicated by absence of perspiration, light headedness and dark colored urine. Other medications can produce hypotension by different mechanisms. Chronic use of alpha blockers or beta blockers can lead to hypotension. Beta blockers can cause hypotension both by slowing the heart rate and by decreasing the pumping ability of the heart muscle. Decreased cardiac output despite normal blood volume, due to severe congestive heart failure, large myocardial infarction, heart valve problems, or extremely low heart rate (bradycardia), often produces hypotension and can rapidly progress to cardiogenic shock. Arrhythmias often result in hypotension by this mechanism. Excessive vasodilation, or insufficient constriction of the blood vessels (mostly arterioles), causes hypotension. This can be due to decreased sympathetic nervous system output or to increased parasympathetic activity occurring as a consequence of injury to the brain or spinal cord. Dysautonomia, an intrinsic abnormality in autonomic system functioning, can also lead to hypotension. Excessive vasodilation can also result from sepsis, acidosis, or medications, such as nitrate preparations, calcium channel blockers, or AT1 receptor antagonists (Angiotensin II acts on AT1 receptors). Many anesthetic agents and techniques, including spinal anesthesia and most inhalational agents, produce significant vasodilation. Lower blood pressure is a side effect of certain herbal medicines, which can also interact with several medications. An example is the theobromine in Theobroma cacao, which lowers blood pressure through its actions as both a vasodilator and a diuretic, and has been used to treat high blood pressure. Syndromes Orthostatic hypotension Orthostatic hypotension, also called postural hypotension, is a common form of low blood pressure. It occurs after a change in body position, typically when a person stands up from either a seated or lying position. It is usually transient and represents a delay in the normal compensatory ability of the autonomic nervous system. It is commonly seen in hypovolemia and as a result of various medications. In addition to blood pressure-lowering medications, many psychiatric medications, in particular antidepressants, can have this side effect. Simple blood pressure and heart rate measurements while lying, seated, and standing (with a two-minute delay in between each position change) can confirm the presence of orthostatic hypotension. Taking these measurements is known as orthostatic vitals. Orthostatic hypotension is indicated if there is a drop of 20 mmHg in systolic pressure (and a 10 mmHg drop in diastolic pressure in some facilities) and a 20 beats per minute increase in heart rate. Vasovagal syncope Vasovagal syncope is a form of dysautonomia characterized by an inappropriate drop in blood pressure while in the upright position. Vasovagal syncope occurs as a result of increased activity of the vagus nerve, the mainstay of the parasympathetic nervous system. Patients will feel sudden, unprovoked lightheadedness, sweating, changes in vision, and finally a loss of consciousness. Consciousness will often return rapidly once patient is lying down and the blood pressure returns to normal. Other Another, but rarer form, is postprandial hypotension, a drastic decline in blood pressure that occurs 30 to 75 minutes after eating substantial meals. When a great deal of blood is diverted to the intestines (a kind of "splanchnic blood pooling") to facilitate digestion and absorption, the body must increase cardiac output and peripheral vasoconstriction to maintain enough blood pressure to perfuse vital organs, such as the brain. Postprandial hypotension is believed to be caused by the autonomic nervous system not compensating appropriately, because of aging or a specific disorder. Hypotension is a feature of Flammer syndrome, which is characterized by cold hands and feet and predisposes to normal tension glaucoma. Hypotension can be a symptom of relative energy deficiency in sport, sometimes known as the female athlete triad, although it can also affect men. Pathophysiology Blood pressure is continuously regulated by the autonomic nervous system, using an elaborate network of receptors, nerves, and hormones to balance the effects of the sympathetic nervous system, which tends to raise blood pressure, and the parasympathetic nervous system, which lowers it. The vast and rapid compensation abilities of the autonomic nervous system allow normal individuals to maintain an acceptable blood pressure over a wide range of activities and in many disease states. Even small alterations in these networks can lead to hypotension. Diagnosis For most adults, the optimal blood pressure is at or below 120/80 mmHg. If the systolic blood pressure is <90 mmHg or the diastolic blood pressure is <60 mmHg, it would be classified as hypotension. However, occasional blood pressure readings below 90/60 mmHg are not infrequent in the general population, and, in the absence of some pathological cause, hypotension appears to be a relatively benign condition in most people. The diagnosis of hypotension is usually made by measuring blood pressure, either non-invasively with a sphygmomanometer or invasively with an arterial catheter (mostly in an intensive care setting). Another way to diagnose low blood pressure is by using the mean arterial pressure (MAP) measured using an arterial catheter or by continuous, non-invasive hemodynamic monitoring which measures intra-operative blood pressure beat-by-beat throughout surgery. A MAP <65 mmHg is considered hypotension. Intra-operative hypotension <65 mmHg can lead to an increased risk of acute kidney injury, myocardial injury or post-operative stroke. While an incidental finding of hypotension during a routine blood pressure measurement may not be particularly worrying, a substantial drop in blood pressure following standing, exercise or eating can be associated with symptoms and may have implications for future health. A drop in blood pressure after standing, termed postural or orthostatic hypotension, is defined as a decrease in supine-to-standing BP >20 mm Hg systolic or >10 mm Hg diastolic within 3 minutes of standing. Orthostatic hypotension is associated with increased risk of future cardiovascular events and mortality. Orthostatic vitals are frequently measured to assist with the diagnosis of orthostatic hypotension, and may involve the use of a tilt table test to evaluate vasovagal syncope. Treatment Treatment depends on what causes low blood pressure. Treatment may not be needed for asymptomatic low blood pressure. Depending on symptoms, treatment may include drinking more fluids to prevent dehydration, taking medicines to raise blood pressure, or adjusting medicines that cause low blood pressure. Adding electrolytes to a diet can relieve symptoms of mild hypotension, and a morning dose of caffeine can also be effective. Chronic hypotension rarely exists as more than a symptom. In mild cases, where the patient is still responsive, laying the person on their back and lifting the legs increases venous return, thus making more blood available to critical organs in the chest and head. The Trendelenburg position, though used historically, is no longer recommended. Hypotensive shock treatment always follows the first four following steps. Outcomes, in terms of mortality, are directly linked to the speed that hypotension is corrected. Still-debated methods are in parentheses, as are benchmarks for evaluating progress in correcting hypotension. A study on septic shock provided the delineation of these general principles. However, since it focuses on hypotension due to infection, it is not applicable to all forms of severe hypotension. Volume resuscitation (usually with crystalloid or blood products) Blood pressure support with a vasopressor (all seem equivalent with respect to risk of death, with norepinephrine possibly better than dopamine). Trying to achieve a mean arterial pressure (MAP) of greater than 70 mmHg does not appear to result in better outcomes than trying to achieve a MAP of greater than 65 mmHg in adults. Ensure adequate tissue perfusion (maintain SvO2 >70 with use of blood or dobutamine) Address the underlying problem (i.e., antibiotic for infection, stent or CABG (coronary artery bypass graft surgery) for infarction, steroids for adrenal insufficiency, etc...) The best way to determine if a person will benefit from fluids is by doing a passive leg raise followed by measuring the output from the heart. Medication Chronic hypotension sometimes requires the use of medications. Some medications that are commonly used include Fludrocortisone, Erythropoietin, and Sympathomimetics such as Midodrine and Noradrenaline and precursor (L-DOPS). Fludrocortisone is the first-line therapy (in the absence of heart failure) for patients with chronic hypotension or resistant orthostatic hypotension. It works by increasing the intravascular volume. Midodrine is a therapy used for severe orthostatic hypotension, and works by increasing peripheral vascular resistance. Noradrenaline and its precursor L-DOPS are used for primary autonomic dysfunction by increasing vascular tone. Erythropoietin is given to patients with neurogenic orthostatic hypotension and it works through increasing vascular volume and viscosity. Pediatrics The definition of hypotension changes in the pediatric population depending on the child's age as seen in the table below. The clinical history provided by the caretaker is the most important part in determining the cause of hypotension in pediatric patients. Symptoms for children with hypotension include increased sleepiness, not using the restroom as much (or at all), having difficulty breathing or breathing rapidly, or syncope. The treatment for hypotension in pediatric patients is similar to the treatment in adults by following the four first steps listed above (see Treatment). Children are more likely to undergo intubation during the treatment of hypotension because their oxygen levels drop more rapidly than adults. The closing of fetal shunts following birth can create instability in the "transitional circulation" of the fetus, and often creates a state of hypotension following birth; while many infants can overcome this hypotension through the closing of shunts, a mean blood pressure (MBP) of lower than 30 mmHg is correlated with severe cerebral injury and can be experienced by premature infants who have poor shunt closure. Etymology Hypotension, from Ancient Greek hypo-, meaning "under" or "less" + English tension, meaning "'strain" or "tightness". This refers to the under-constriction of the blood vessels and arteries which leads to low blood pressure.
Biology and health sciences
Cardiovascular disease
Health
501058
https://en.wikipedia.org/wiki/Jarosite
Jarosite
Jarosite is a basic hydrous sulfate of potassium and ferric iron (Fe-III) with a chemical formula of KFe3(SO4)2(OH)6. This sulfate mineral is formed in ore deposits by the oxidation of iron sulfides. Jarosite is often produced as a byproduct during the purification and refining of zinc and is also commonly associated with acid mine drainage and acid sulfate soil environments. Physical properties Jarosite has a trigonal crystal structure and is brittle, with basal cleavage, a hardness of 2.5–3.5, and a specific gravity of 3.15–3.26. It is translucent to opaque with a vitreous to dull luster, and is colored dark yellow to yellowish-brown. It can sometimes be confused with limonite or goethite with which it commonly occurs in the gossan (oxidized cap over an ore body). Jarosite is an iron analogue of the potassium aluminium sulfate, alunite. Solid solution series The alunite supergroup includes the alunite, jarosite, beudantite, crandallite and florencite subgroups. The alunite supergroup minerals are isostructural with each other and substitution between them occurs, resulting in several solid solution series. The alunite supergroup has the general formula AB3(TO4)2(OH)6. In the alunite subgroup B is Al, and in the jarosite subgroup B is Fe3+. The beudantite subgroup has the general formula AB3(XO4)(SO4)(OH)6, the crandallite subgroup AB3(TO4)2(OH)5•H2O and the florencite subgroup AB3(TO4)2(OH)5 or 6. In the jarosite-alunite series Al may substitute for Fe and a complete solid solution series between jarosite and alunite, KAl3(SO4)2(OH)6, probably exists, but intermediate members are rare. The material from Kopec, Czech Republic, has about equal Fe and Al, but the amount of Al in jarosite is usually small. When jarosite forms from pyrite oxidation in sedimentary clays, the main sources of K+ are illite, a non-swelling clay, or K-feldspar. In other geological settings mica's alteration can also be a source of potassium. In the jarosite-natrojarosite series Na substitutes for K to at least Na/K = 1:2.4 but the pure sodium end member NaFe3(SO4)2(OH)6 is not known in nature. Minerals with Na > K are known as natrojarosite. End member formation (jarosite and natrojarosite) is favoured by a low temperature environment, less than 100 °C, and is illustrated by the oscillatory zoning of jarosite and natrojarosite found in samples from the Apex Mine, Arizona, and Gold Hill, Utah. This indicates that there is a wide miscibility gap between the two end members, and it is doubtful whether a complete series exists between jarosite and natrojarosite. In hydroniumjarosite the hydronium ion H3O+ can also substitute for K+, with increased hydronium ion content causing a marked decrease in the lattice parameter c, although there is little change in a. Hydroniumjarosite will only form from alkali-deficient solutions, as alkali-rich jarosite forms preferentially. Divalent cations may also substitute for the monovalent cation K+ in the A site. Charge balance may be achieved in three ways. Firstly by replacing two monovalent cations by one divalent cation, and leaving an A site vacancy, as in plumbogummite, Pb2+Al3(PO4)2(OH)5.H2O, which is a member of the crandallite subgroup. Secondly by incorporating divalent ions in the B sites, as in osarizawaite, Pb2+Cu2+Al2(SO4)2(OH)6, alunite subgroup, and beaverite, Pb2+Cu2+(Fe3+,Al)2(SO4)2(OH)6, jarosite subgroup. Thirdly by replacing divalent anions with trivalent anions, as in beudantite, PbFe3+3(AsO4)3−(SO4)(OH)6, beudantite subgroup. History Jarosite was first described in 1852 by August Breithaupt in the Barranco del Jaroso in the Sierra Almagrera (near Los Lobos, Cuevas del Almanzora, Almería, Spain). The name jarosite is directly derived from "jara", the Spanish name of a yellow flower that belongs to the genus Cistus and grows in the sierra. The mineral and the flower have the same color. Mysterious spheres of clay, in diameter and covered with jarosite, have been found beneath the Temple of the Feathered Serpent, an ancient six-level stepped pyramid from Mexico City. Mars exploration Ferric sulfate and jarosite have been detected by three martian rovers: Spirit, Opportunity and Curiosity. These substances are indicative of strongly oxidizing conditions prevailing at the surface of Mars. In May 2009, the Spirit rover became stuck when it drove over a patch of soft ferric sulfate that had been hidden under a veneer of normal-looking soil. Because iron sulfate has very little cohesion, the rover's wheels could not gain sufficient traction to pull the body of the rover out of the iron sulfate patch. Multiple techniques were attempted to extricate the rover, but the wheels eventually sank so deeply into the iron sulfate that the body of the rover came to rest on the Martian surface, preventing the wheels from exerting any force on the material below them. As the JPL team failed to recover the mobility of Spirit, it signified the end of the journey for the rover. Antarctica deep borehole On Earth, jarosite is mainly associated with the ultimate stage of pyrite oxidation in clay environment, and can also be found in mine tailings waste where acidic conditions prevail. Against all expectations, jarosite has also been fortuitously discovered in minute quantities in the form of small dust particles in ice cores recovered from a deep borehole in Antarctica. That surprising discovery was made by geologists who were searching for specific minerals capable of indicating ice age cycles within the layers of a 1620-meter-long ice core. Geologists speculate that jarosite dust could also have accumulated within ice in glaciers on Mars. However, that hypothesis is a matter of controversy because, on Mars, jarosite deposits can be very thick (up to 10 meters). However, Mars is also a very dusty planet and, in the absence of plate tectonics on Mars, glacial dust deposits might have accumulated over long periods of time. Use in materials science Jarosite is also a more generic term denoting an extensive family of compounds of the form AM3(OH)6(SO4)2, where A+ = Na, K, Rb, NH4, H3O, Ag, Tl and M3+ = Fe, Cr, V. In condensed matter physics and materials science they are renowned for containing layers with kagome lattice structure, relating to geometrically frustrated magnets.
Physical sciences
Minerals
Earth science
501176
https://en.wikipedia.org/wiki/Swallowtail%20butterfly
Swallowtail butterfly
Swallowtail butterflies are large, colorful butterflies in the family Papilionidae, and include over 550 species. Though the majority are tropical, members of the family inhabit every continent except Antarctica. The family includes the largest butterflies in the world, the birdwing butterflies of the genus Ornithoptera. Swallowtails have a number of distinctive features; for example, the papilionid caterpillar bears a repugnatorial organ called the osmeterium on its prothorax. The osmeterium normally remains hidden, but when threatened, the larva turns it outward through a transverse dorsal groove by inflating it with fluid. The forked appearance in some of the swallowtails' hindwings, which can be seen when the butterfly is resting with its wings spread, gave rise to the common name swallowtail. As for its formal name, Linnaeus chose Papilio for the type genus, as papilio is Latin for "butterfly". For the specific epithets of the genus, Linnaeus applied the names of Greek figures to the swallowtails. The type species: Papilio machaon honored Machaon, one of the sons of Asclepius, mentioned in the Iliad. Further, the species Papilio homerus is named after the Greek poet, Homer. The Mon of the Taira clan of Japan is an Agehachō (swallowtail butterfly). Taxonomy Subfamilies The genera of extant swallowtails are usually classified into three subfamilies, Baroniinae, Parnassiinae, and Papilioninae, the latter two being further divided into tribes. In swallowtails, besides morphological characteristics, the choice of food plants and ecological lifestyle reflect phylogeny and classification. Baroniinae The Baroniinae are a monotypic subfamily, restricted to a very small region in Mexico and are considered to be the most basal of the subfamilies. Baronia brevicornis is considered to be a relict species, and shares features with a fossil taxon Praepapilio. Baronia is unique among papilionidae' as having an Acacia species (family Leguminosae) as its food plant. Subfamily: Baroniene. Parnassiinae The Parnassiinae are a subfamily of essentially Holarctic butterflies. The vast majority of species, mostly Parnassius, can be found in mountain habitats. Parnassiinines can also be found in other habitats such as "arid deserts (Hypermnestra), humid forests (Luehdorfia) and even lowland meadows (Zerynthia)". The tribes recognized in the Parnassiinae are Parnassiini, Zerynthiini, and Luehdorfiini. Tribe Parnassiini contains two genera, Hypermnestra, largely confined to central Asia and the genus Parnassius (the Apollos), a distinctive group of many species, all of which are alpine and capable of living at high altitudes. Most Parnassius have two small reddish spots on their hindwings. The tribe Luehdorfiini contains the genera Archon of Asia minor and the genus Luehdorfia of China and Japan. These two tribes have evolved to change their food plants, while the third tribe, Zerynthiini, has retained the archetypical papilionid food plant, the lowland vine Aristolochia. Zerynthiini comprises four genera – Sericinus, Bhutanitis, Zerynthia and Allancastria. Subfamily: Parnassiinae. Papilioninae The tribes recognized in the Papilioninae are Leptocircini, Teinopalpini, Troidini, and Papilionini. Subfamily: Papilioninae. Praepapilioninae An additional subfamily, Praepapilioninae, consisting of a single genus Praepapilio, includes two species of extinct butterflies, each member being described from single fossils found in a middle Eocene deposit in Colorado, United States (Durden and Rose, 1978). Phylogeny A phylogeny of the Papilionidae based on Nazari (2007) is given: It is now accepted that the subfamily Papilioninae is monophyletic. The swallowtail butterflies in the nominate tribe Papilionini number about 225 species and studies have been made on their host plant coevolution and phylogeny. Old morphological classifications were also found to be valid in that they formed clusters. Species belonging to the groups that use Rutaceae as host plants formed two groups corresponding to Old World and American taxa. Those that fed on Lauraceae and Magnoliaceae were found to form another cluster which includes both Asian and American taxa. The Parnassinae, like the Papilioninae, were also believed to be monophyletic based on morphological studies but recent studies based on both morphological and molecular characteristics suggest that this is not the case. Of the Parnassiinae, the genera Parnassius and Hypermnestra were found to be extremely close based on molecular studies and are now considered to be part of the tribe Parnassiini. The two taxa, Archon and Luehdorfia, have been found to be closely related through analysis of nuclear and mitochondrial DNA, and, though they share no morphological similarities, have now been united in the tribe Luehdorfiini. The subfamily Baroniinae is represented by the sole representative species Baronia brevicornis. They are unique in the family to use the Fabaceae (Leguminosae) as their larval host plants. The Baronninae and the extinct subfamily Praepapilioninae share many external similarities and are traditionally considered to be the most primitive subfamilies and sister to the rest of the swallowtails. Recent research suggests that this may not be the case, the Baroniinae being closely related to only the Parnassiinae, and Praepapilio to only the Papilionini and neither taxa being sister to the rest of the swallowtails. Distribution As of 2005, 552 extant species have been identified which are distributed across the tropical and temperate regions. Various species inhabit altitudes ranging from sea level to high mountains, as in the case of most species of Parnassius. The majority of swallowtail species and the greatest diversity are found in the tropics and subtropical regions between 20°N and 20°S, particularly Southeast Asia, and between 20°N and 40°N in East Asia. Only 12 species are found in Europe and only one species, Papilio machaon is found in the British Isles. North America has 40 species, including several tropical species and Parnassius. The northernmost swallowtail is the Siberian Apollo (Parnassius arcticus), found in the Arctic Circle in northeastern Yakutia, at altitudes of 1500 meters above sea level. In the Himalayas, various Apollo species such as Parnassius epaphus, have been found at altitudes of 6,000 meters above sea level. Food The caterpillars of various swallowtail butterfly species feed on a wide range of different plants, most depending on only one of five families: Aristolochiaceae, Annonaceae, Lauraceae, Umbelliferae (Apiaceae) and Rutaceae. By eating some of these toxic plants, the caterpillars sequester aristolochic acid which renders both the caterpillars and the butterflies of some of these as toxic, thus protecting them from predators. Similarly, the Parnassius smintheus sequesters sarmentosin from its host plant Sedum lanceolatum for protection from predators. Swallowtail tribes Zerynthiini (Parnassiinae), Luehdorfiini (Parnassiinae) and Troidini (Papilioninae), almost exclusively use the family Aristolochiaceae as their host plants. For example, the eastern black swallowtail's (Papilio polyxenes) main host plant in the wild is Queen Anne's lace, but they also eat garden plants in the carrot family, including carrots, parsley, dill, and fennel. Adult swallowtails sip nectar, but also mud and sometimes manure. Life cycle The detailed descriptions of morphological characteristics of the Papilionidae, as quoted in Bingham (1905) are as follows: Stages of development of a papilionid, the giant swallowtail (Papilio cresphontes) Distinguishing characteristics The key characteristics that differentiate the Papilionidae from the other butterfly families are: The osmeterium is a forked, fleshy eversible organ found in the prothoracic segment of caterpillars. Venation – in swallowtails, the second anal vein, 2A, extends up to the wing margin and does not link with the first anal vein, 1A. These veins are fused in other butterfly families and 2A does not reach the wing margin. The sclerites of the cervix (membranous neck between the head and thorax) are fused beneath the neck where the muscles for head movement are anchored. Special adaptations and defense Swallowtail butterflies practice Batesian mimicry, a behavior in which the butterflies' appearance closely resemble that of distasteful species that prevents predation. Swallowtails differ from many animals that practice mimicry. The tiger swallowtail butterfly (Papilio glaucus), exhibits a female-limited polymorphism for Batesian mimicry and others, such as the Canadian tiger swallowtail (Papilio canadensis) do not display any form of mimicry. Predators include the red-winged blackbird, Pennsylvania firefly, five-lined skink, green darner, goldenrod spider, Chinese mantis, fiery searcher, and striped skunk. Biological basis for polymorphisms in mimicry Not all individuals in some species are identical in appearance. For example, Papilio glaucus (eastern tiger swallowtail), Y-linkage determines whether the females are either wild-type (yellow and black) or melanic (dark melanin replaces the yellow background). This genetic difference stems from the fact that melanism is controlled by a single gene, which controls the level of dopamine in the organism. The enzyme BAS, which assists dopamine in producing the yellow pigmentation, normally found on the wings' background, is suppressed. Without the pigmentation, the butterfly appears mostly black (the melanic form) and is a Batesian mimic of Battus philenor, the pipevine swallowtail. There are also Papilio glaucus that are not wholly black; several possess an intermediate "sooty" color and are sensitive to temperature. The different polymorphisms (wild-type, melanic, and the 'sooty' intermediate) depend upon the geographical distribution and abundance of its mimic, the Battus philenor, whose wing color varies depending on its geographical location. In order to be successfully confused for the B. philenor by predators, the Papilio glaucus's background wing color matches that of the B. philenor residing in the same regional area. Studies support this theory; in the southeastern United States, the relative abundance of melanic females has been found to geographically correlate with B. philenor. Mimicry Only certain subsets of swallowtails practice mimicry. Species differ in whether one or both sexes is mimetic, and whether the mimicry is monomorphic or polymorphic. A phenomenon which has received particular attention is female-limited polymorphism, in which only the females of a species are mimetic and polymorphic, often mimicking different, distantly-related aposematic butterflies. This polymorphism is seen in Papilio dardanus, the African swallowtail butterfly, whose females have three different morphs for wing color pattern: a black-and-white pattern for Batesian mimicry, a black-and-yellow pattern that resembles the males of the species, and a pattern with orange patches that resembles the elderly males of the species. Given that the males of the species, which do not have Batesian mimicry, are preyed upon much more frequently by predators than the females, an ongoing question is why females would exhibit the non-mimetic wing pattern, which would seemingly lower their fitness compared to the mimicry form. The pipevine swallowtail exhibits Batesian mimicry as well. Several hypotheses for this phenomenon were made, the two noteworthy being the pseudosexual selection hypothesis and the male avoidance hypothesis. In the pseudosexual hypothesis, male butterflies aggressively approached the male-looking females and then mellowed their behavior into sexual behavior when they were close enough to identify them as females. In the male avoidance hypothesis, female butterflies disguise themselves in an attempt to evade male harassment, as courtship can be harmful, time-consuming, and attract predators. One study recorded male responses to females of each morphs and found that the males consistently favored the Batesian mimics, then the black and yellow, and then the morph with orange patches. The scientists concluded that frequency-dependent selection did lead to equal success for all three alternative strategies: the Batesian females suffered the fewest predators but their fitness was reduced the most by sexual harassment, while the other two faced lower sexual harassment but also lost fitness from predators' attacks. Mating and young After mating, the male Parnassines produce a glue-like substance that is used to seal the female genital opening and prevent other males from mating. They lay individual eggs on the underside of the leaves of their food plants. There is no parental investment once the eggs have been laid. The pupae are typically attached to the substrate by the cremaster but with head up held by a silk girdle. The Apollos, however, pupate in debris on the ground and also build a loose cocoon. In the temperate regions, the winters are passed in a pupal diapause stage. In culture Since swallowtail butterflies are large, colorful, and attractive, they have been targeted by butterfly collectors. The largest of these, the birdwing butterflies are particularly sought after and are cultured in butterfly farms to supply collectors. Many members of the family, as larvae, feed on plants of the citrus family Rutaceae, making some of these attractive butterflies pests in citrus orchards. The Oregon swallowtail is the state insect of Oregon. The eastern tiger swallowtail is the state insect of Virginia and the state butterfly of Georgia, Delaware, and South Carolina. The black swallowtail is the state butterfly of Oklahoma.
Biology and health sciences
Lepidoptera
null
501188
https://en.wikipedia.org/wiki/Mucus
Mucus
Mucus (, ) is a slippery aqueous secretion produced by, and covering, mucous membranes. It is typically produced from cells found in mucous glands, although it may also originate from mixed glands, which contain both serous and mucous cells. It is a viscous colloid containing inorganic salts, antimicrobial enzymes (such as lysozymes), immunoglobulins (especially IgA), and glycoproteins such as lactoferrin and mucins, which are produced by goblet cells in the mucous membranes and submucosal glands. Mucus serves to protect epithelial cells in the linings of the respiratory, digestive, and urogenital systems, and structures in the visual and auditory systems from pathogenic fungi, bacteria and viruses. Most of the mucus in the body is produced in the gastrointestinal tract. Amphibians, fish, snails, slugs, and some other invertebrates also produce external mucus from their epidermis as protection against pathogens, to help in movement, and to line fish gills. Plants produce a similar substance called mucilage that is also produced by some microorganisms. Respiratory system In the human respiratory system, mucus is part of the airway surface liquid (ASL), also known as epithelial lining fluid (ELF), that lines most of the respiratory tract. The airway surface liquid consists of a sol layer termed the periciliary liquid layer and an overlying gel layer termed the mucus layer. The periciliary liquid layer is so named as it surrounds the cilia and lies on top of the surface epithelium. The periciliary liquid layer surrounding the cilia consists of a gel meshwork of cell-tethered mucins and polysaccharides. The mucus blanket aids in the protection of the lungs by trapping foreign particles before they can enter them, in particular through the nose during normal breathing. Mucus is made up of a fluid component of around 95% water, the mucin secretions from the goblet cells, and the submucosal glands (2–3% glycoproteins), proteoglycans (0.1–0.5%), lipids (0.3–0.5%), proteins, and DNA. The major mucins secreted – MUC5AC and MUC5B - are large polymers that give the mucus its rheologic or viscoelastic properties. MUC5AC is the main gel-forming mucin secreted by goblet cells, in the form of threads and thin sheets. MUC5B is a polymeric protein secreted from submucosal glands and some goblet cells, and this is in the form of strands. In the airways—the trachea, bronchi, and bronchioles—the lining of mucus is produced by specialized airway epithelial cells called goblet cells, and submucosal glands. Small particles such as dust, particulate pollutants, and allergens, as well as infectious agents and bacteria are caught in the viscous nasal or airway mucus and prevented from entering the system. This process, together with the continual movement of the cilia on the respiratory epithelium toward the oropharynx (mucociliary clearance), helps prevent foreign objects from entering the lungs during breathing. This explains why coughing often occurs in those who smoke cigarettes. The body's natural reaction is to increase mucus production. In addition, mucus aids in moisturizing the inhaled air and prevents tissues such as the nasal and airway epithelia from drying out. Mucus is produced continuously in the respiratory tract. Mucociliary action carries it down from the nasal passages and up from the rest of the tract to the pharynx, with most of it being swallowed subconsciously. Sometimes in times of respiratory illness or inflammation, mucus can become thickened with cell debris, bacteria, and inflammatory cells. It is then known as phlegm which may be coughed up as sputum to clear the airway. Respiratory tract Increased mucus production in the upper respiratory tract is a symptom of many common ailments, such as the common cold, and influenza. Nasal mucus may be removed by blowing the nose or by using nasal irrigation. Excess nasal mucus, as with a cold or allergies, due to vascular engorgement associated with vasodilation and increased capillary permeability caused by histamines, may be treated cautiously with decongestant medications. Thickening of mucus as a "rebound" effect following overuse of decongestants may produce nasal or sinus drainage problems and circumstances that promote infection. During cold, dry seasons, the mucus lining nasal passages tends to dry out, meaning that mucous membranes must work harder, producing more mucus to keep the cavity lined. As a result, the nasal cavity can fill up with mucus. At the same time, when air is exhaled, water vapor in breath condenses as the warm air meets the colder outside temperature near the nostrils. This causes an excess amount of water to build up inside nasal cavities. In these cases, the excess fluid usually spills out externally through the nostrils. In the lower respiratory tract impaired mucociliary clearance due to conditions such as primary ciliary dyskinesia may result in mucus accumulation in the bronchi. The dysregulation of mucus homeostasis is the fundamental characteristic of cystic fibrosis, an inherited disease caused by mutations in the CFTR gene, which encodes a chloride channel. This defect leads to the altered electrolyte composition of mucus, which triggers its hyperabsorption and dehydration. Such low-volume, viscous, acidic mucus has a reduced antimicrobial function, which facilitates bacterial colonisation. The thinning of the mucus layer ultimately affects the periciliary liquid layer, which becomes dehydrated, compromising ciliary function, and impairing mucociliary clearance. A respiratory therapist can recommend airway clearance therapy which uses a number of clearance techniques to help with the clearance of mucus. Mucus hypersecretion In the lower respiratory tract excessive mucus production in the bronchi and bronchioles is known as mucus hypersecretion. Chronic mucus hypersecretion results in the chronic productive cough of chronic bronchitis, and is generally synonymous with this. Excessive mucus can narrow the airways, limit airflow, and accelerate a decline in lung function. Digestive system In the human digestive system, mucus is used as a lubricant for materials that must pass over membranes, e.g., food passing down the esophagus. Mucus is extremely important in the gastrointestinal tract. It forms an essential layer in the colon and in the small intestine that helps reduce intestinal inflammation by decreasing bacterial interaction with intestinal epithelial cells. The layer of mucus of the gastric mucosa lining the stomach is vital to protect the stomach lining from the highly acidic environment within it. Reproductive system In the human female reproductive system, cervical mucus prevents infection and provides lubrication during sexual intercourse. The consistency of cervical mucus varies depending on the stage of a woman's menstrual cycle. At ovulation cervical mucus is clear, runny, and conducive to sperm; post-ovulation, mucus becomes thicker and is more likely to block sperm. Several fertility awareness methods rely on observation of cervical mucus, as one of three primary fertility signs, to identify a woman's fertile time at the mid-point of the cycle. Awareness of the woman's fertile time allows a couple to time intercourse to improve the odds of pregnancy. It is also proposed as a method to avoid pregnancy. Clinical significance In general, nasal mucus is clear and thin, serving to filter air during inhalation. During times of infection, mucus can change color to yellow or green either as a result of trapped bacteria or due to the body's reaction to viral infection. For example, Staphylococcus aureus infection may turn the mucus yellow. The green color of mucus comes from the heme group in the iron-containing enzyme myeloperoxidase secreted by white blood cells as a cytotoxic defense during a respiratory burst. In the case of bacterial infection, the bacterium becomes trapped in already-clogged sinuses, breeding in the moist, nutrient-rich environment. Sinusitis is an uncomfortable condition that may include congestion of mucus. A bacterial infection in sinusitis will cause discolored mucus and would respond to antibiotic treatment; viral infections typically resolve without treatment. Almost all sinusitis infections are viral and antibiotics are ineffective and not recommended for treating typical cases. In the case of a viral infection such as cold or flu, the first stage and also the last stage of the infection cause the production of a clear, thin mucus in the nose or back of the throat. As the body begins to react to the virus (generally one to three days), mucus thickens and may turn yellow or green. Obstructive lung diseases often result from impaired mucociliary clearance that can be associated with mucus hypersecretion, and these are sometimes referred to as mucoobstructive lung diseases. Techniques of airway clearance therapy can help to clear secretions, maintain respiratory health, and prevent inflammation in the airways. A unique umbilical cord lining epithelial stem cell expresses MUC1, termed (CLEC-muc). This has been shown to have good potential in the regeneration of the cornea. Properties of mucus Tunable swelling capacity Mucus is able to absorb water or dehydrate through pH variations. The swelling capacity of mucus stems from the bottlebrush structure of mucin within which hydrophilic segments provide a large surface area for water absorption. Moreover, the tunability of swelling effect is controlled by polyelectrolyte effect. Polyelectrolyte effect in mucus Polymers with charged molecules are called polyelectrolytes. Mucins, a kind of polyelectrolyte proteoglycans, are the main component of mucus, which provides the polyelectrolyte effect in mucus. The process of inducing this effect comprises two steps: attraction of counter-ions and water compensation. When exposed in physiological ionic solution, the charged groups in the polyelectrolytes attract counter-ions with opposite charges, thereby leading to a solute concentration gradient. An osmotic pressure is introduced to equalize the concentration of solute throughout the system by driving water to flow from the low concentration areas to the high concentration areas. In short, the influx and outflux of water within mucus, managed by the polyelectrolyte effect, contribute to mucus' tunable swelling capacity. Mechanism of pH-tunable swelling The ionic charges of mucin are mainly provided by acidic amino acids including aspartic acid (pKa=3.9) and glutamic acid (pKa=4.2). The charges of acidic amino acids will change with environmental pH value due to acid dissociation and association. Aspartic acid, for example, has a negative side chain when the pH value is above 3.9, while a neutrally charged side chain will be introduced as pH value drops below 3.9. Thus, the number of negative charges in mucus is influenced by the pH value of surrounding environment. That is, the polyelectrolyte effect of mucus is largely affected by the pH value of solution due to the charge variation of acidic amino acid residues on the mucin backbone. For instance, the charged residue on mucin is protonated at a normal pH value of the stomach, approximately pH 2. In this case, there is scarcely polyelectrolyte effect, thereby causing compact mucus with little swelling capacity. However, a kind of bacteria, Helicobacter pylori, is prone to producing base to elevate the pH value in stomach, leading to the deprotonation of aspartic acids and glutamic acids, i.e., from neutral to negative-charged. The negative charges in the mucus greatly increase, thus inducing the polyelectrolyte effect and the swelling of the mucus. This swelling effect increases the pore size of the mucus and decreases mucus' viscosity, which allows bacteria to penetrate and migrate into the mucus and cause disease. Charge selectivity The high selective permeability of mucus plays a crucial role in the healthy state of human beings by limiting the penetration of molecules, nutrients, pathogens, and drugs. The charge distribution within mucus serves as a charge selective diffusion barrier, thus significantly affecting the transportation of agents. Among particles with various surface zeta potentials, cationic particles tend to have a low depth of penetration, neutral ones possess medium penetration, and anionic ones have the largest penetration depth. Furthermore, the effect of charge selectivity changes when the status of the mucus varies, i.e., native mucus has a threefold higher potential to limit agent penetration than purified mucus. Other animals Mucus is also produced by a number of other animals. All fish are covered in mucus secreted from glands all over their bodies. Invertebrates such as snails and slugs secrete mucus called snail slime to enable movement, and to prevent their bodies from drying out. Their reproductive systems also make use of mucus for example in the covering of their eggs. In the unique mating ritual of Limax maximus the mating slugs lower themselves from elevated locations by a mucus thread. Mucus is an essential constituent of hagfish slime used to deter predators. Mucus is produced by the endostyle in some tunicates and larval lampreys to help in filter feeding.
Biology and health sciences
Exocrine system
Biology
501250
https://en.wikipedia.org/wiki/Naproxen
Naproxen
Naproxen, sold under the brand name Aleve among others, is a nonsteroidal anti-inflammatory drug (NSAID) used to treat pain, menstrual cramps, and inflammatory diseases such as rheumatoid arthritis, gout and fever. It is taken orally. It is available in immediate and delayed release formulations. Onset of effects is within an hour and lasts for up to twelve hours. Naproxen is also available in salt form, naproxen sodium, which has better solubility when taken orally. Common side effects include dizziness, headache, bruising, allergic reactions, heartburn, and stomach pain. Severe side effects include an increased risk of heart disease, stroke, gastrointestinal bleeding, and stomach ulcers. The heart disease risk may be lower than with other NSAIDs. It is not recommended in people with kidney problems. Use is not recommended in the third trimester of pregnancy. Naproxen is a nonselective COX inhibitor. As an NSAID, naproxen appears to exert its anti-inflammatory action by reducing the production of inflammatory mediators called prostaglandins. It is metabolized by the liver to inactive metabolites. Naproxen was patented in 1967 and approved for medical use in the United States in 1976. In the United States it is available over-the-counter and as a generic medication. In 2022, it was the 88th most commonly prescribed medication in the United States, with more than 7million prescriptions. Medical uses Naproxen's medical uses are related to its mechanism of action as an anti-inflammatory compound. Naproxen is used to treat a variety of inflammatory conditions and symptoms that are due to excessive inflammation, such as pain and fever (naproxen has fever-reducing, or antipyretic, properties in addition to its anti-inflammatory activity). Naproxen's anti-inflammatory properties may relieve pain caused by inflammatory conditions such as migraine, osteoarthritis, kidney stones, rheumatoid arthritis, psoriatic arthritis, gout, ankylosing spondylitis, menstrual cramps, tendinitis, and bursitis. Naproxen sodium is used as a "bridge therapy" in medication-overuse headache to slowly take patients off other medications. Available formulations Naproxen sodium is available as both an immediate-release and an extended-release tablet. The extended-release formulations (sometimes called "sustained release", or "enteric coated") take longer to take effect than the immediate-release formulations and therefore are less useful when immediate pain relief is desired. Extended-release formulations are more useful for the treatment of chronic, or long-lasting, conditions, in which long-term pain relief is desirable. Pregnancy and lactation As with all non-steroidal anti-inflammatory medications (NSAIDs), naproxen use should be avoided in pregnancy due to the importance of prostaglandins in vascular and renal function in the fetus. NSAIDs should especially be avoided in the third trimester. Small amounts of naproxen are excreted in breast milk. However, adverse effects are uncommon in infants breastfed from a mother taking naproxen. Adverse effects Common adverse effects include dizziness, drowsiness, headache, rash, bruising, and gastrointestinal upset. Heavy use is associated with an increased risk of end-stage renal disease and kidney failure. Naproxen may cause muscle cramps in the legs in 3% of people. In October 2020, the U.S. Food and Drug Administration (FDA) required the drug label to be updated for all nonsteroidal anti-inflammatory medications to describe the risk of kidney problems in unborn babies that result in low amniotic fluid. They recommend avoiding NSAIDs in pregnant women at 20 weeks or later in pregnancy. Gastrointestinal As with other non-COX-2 selective NSAIDs, naproxen can cause gastrointestinal problems, such as heartburn, constipation, diarrhea, ulcers and stomach bleeding. Naproxen should be taken orally with, or just after food, to decrease the risk of gastrointestinal side effects. Persons with a history of ulcers or inflammatory bowel disease should consult a doctor before taking naproxen. In U.S. markets, naproxen is sold with boxed warnings about the risk of gastrointestinal ulceration or bleeding. Naproxen poses an intermediate risk of stomach ulcers compared with ibuprofen, which is low-risk, and indometacin, which is high-risk. To reduce stomach ulceration risk, it is often combined with a proton-pump inhibitor (a medication that reduces stomach acid production) during long-term treatment of those with pre-existing stomach ulcers or a history of developing stomach ulcers while on NSAIDs. Cardiovascular COX-2 selective and nonselective NSAIDs have been linked to increases in the number of serious and potentially fatal cardiovascular events, such as myocardial infarctions and strokes. Naproxen is, however, associated with the smallest overall cardiovascular risks. Cardiovascular risk must be considered when prescribing any nonsteroidal anti-inflammatory drug. The drug had roughly 50% of the associated risk of stroke compared with ibuprofen and was also associated with a reduced number of myocardial infarctions compared with control groups. A study found that high-dose naproxen induced near-complete suppression of platelet thromboxane throughout the dosing interval and appeared not to increase cardiovascular disease (CVD) risk, whereas other non-aspirin high-dose NSAID regimens had only transient effects on platelet COX-1 and were associated with a small but definite vascular hazard. Conversely, naproxen was associated with higher rates of upper gastrointestinal bleeding complications compared with other NSAIDs. Interactions Drug–drug interactions Naproxen may interact with antidepressants, lithium, methotrexate, probenecid, warfarin and other blood thinners, heart or blood pressure medications, including diuretics, or steroid medicines such as prednisone. NSAIDs such as naproxen may interfere with and reduce the efficacy of SSRI antidepressants, as well as increase the risk of bleeding greater than the individual bleeding risk of either class of agent, when taken together. Naproxen is not contraindicated in the presence of SSRIs, though concomitant use of the medications should be done with caution. Alcohol consumption increases the risk of gastrointestinal bleeding when combined with NSAIDs like naproxen in a dose-dependent manner (that is, the higher the dose of naproxen, the higher the risk of bleeding). The risk is highest for people who are heavy drinkers. Pharmacology Mechanism of action Naproxen works by reversibly inhibiting both the COX-1 and COX-2 enzymes as a non-selective coxib. Pharmacokinetics Naproxen is a minor substrate of CYP1A2 and CYP2C9. It is extensively metabolized in the liver to 6-O-desmethylnaproxen, and both the parent drug and the desmethyl metabolite undergo further metabolism to their respective acylglucuronide conjugated metabolites. An analysis of two clinical trials shows that naproxen's time to peak plasma concentration occurs between 2 and 4 hours after oral administration, though naproxen sodium reaches peak plasma concentrations within 1–2 hours. Pharmacogenetics The pharmacogenetics of naproxen has been studied to better understand its adverse effects. In 1998, a small pharmacokinetic (PK) study failed to show that differences in a patient's ability to clear naproxen from the body could account for differences in a patient's risk of experiencing the adverse effect of a serious gastrointestinal bleed while taking naproxen. However, the study failed to account for differences in the activity of CYP2C9, a drug-metabolizing enzyme that is necessary for clearing naproxen. Studies on the relationship between CYP2C9 genotype and NSAID-induced gastrointestinal bleeds have shown that genetic variants in CYP2C9 that reduce the clearance of major CYP2C9 substrates (like naproxen) increase the risk of NSAID-induced gastrointestinal bleeds, especially for homozygous defective variants. Chemistry Naproxen is a member of the 2-arylpropionic acid (profen) family of NSAIDs. The free acid is an odorless, white to off-white crystalline substance. Naproxen free base is lipid-soluble and practically insoluble in water, while naproxen sodium and many other salts are freely soluble in water, often soluble in methanol, and sparingly soluble in alcohol; check the specific solubility of each salt before use. Naproxen has a melting point of 152–155 °C, while naproxen salts tend to have higher melting points. Synthesis Naproxen has been industrially produced by Syntex starting from 2-naphthol as follows: Society and culture Brand names Naproxen and naproxen sodium are marketed under various brand names, including Accord, Aleve, Anaprox, Antalgin, Apranax, Feminax Ultra, Flanax, Inza, Maxidol, Nalgesin, Naposin, Naprelan, Naprogesic, Naprosyn, Narocin, Pronaxen, Proxen, Soproxen, and Xenifar. It is also available as the combination naproxen/esomeprazole magnesium in delayed-release tablets under the brand name Vimovo. Access restrictions Syntex first marketed naproxen in 1976, as the prescription drug Naprosyn. They first marketed naproxen sodium under the brand name Anaprox in 1980. It remains a prescription-only drug in much of the world. In the United States, the Food and Drug Administration (FDA) approved it as an over-the-counter (OTC) drug in 1994. OTC preparations of naproxen in the U.S. are mainly marketed by Bayer HealthCare under the brand name Aleve and generic store brand formulations in 220mg tablets. In Australia, packets of 275mg tablets of naproxen sodium are Schedule 2 pharmacy medicines, with a maximum daily dose of five tablets or 1375mg. In the United Kingdom, 250mg tablets of naproxen were approved for OTC sale under the brand name Feminax Ultra in 2008, for the treatment of primary dysmenorrhoea in women aged 15 to 50. In the Netherlands, 220mg and 275mg tablets are available OTC in drugstores, 550mg is OTC only at pharmacies. Aleve became available over the counter in some provinces in Canada on 14 July 2009, but not British Columbia, Quebec or Newfoundland and Labrador; it subsequently became available OTC in British Columbia in January 2010. Toxicology scandal Naproxen was one of the four substances named in the prosecution of Industrial Bio-Test Laboratories (IBT) for fraudulent toxicology testing. Naproxen passed subsequent legitimate toxicology testing. Ecological effects Naproxen has been found in groundwater and drinking water in concentrations high enough to have adverse effects on invertebrates including fungi, algae, bacteria and . Naproxen is not thoroughly removed by conventional water treatment methods, and its degradation pathways in the environment are limited. Some methods more successfully remove naproxen from wastewater, including metal-organic complexes and porous carbon. Although the levels are generally low enough to not be acutely toxic, sub-lethal effects may still occur, such as reduced photosynthetic ability. Research Naproxen may have antiviral activity against influenza. In laboratory research, it blocks the RNA-binding groove of the nucleoprotein of the virus, preventing the formation of the ribonucleoprotein complex—thus taking the viral nucleoproteins out of circulation. Veterinary use Horses Naproxen is given by mouth to horses at a dose of 10mg/kg and has shown to have a wide safety margin (no toxicity when given at three times the recommended dose for 42 days). It is more effective for myositis than the commonly used NSAID phenylbutazone, and has shown especially good results for treatment of equine exertional rhabdomyolysis, a disease of muscle breakdown; it is less commonly used for musculoskeletal disease.
Biology and health sciences
Pain treatments
Health
501478
https://en.wikipedia.org/wiki/Stable
Stable
A stable is a building in which livestock, especially horses, are kept. It most commonly means a building that is divided into separate stalls for individual animals and livestock. There are many different types of stables in use today; the American-style barn, for instance, is a large barn with a door at each end and individual stalls inside or free-standing stables with top and bottom-opening doors. The term "stable" is additionally utilised to denote a collection of animals under the care of a single owner, irrespective of their housing or whereabouts. The exterior design of a stable can vary widely, based on climate, building materials, historical period and cultural styles of architecture. A wide range of building materials can be used, including masonry (bricks or stone), wood and steel. Stables also range widely in size, from a small building housing one or two animals to facilities at agricultural shows or race tracks that can house hundreds of animals. History The stable is typically historically the second-oldest building type on the farm. The world's oldest horse stables were discovered in the ancient city of Pi-Ramesses in Qantir, in Ancient Egypt, and were established by Ramesses II (c. 1304–1213 BC). These stables covered approximately 182,986 square feet, had floors sloped for drainage, and could contain about 480 horses. Free-standing stables began to be built from the 16th century. They were well built and placed near the house because these animals were highly valued and carefully maintained. They were once vital to the economy and an indicator of their owners' position in the community. Relatively few examples survive of complete interiors (i.e. with stalls, mangers and feed racks) from the mid-19th century or earlier. Traditionally, stables in Great Britain had a hayloft on their first (i.e. upper) floor and a pitching door at the front. Doors and windows were symmetrically arranged. Their interiors were divided into stalls and usually included a large stall for a foaling mare or sick horse. The floors were cobbled (or, later, bricked) and featured drainage channels. An outside stone stairway constructed against the side of the building was common for reaching the upper level. Horses For horses, stables are often part of a larger complex which includes trainers, vets and farriers. Other uses The word stable is also used metonymically to refer to the collection of horses that the building contains (for example, the college's stable includes a wide variety of breeds) and even, by extension, metaphorically to refer to a group of people—often (but not exclusively) athletes—trained, coached, supervised or managed by the same person or organisation. For example, art galleries typically refer to the artists they represent as their stable of artists. Analogously, car enthusiast magazines sometimes speak of collectible cars in this way, referring to the cars in a collector's stable (most especially when the metaphor can play on the word association of pony cars). Historically, the headquarters of a unit of cavalry, not simply their horses' accommodation, was known as a "stable". Gallery
Technology
Buildings and infrastructure
null
501723
https://en.wikipedia.org/wiki/Takifugu
Takifugu
Takifugu, also known by the Japanese name , is a genus of pufferfish with 25 species, most of which are native to salt and brackish waters of the northwest Pacific, but a few species are found in freshwater in Asia or more widely in the Indo-Pacific region. Their diet consists mostly of algae, molluscs, invertebrates and sometimes crustaceans. The fish contains lethal amounts of the poison tetrodotoxin in the internal organs, especially the liver and the ovaries, but also in the skin and the testes. The poison paralyzes the muscles while the victim stays fully conscious, and eventually dies from asphyxiation. There is currently no antidote, and the standard medical approach is to try to support the respiratory and circulatory system until the effect of the poison wears off. Distribution and conservation status There are 25 species belonging to the genus Takifugu. Most species are restricted to salt and brackish waters of the northwest Pacific, but a few occur more widely in the Indo-Pacific region or in freshwater of Asia. Although several are euryhaline (can adapt to various salinities) to some extent, most are unable to live in freshwater. Two exceptions are the anadromous Takifugu obscurus and the Takifugu ocellatus, which lives in coastal marine waters but migrates into fresh water to spawn in rivers. Most species in the genus are not considered threatened, but there are two notable exceptions: the critically endangered Takifugu chinensis and the endangered Takifugu plagiocellatus. Takifugu rubripes serves as a model organism in biological research. Morphology and behaviour Not all Takifugu have been studied in detail, but the most researched species is Takifugu rubripes, due to the commercial farming of this fish for human consumption. Takifugu rubripes, for example, breeds from March to May and lays eggs attached to rocks at a depth of around . As far as known, most species live exclusively in marine and brackish water, also breeding in this habitat. The anadromous Takifugu obscurus migrates from its coastal marine habitat into fresh water to spawn. An even more exceptional and unique breeding behavior is displayed by Takifugu niphobles. They gather in groups at certain beaches, throw themselves onto land where fertilization happens and then return to the water. The eggs either float back into the water or may stay on land under rocks for a period, only hatching when again submerged by high tide. This breeding behavior is unique among pufferfish, but found in a few other unrelated fish like capelin and grunion. Fugu can also change color over time. This helps them to camouflage. Toxicity The fish's main defense is the neurotoxin contained in its internal organs, mainly the ovaries and the liver, to a lesser extent in the intestines and the skin, and only minute amounts in the muscles and blood. This makes the fugu a lethal meal for most predators, including the occasional human. The toxin is called tetrodotoxin, or more precisely anhydrotetrodotoxin 4-epitetrodotoxin and is about 1200 times deadlier than cyanide. This poison can also be found in other animals such as the blue-ringed octopus, cone snails, and even some newts. The pufferfish does not create the poison itself; rather it is generated by bacteria e.g. Pseudomonas within the fish. The fish obtains the bacteria by eating food containing these bacteria. Pufferfish that are born and grown in captivity do not produce tetrodotoxin until they receive some of the poison-producing bacteria, often by eating tissues from a toxin-producing fish. Also, some fish are more poisonous than others. Each fish has enough poison to kill around thirty adult humans. Genome Apparently due to some unknown selection pressure, intronic and extragenic sequences have been drastically reduced within this family. As a result, they have the smallest-known genomes yet found amongst the vertebrate animals, while containing a genetic repertoire very similar to other fishes and thus comparable to vertebrates generally. Since these genomes are relatively compact it is relatively fast and inexpensive to compile their complete sequences, as has been done for two species of pufferfishes (Takifugu rubripes and Tetraodon nigroviridis). The former species was the second vertebrate in history to have its genome mapped, after humans. Species , there are 25 recognized species in the genus Takifugu: * Fish that have edible body parts according to the Japanese Ministry of Health and Welfare
Biology and health sciences
Acanthomorpha
Animals
501780
https://en.wikipedia.org/wiki/Quoll
Quoll
Quolls (; genus Dasyurus) are carnivorous marsupials native to Australia and New Guinea. They are primarily nocturnal, and spend most of the day in a den. Of the six species of quoll, four are found in Australia and two in New Guinea. Another two species are known from fossil remains in Pliocene and Pleistocene deposits in Queensland. Genetic evidence indicates that quolls evolved around 15 million years ago in the Miocene, and that the ancestors of the six species had all diverged by around four million years ago. The six species vary in weight and size, from to . They have brown or black fur and pink noses. They are largely solitary, but come together for a few social interactions, such as mating, which occurs during the winter season. A female gives birth to up to 30 pups, but the number that can be raised to adulthood is limited by the number of teats (6–7). They have a life span of 1–5 years (species dependent). Quolls eat smaller mammals, small birds, lizards, and insects. All species have drastically declined in numbers since Australasia was colonised by Europeans, with one species, the eastern quoll, becoming extinct on the Australian mainland in the 1960s. Major threats to their survival include the toxic cane toad, predators such as feral cats and foxes, urban development, and poison baiting. Conservation efforts include captive breeding programs and reintroductions. Taxonomy The name Dasyurus (from Greek δασύουρος, dasýouros) means "hairy-tail", and was coined by Étienne Geoffroy Saint-Hilaire in 1796 (from δασύς : dasýs "hairy" and οὐρά : ourá "tail"). In 1770, Captain Cook collected quolls on his exploration of the east coast of Australia, adopting an Aboriginal name for the animals. Although the origin of Cook's specimens are unclear, the word and its variants je-quoll, jaquol or taquol are derived from the word dhigul in the language of the Guugu Yimithirr people of far north Queensland. No evidence indicates the local indigenous people used the word in the Sydney area. They were likened in appearance to a polecat or marten in the earliest reports, the tiger quoll (spotted-tailed) being called "spotted marten" and eastern quoll "spotted opossum", but by 1804, the names "native fox", "native cat" and "tiger cat" had been adopted by early settlers; quolls are still called "marsupial foxes" or "marsupial cats". In the 1960s, noted naturalist David Fleay pushed for the revival of the term "quoll" to replace the then-current vernacular names that he felt were misleading. Four species have been recovered from Pleistocene cave deposits from Mount Etna Caves National Park near Rockhampton in central Queensland. Remains of the spotted-tailed quoll and the northern quoll, and a species either identical or very similar to the eastern quoll, as well as a prehistoric species as yet undescribed, all lived in what was a rainforest climate. The northern quoll is still found in the region. The fossil species D. dunmalli, described by Bartholomai in 1971, is the oldest species recovered to date. Its remains were found in Pliocene deposits near Chinchilla in southeastern Queensland. Known only from a lower jaw and some teeth, it was a relative of the spotted-tailed quoll. The first species described, the eastern quoll, was originally placed in the American opossum genus Didelphis by an anonymous author, and named Didelphis maculata. This name is no longer considered valid, and the second part of the name is now given to a different species, the spotted-tailed quoll, Dasyurus maculatus, while the eastern quoll was renamed Dasyurus viverrinus by George Shaw in 1800. The tribe Dasyurini, to which quolls belong, also includes the Tasmanian devil, the antechinus, the kowari, and the mulgara. Genetic analysis of cytochrome b DNA and 12S rRNA of the mitochondria indicates the quolls evolved and diversified in the late Miocene between 15 and 5 million years ago, a time of great diversification in marsupials. The ancestors of all current species had diverged by the early Pliocene, around 4 million years ago. Species The genus Dasyurus consists of six species of quoll: The following is a phylogenetic tree based on mitochondrial genome sequences: Description Quolls are solitary, nocturnal animals. Depending on the species, adult quolls can be long, with hairy tails about long. Average weight differs greatly depending on the species; male western and eastern quolls weigh about and females . The spotted-tailed quoll is the largest, with the male weighing about and the female . The northern quoll is the smallest, and the male weighs on average , and the female . Their coats are sandy, brown, or black, with a sparse scattering of white spots. They have bright pink noses and long snouts. Females have >8 teats and develop a pouch during the breeding season, which opens toward the tail (with the exception of the spotted-tailed quoll, which has a true pouch) when they are rearing young. Their natural lifespans are 1–5 years; the larger species tend to live longer. Distribution and habitat Quolls are indigenous to mainland Australia, the island state of Tasmania, and New Guinea. The six species were once widely distributed across the three land masses, but are now restricted to only a few areas. Although primarily ground-dwelling, the genus has developed secondary arboreal characteristics. Each species of quoll lives in distinct geographical areas. The spotted-tailed quoll is an exclusively mesic zone species; inhabiting wetter habitats. The western quoll also inhabits mesic habitat, but has adapted to arid regions across inland Australia, while the northern quoll inhabits tropical habitat of high rainfall. Behaviour Quolls are carnivorous marsupials. They are primarily nocturnal, sleeping in hollowed-out logs or rocky dens and coming out to hunt during the night, though on rare occasions they can be seen looking for prey during the day. They are mostly ground-dwelling, but it is not uncommon to see a quoll climbing a tree. Quolls mark their territory several kilometres away from their dens. A male's territory often overlaps many females' territories, and male and female quolls only meet for mating. Some quolls use communal latrines, usually on an outcropping used for marking territory and social functions, which may have up to 100 droppings in them. Quolls are mostly solitary, limiting contact with other quolls to mating or other social activities. Diet Quolls are mostly carnivorous. The smaller quolls primarily eat insects, birds, frogs, lizards and fruit; the larger species eat birds, reptiles, and mammals, including echidnas and possums. The spotted-tailed quoll's diet is dominated by mammals such as brushtail possums, rabbits, hares and invertebrates. The exact mix is variable depending on the availability of prey after bushfires, and can include carrion or bandicoots when food is scarce. The other species of quoll have also been known to eat carrion. Quolls hunt by stalking. Quolls pin small prey down with their front paws while devouring it, and jump onto larger prey, sinking in their claws and closing their jaws around the neck. The paws and vibrissae of quolls allow them to reach into small burrows to find prey. Quolls can obtain all the water they need from their food, making them adaptable during droughts or other periods of water shortage. A study of historical records revealed 111 written accounts of quolls opportunistically feeding on human remains in Australia. Reproduction Mating occurs during the winter months. Once a female quoll has been impregnated, the folds on her abdomen convert into a pouch that opens at the back. The gestation period is ~21 days (species dependent). A newborn quoll, or joey, is the size of a grain of rice at birth. Up to 30 quolls (species dependent) can be born in each litter, but the number that can be raised is limited by the number of teats. The survivors fuse to the teats and suckle milk in their mother's pouch for 6–8 weeks. After this, the pups unfuse from the teats and the mother can deposit them in a den where they can remain for over a month. Quolls reach maturity at one year old, and have a natural lifespan of 1–5 years (species dependent). The appearance of their pouches have been reported to be a reliable indicator of reproductive status: during the follicular phase pouches are red and have secretions, and after ovulation pouches are deep and wet. This can determine where a female quoll is in her ovarian cycle, which is anticipated to be helpful in breeding management. Threats Cane toads were introduced into Queensland in 1935; their numbers have since grown exponentially. These poisonous toads pose a significant threat to the northern quoll, which may die after consuming one. The Department of Sustainability, Environment, Water, Population and Communities has stated that cane toads are highly invasive and are major threats to the survival of northern quolls. Predators such as red foxes and feral cats prey on quolls and compete with them for food. For example, both quolls and foxes catch and consume rabbits. Since the introduction of foxes, quoll populations have declined dramatically. Foxes have been eradicated from many of the islands off the coast of Australia in an effort to protect quolls. Quoll habitat suffers from urbanisation, housing development, mining development, and agricultural expansion. Habitats are also being destroyed by large herbivores trampling the grass and overgrowth, making camouflage difficult. Bushfires and weeds also contribute to habitat destruction. The natural poison fluoroacetate (Compound 1080) is commonly used in Australia to control introduced pests such as European rabbits, foxes, feral predators, and dingoes. The poison is extremely toxic to introduced pests, but less so to native animals as it is found naturally in many Australian plants. However, juvenile quolls may be susceptible to the poison. research was being undertaken to determine whether the number of quolls protected from predators may be less than those killed by the poison. Conservation Since 1770, all Australian quoll species have declined due to habitat destruction through urbanisation. European rabbits were introduced to Australia with the arrival of the First Fleet in 1788 as part of biodiversity enrichment efforts. The native quolls predated upon rabbits and prior to 1870, many accounts recorded quolls impeding their establishment on the mainland while island colonies thrived. In response, quolls were systematically exterminated by colonists to defend introduced species such as chickens; rabbits populations subsequently reached plague proportions. Quolls have been studied in captivity, with the ultimate aim of supporting conservation of the species, and future translocations. These studies include investigations into their haematology and blood biochemistry profiles, and dietary studies. Creating a native pet industry in Australia related to quolls could aid in their conservation. However, concerns exist about this methodology in regards to animal husbandry, conservation benefits, and other issues. Some scientists believe that keeping quolls as pets could aid in their long-term conservation, but further research is needed. Spotted-tailed quoll In late October 2011, a litter of five spotted-tailed quoll pups was born at Wild Life Sydney in Darling Harbour, Australia. The pups were born to inexperienced parents, both just one year old. The reason for the young parents was because older male quolls can become violent and kill the female if they do not want to mate. By breeding one-year-old quolls, there was no threat of violence. Four of the quoll pups will be sent to other zoos or wildlife parks across Australia, but one, which the researchers named Nelson, will stay at the centre to become an "ambassador for all quolls". On 28 September 2023, it was reported that a farmer in Beachport, South Australia set up a trap to catch what he thought was a fox or a cat eating his chickens and caught a Spotted-tailed quoll. It is the first time in 130 years that a quoll has been found in South Australia. It was considered to be extinct in South Australia. The captured quoll was handed over to the National Parks and Wildlife Service where it will be DNA tested and treated by a veterinarian. Western quoll Fox control programs have benefited the western quoll. The Department of Environment and Conservation (Western Australia) monitors western quoll populations in the Jarrah Forest as part of its faunal management programs, as well as ongoing research into fox control, timber harvesting, and prescribed burning. The Perth Zoo has been monitoring a successful captive-breeding program since 1989. It has successfully bred more than 60 western quolls, most of which it transferred to Julimar Conservation Park, with proposals to translocate to Wheatbelt reserves and Shark Bay. Eastern quoll In 2003, the eastern quoll was reintroduced to a 473 ha fox-proof fenced sanctuary at Mt Rothwell Biodiversity Interpretation Centre at Mount Rothwell in Victoria. In 2016, the eastern quoll was also successfully reintroduced to Mulligans Flat Woodland Sanctuary in the Australian Capital Territory. Bristol Zoo was the first zoo in the UK to successfully breed eastern quolls. In March 2018, twenty eastern quolls bred in a wildlife park in Tasmania were released into the Booderee National Park on the south coast of NSW. In May 2021, the reintroduction of eastern quolls to Booderee National Park has been reported to have failed when numbers were down to one male. Northern quoll The northern quoll is threatened by toxic cane toads, but a University of Sydney project revealed in 2010 is teaching them to avoid eating the invasive amphibians. In 2008, the Northern Territory Wildlife Park in Australia recorded their first litter of northern quoll pups in the park. The quolls bred well in captivity, with over 15 litters in the 2008 breeding season alone. Bronze quoll The bronze quoll occurs in a few protected areas, such as Wasur National Park and Tonda Wildlife Management Area. More research on distribution and threats is needed for further conservation. Culture contexts Tjilpa is the name given to the quoll amongst the Northern Arrernte language group of Australian Aboriginal people.
Biology and health sciences
Marsupials
Animals
501921
https://en.wikipedia.org/wiki/Rinderpest
Rinderpest
Rinderpest (also cattle plague or steppe murrain) was an infectious viral disease of cattle, domestic water buffalo, and many other species of even-toed ungulates, including gaurs, buffaloes, large antelope, deer, giraffes, wildebeests, and warthogs. The disease was characterized by fever, oral erosions, diarrhea, lymphoid necrosis, and high mortality. Death rates during outbreaks were usually extremely high, approaching 100% in immunologically naïve populations. Rinderpest was mainly transmitted by direct contact and by drinking contaminated water, although it could also be transmitted by air. Rinderpest is believed to have originated in Asia, and to have spread by transport of cattle. The term Rinderpest () is a German word meaning 'cattle plague'. The rinderpest virus (RPV) is closely related to the measles and canine distemper viruses. The measles virus may have emerged from rinderpest as a zoonotic disease around 600 BC, a period that coincides with the rise of large human settlements. After a global eradication campaign that began in the mid-20th century, the last confirmed case of rinderpest was diagnosed in 2001. In 2010, the United Nations Food and Agriculture Organization (FAO) announced that field activities in the decades-long, worldwide campaign to eradicate the disease were ending, paving the way for a formal declaration in June 2011 of the global eradication of rinderpest. This makes it only the second disease in history to be fully wiped out, following smallpox. Virus Rinderpest virus (RPV), a member of the genus Morbillivirus, is closely related to the measles and canine distemper viruses. Like other members of the Paramyxoviridae family, it produces enveloped virions, and is a negative-sense single-stranded RNA virus. The virus is particularly fragile and is quickly inactivated by heat, desiccation, and sunlight. Measles virus evolved from the then-widespread rinderpest virus most probably between the 11th and 12th centuries. The earliest likely origin is during the seventh century; some linguistic evidence exists for this earlier origin. In 2020 research on the measles virus has suggested a modified understanding of the evolution of rinderpest. Work on preserved older samples of measles (1912 and following) have been tested in various ways to determine the likely trajectory of the measles virus' divergence from rinderpest. It is thought based on this study that the earliest date at which the divergence could have occurred is the sixth century BC. Disease and symptoms Death rates during outbreaks were usually extremely high, approaching 100% in immunologically naïve populations. The disease was mainly spread by direct contact and by drinking contaminated water, although it could also be transmitted by air. Initial symptoms include fever, loss of appetite, and nasal and eye discharges. Subsequently, irregular erosions appear in the mouth, the lining of the nose, and the genital tract. Acute diarrhea, preceded by constipation, is also a common feature. Most animals die six to twelve days after the onset of these clinical signs. The delayed appearance of these signs of illness account for the steady spread of the disease once a historical outbreak began: an animal infected by rinderpest undergoes an incubation period of 3–15 days. Signs of the disease only manifest at the end of that time. Cattle and wild ungulates will normally die 8–12 days after signs of the disease emerge, by which time the animals may have travelled far from the place of infection and been mixed with many other animals. History and epizootics Early history The disease is believed to have originated in Asia, later spreading through the transport of cattle. Other cattle epizootics are noted in ancient times: a cattle plague is thought to be one of the 10 plagues of Egypt described in the Hebrew Bible. By around 3,000 BC, a cattle plague had reached Egypt, and rinderpest later spread throughout the remainder of Africa, following European colonization. In the 4th century, Roman writer Severus Sanctus Endelechius described rinderpest in his book, On the Deaths of Cattle. 18th century Cattle plagues recurred throughout history, often accompanying wars and military campaigns. They hit Europe especially hard in the 18th century, with three long panzootics, which although varying in intensity and duration from region to region, took place in the periods of 1709–1720, 1742–1760, and 1768–1786. In the 18th century a deadly outbreak between 1769 and 1785 resulted in universal governmental action, but with somewhat divergent responses. The Dutch and the German principalities demanded quarantines and strict burial practices; England and large parts of Italy (the Papal States) saw slaughter of infected animals; in the Austrian Netherlands (Flanders) the response was inspection and precautionary slaughter coupled with compensation to the owners. There was no code of practice and no standard response. But for a hundred years thereafter in German-speaking countries there was intense focus on the problem of Rinderpest. Inoculation In the early 18th century, the disease was seen as similar to smallpox, due to its analogous symptoms. The personal physician of the pope, Giovanni Maria Lancisi, recommended the destruction of all infected and exposed animals. This policy was not very popular and was used only sparingly in the first part of the century. Later, it was used successfully in several countries, although it was sometimes seen as too costly or drastic, and depended on a strong central authority to be effective (which was notably lacking in the Dutch Republic). Because of these downsides, numerous attempts were made to inoculate animals against the disease. These attempts met with varying success, but the procedure was not widely used and was no longer practiced at all in 19th-century Western or Central Europe. Rinderpest was an immense problem, but inoculation was not a valid solution. In many cases, it caused too many losses. Even more importantly, it perpetuated the circulation of the virus in the cattle population. The pioneers of inoculation did contribute significantly to knowledge about infectious diseases. Their experiments confirmed the concepts of those who saw infectious diseases as caused by specific agents, and were the first to recognize maternally derived immunity. Early English experimentation The first written report of rinderpest inoculation was published in a letter signed "T.S." in the November 1754 issue of The Gentleman's Magazine, a widely read journal which also supported the progress of smallpox inoculation. This letter reported that a Mr Dobsen had inoculated his cattle and had thus preserved 9 out of 10 of them, although this was retracted in the next issue, as it was apparently a Sir William St. Quintin who had done the inoculating (this was done by placing bits of material previously dipped in morbid discharge into an incision made in the dewlap of the animal). These letters encouraged further application of inoculation in the fight against diseases. The first inoculation against measles was made three years after their publication. From early 1755 onwards, experiments were taking place in the Netherlands, as well, results of which were also published in The Gentleman's Magazine. As in England, the disease was seen as analogous with smallpox. While these experiments were reasonably successful, they did not have a significant impact: the total number of inoculations in England appears to have been very limited, and after 1780, the English interest in inoculation disappeared almost entirely. Almost all further experimentation was done in the Netherlands, northern Germany and Denmark. Further trials in the Netherlands Due to a very severe outbreak at the end of the 1760s, some of the best-known names in Dutch medicine became involved in the struggle against the disease. Several independent trials were begun, most notably by Pieter Camper in Groningen and Friesland. The results of his experiment in Friesland were encouraging, but they proved to be the exception; testing by others in the provinces of Utrecht and Friesland obtained disastrous results. As a result, the Frisian authorities concluded in 1769 that the cause of rinderpest was God's displeasure with the sinful behavior of the Frisian people and proclaimed 15 November a day of fasting and prayer. Interest in inoculation declined sharply across the country. In this climate of discouragement and scepticism, Geert Reinders, a farmer in the province of Groningen and a self-taught man, decided to continue the experiments. He collaborated with Wijnold Munniks, who had supervised earlier trials. They tried different inoculation procedures and a variety of treatments to lighten the symptoms, all of them without significant effect. Although they were not able to perfect the inoculation procedure, they did make some useful observations. Reinders resumed his experiments in 1774, concentrating on the inoculation of calves from cows that had recovered from rinderpest. He was probably the first to make practical use of maternally derived immunity. The detailed results of his trials were published in 1776 and reprinted in 1777. His inoculation procedure did not differ much from what had been used previously, except for the use of three separate inoculations at an early age. This produced far better results, and the publication of his work renewed interest in inoculation. For the period of 1777 to 1781, 89% of inoculated animals survived, compared to a 29% survival rate after natural infection. In the Netherlands, too, interest in rinderpest inoculation declined in the 1780s because the disease itself decreased in intensity. In other countries Apart from the Dutch Republic, the only other regions where inoculation was used to any significant level were northern Germany and Denmark. Experiments started in Mecklenburg during the epizootic of the late 1770s. "Insurance companies" were created which provided inoculation in special "institutes". Although these were private initiatives, they were created with full encouragement from the authorities. Though neighboring states followed this practice with interest, the practice never caught on outside Mecklenburg; many were still opposed to inoculation. While some experimentation occurred in other countries (most extensively in Denmark), in the majority of European countries, the struggle against the disease was based on stamping it out. Sometimes, this could be done with minimal sacrifices; at other times, it required slaughter at a massive scale. 19th century There were major outbreaks of cattle plague documented from the mid-century onwards. Responses to these outbreaks differed across the world. Rinderpest in 19th-century Europe A major outbreak affected the whole of the British Isles for three years after 1865. In August 1865 an Order of the British Privy Council required the slaughter of rinderpest-affected cattle. By early May 1867, the overall slaughter total was around 75,000 cattle, which at that time had a value of approximately £10 per head. Initially, £55,000 was granted (after a period of delay) to compensate farmers where they complied with the slaughter directive but had no other source of compensation. In certain areas, such as Aberdeenshire and Norfolk, farmers had banded together to provide mutual assurance by creating a resource pool against the risk of rinderpest. Because the initial slaughter regime was not backed by compensation, it was the presence of a voluntary mutual assurance scheme that drove down the infection rates by guaranteeing payment for compliance with the government instruction. The Privy Council ordered a detailed investigation of the disaster, which reported in 1868. In 1871, there was held an international Rinderpest convention in Vienna. It was purposed to establish mechanisms for reporting outbreaks to warn neighbouring countries, and so as to establish policies for inspections, quarantines and disinfections as well as monitoring the cattle trade. In 1879, there was a notable cattle plague outbreak in Congress Poland and parts of Prussia, resulting in the slaughter of animals. Impacted cities included Warsaw, Posen, and Sochaczew. Prussian authorities considered military border guards to help hinder the spread of the disease. Rinderpest in 19th-century Africa Around the turn of the century, a plague struck in Southern Africa. Spinage establishes a critical commentary on the theory that in 1888, rinderpest was introduced into Abyssinia (modern Ethiopia) by the invading Italian army, which supposedly brought with them infected cattle from India. The procurement chain is not traced beyond an Egyptian businessman from Cairo, but it is possible that the British Army got their draft oxen from India. However, the documentary chain only supported limited negative conclusions. "There is therefore no evidence in contemporary accounts that the rinderpest panzootic was imported from India with infected oxen to provision the Italian landing at Massawa." It may now be impossible to disentangle the probabilities of where rinderpest initially came from- invading Italians, invading Egyptians or local break-outs in Eritrea. Once in progress, the infection eventually spread to the shores of Lake Victoria and into German Tanzania. Sunseri concentrates on the detailed progress of the epizootic in German Tanzania, endeavouring to show that the disease was known to be present but was not officially recognised as being rinderpest. He emphasises in particular the failure by the German government to rely on or accept a post mortem in 1892 professionally medically conducted on an affected animal that had been duly diagnosed as having rinderpest. The diagnosis was procured at the personal behest of the governor and remitted to Berlin. It appears that awareness of a cattle plague in general did not amount to the German government accepting that the plague was rinderpest, for which measures of a strict kind were prescribed in Germany itself. The governor, Julius von Soden, personally lost his own herd, and this may have led him to secure the post-mortem so as to challenge the official diagnostic silence. The impact on African-owned herds was drastic. The disease was locally described as "sadoka" and it also affected local wildlife. Sunseri's thesis basically explains the German government's failure to recognise the true nature of the disease as permitting ineffective policies. The local German government was short of cash, without a vet until the late 1890s and surrounded by innumerable serious cattle diseases apart from rinderpest. The 1885 protectorate status of Tanzania (ruled by the German East Africa Company) had been interrupted by coastal rebellion: when formal German rule began and the military went inland in 1891 to pacify areas, they encountered massive cattle deaths ostensibly due to viral spread from wildlife (one assumes at waterholes). Some observers themselves described the outbreak as rinderpest, whereas argument and debate continued because of essentially lack of consistent information and detailed investigation. When the German governor requested confirmation as to a course of action, he would have been fully aware of the administrative consequences, had matters been dealt with in Germany (quarantines, slaughter policies, disinfection controls of cattle transport and control of products suspected of contact with contaminated animals). In the event, the post-mortem was reviewed in Berlin and determined to be incomplete: a diagnosis could only be made on the ground by a vet. Funding vets was not a priority as most of the cattle by then (1892) had died. Meanwhile, a German staff doctor with an interest in animal diseases opined (two long Reports for the German Colonial Service) that the problem must be an Africa-specific matter not the familiar rinderpest. His confusion may derive from the absence of impact of rinderpest on German wildlife. This is now explained by the fenced and manicured German agricultural landscape of the day being insufficiently "wild" and livestock normally being kept apart. By 1893, government regulatory response was as though the disease had been rinderpest in Germany (and included preventive slaughter). Cattle exports were banned in 1893 (to improve local stocks not on grounds of confining spread, as some cattle were exempt). Nevertheless importation, legal or illegal or rebranded via Zanzibar, reached the British colonies in the south. Marquardt concentrates on the detailed progress of the disease in South Africa during the 1896 outbreak. Between 1896 and 1897, 95% of the cattle in South Africa were killed by the disease. The primary spreading agency seems to be the common use of waterholes by wild ungulates and herded cattle. The herded cattle were normally in transit and the long incubation period and delayed symptoms meant that spreading had taken place before illness was realised. His initial case study is Southern Bechuanaland settled as it then was by two distinct cattle-focused groups: the Tswana people and the Boers. It was flat, hot and dry and was considered good cattle-raising country. Water was regularly available by drilling 20-30 feet below the surface, though many farms had water only by drilling 50-100 feet down. From 1882 onwards, designated Tswana reserves were created adjoining white farms in many instances. African pastoralism was constrained by this. From 1895, increasing numbers of white settlers (now administered from the Cape) evicted the Tswana and tension between these groups was inevitable. The 1896 drought resulted in fewer watering places being available, and a greater density of usage including both groups of cattle-owners and the wild animals. By May 1896, the vast Clober farm had become a focus of infection with immediate slaughter policies in place. Three river drinking places, mainly used by the Tswana group, recorded over 12,000 head of cattle regularly each; the government was reluctant to embark on wholesale destruction. The government tried, and failed, to stop herds crossing rivers and perpetuating stock-mingling. The spread of the disease was relentless in the Bechuanaland Protectorate. The connection between rinderpest and starvation was recognised by the British government as cause for urgent intervention by delivery of food relief. In 1896, 30,000 tons of mealies (corn) were delivered for the relief of the Bechuanaland Protectorate. Meanwhile, the Crocodile River in the Transvaal was reported as choked with cattle and other animal corpses, but remained in use. During the dry season, the government made no attempt to control use of the watering holes, fearing the consequences if they did. The Boers essentially did no better, mainly because they continued to migrate their cattle between parcels of land rather than remaining stationery within a particular parcel. Complaint by both Boer and Tswana groups was focused on the government rather than mutual hostility. Fencing, and quarantining coupled with killing of infected cattle, was a policy barely controllable in the expanses of the colony, though it had some success in England. However, fencing resulted in herd-mingling and consequent infection. The Tswana herds were quarantined together; the Boer herds were also quarantined but on their own land. The system was very unpopular. The policy was scorned and pilloried in the press: plenty of reports came out to the effect that the disease was spread by the quarantine guards and by the vets, all of whom were less than careful about disinfecting themselves. It is plausible that the major spreader of disease should be negligent government officials or contractors moving directly from areas known to be diseased to other areas in protective quarantine. In Southern Bechuanaland alone, over 400 men were hired as quarantine guards. Owners from both groups resisted the guards and the Boers vigorously resisted the killing of their cattle. It is likely both groups raised the fences, and several Boer groups deliberately spread the disease in order to claim the compensation. By 1896, it was generally recognised the government campaign had completely failed, overwhelmed by a storm of contributory causes to the spread of the disease. The outbreak in the 1890s killed an estimated 80–90% of all cattle in eastern and southern Africa. Sir Arnold Theiler was instrumental in developing a vaccine that curbed the epizootic. The consequences for the Africans were especially severe. Though cattle numbers revived subsequently, the consequent human toll was mass starvation in the absence of herding, hunting and farming. It is estimated that the human losses were as high as one-third of the population of Ethiopia and two-thirds of the Maasai people of Tanzania. This famine caused significant depopulation in sub-Saharan Africa, allowing thornbush to colonise. This formed ideal habitat for tsetse fly, which carries sleeping sickness, and is unsuitable for livestock; "hence the European view of an empty unspoiled Africa teeming with game". Rinderpest in 19th-century Asia Japan also sustained the presence of rinderpest in the 19th century as illustrated in an anonymous print. The disease was present for centuries in China, Japan and Korea. Japanese black and Korean yellow breed cattle were known to be especially susceptible to it. In 1868, there was a serious outbreak of rinderpest in India, which was investigated by Colonel James Hallen of the Indian Cattle Plague Commission leading to the publication of his survey in 1871. The Imperial Bacteriological Laboratory from 1893 was at Mukteshwar in India. It hosted much research work and many samples. Its founding director was British pathologist Alfred Lingard. In India, some farmers were reported as not hostile to tigers because of the consideration that their attacks on diseased or weaker animals reduced the risk of rinderpest. 20th century In his classic study of the Nuer of southern Sudan, E. E. Evans-Pritchard suggested rinderpest might have affected the Nuer's social organization before and during the 1930s. Since the Nuer were pastoralists, much of their livelihood was based on cattle husbandry, and bride-prices were paid in cattle; prices may have changed as a result of cattle depletion. Rinderpest might also have increased dependence on horticulture among the Nuer. Rinderpest was eradicated from Japan in 1922, as recorded by the Nippon Institute for Biological Science. Distinguished Japanese scientist and Director of the Nippon Institute for Biological Science, Junji Nakamura (1903–1975), was a major researcher into rinderpest, and the contribution of his work to the worldwide eradication of rinderpest was acknowledged by the Food and Agriculture Organisation of the United Nations. The FAO posthumously presented a certificate of appreciation in 2011. A more recent rinderpest outbreak in Africa in 1982–1984 resulted in an estimated US$2 billion in stock losses. Vaccination In 1917–18, William Hutchins Boynton (1881–1959), the chief veterinary pathologist with the Philippine Bureau of Agriculture, developed an early vaccine for rinderpest, based on treated animal organ extracts. In 1959, rinderpest vaccine was prepared at government laboratories in Abuko in The Gambia from the spleen of infected cattle. Walter Plowright worked on a vaccine for the RBOK strain of the rinderpest virus for multiple years, from 1956 to 1962. Plowright was awarded the World Food Prize in 1999 for developing a vaccine against a strain of rinderpest. In 1999, the FAO predicted that with vaccination, rinderpest would be eradicated by 2010. Eradication Widespread eradication efforts began in the early 20th century although, until the 1950s, they mostly took place on an individual country basis, using vaccination campaigns. In 1924, the World Organisation for Animal Health (OIE) was formed in response to rinderpest. In 1950, the Inter-African Bureau of Epizootic Diseases was formed, with the stated goal of eliminating rinderpest from Africa. With the loss of its wildebeest population, the Serengeti experienced radical fire regime shift to intense annual wildfires. During the 1960s, a program called JP 15 attempted to vaccinate all cattle in participating countries and, by 1979, only one of the countries involved, Sudan, reported cases of rinderpest. In the decades since, the wildebeest have returned to the Serengeti and tree cover has returned with them. In 1969, an outbreak of the disease originated in Afghanistan, travelling westwards and promoting a mass vaccination plan, which by 1972, had eliminated rinderpest in all areas of Asia except for Lebanon and India; both countries were the site of further occurrences of the disease in the 1980s. During the 1980s, however, an outbreak of rinderpest from Sudan spread throughout Africa, killing millions of cattle, as well as wildlife. In response, the Pan-African Rinderpest Campaign was initiated in 1987, using vaccination and surveillance to combat the disease. By the 1990s, nearly all of Africa, with the exception of parts of Sudan and Somalia, was declared free of rinderpest. Worldwide, the Global Rinderpest Eradication Programme was initiated in 1994, supported by the Food and Agriculture Organization, the OIE, and the International Atomic Energy Agency. This program was successful in reducing rinderpest outbreaks to few and far between by the late 1990s. The program is estimated to have saved affected farmers approximately 58 million net euros. The end was in sight by 2000 when only the Horn of Africa and Pakistan appeared to have a continued presence. Mariner et al., 2000 introduced participatory disease surveillance to rinderpest efforts. The last confirmed case of rinderpest was reported in Kenya in 2001. Since then, while no cases have been confirmed, the disease is believed to have been present in parts of Somalia past that date. The final vaccinations were administered in 2006, and the last surveillance operations took place in 2009, failing to find any evidence of the disease. The Mariner method continued to be used in those two locations (the Horn and Pakistan) to track down possible lingering refugia in the coming years. In 2008, scientists involved in rinderpest eradication efforts believed a good chance existed that rinderpest would join smallpox as officially "wiped off the face of the planet". The FAO, which had been co-ordinating the global eradication program for the disease, announced in November 2009 that it expected the disease to be eradicated within 18 months. In October 2010, the FAO announced it was confident the disease has been eradicated. The agency said that "[a]s of mid 2010, FAO is confident that the rinderpest virus has been eliminated from Europe, Asia, Middle East, Arabian Peninsula, and Africa," which were the locations where the virus had been last reported. Eradication was confirmed by the World Organization for Animal Health on 25 May 2011. On 28 June 2011, FAO and its members countries officially recognized global freedom from the deadly cattle virus. On this day, the FAO Conference, the highest body of the UN agency, adopted a resolution declaring the eradication of rinderpest. The resolution also called on the world community to follow up by ensuring that samples of rinderpest viruses and vaccines be kept under safe laboratory conditions and that rigorous standards for disease surveillance and reporting be applied. "While we are celebrating one of the greatest successes for FAO and its partners, I wish to remind you that this extraordinary achievement would not have been possible without the joint efforts and strong commitments of governments, the main organizations in Africa, Asia and Europe, and without the continuous support of donors and international institutions", FAO Director-General Jacques Diouf commented. The rinderpest eradication effort is estimated to have cost $5 billion. Stocks of the rinderpest virus are still maintained by highly specialized laboratories. In 2015, FAO launched a campaign calling for the destruction or sequestering of the remaining stocks of rinderpest virus in laboratories in 24 countries, citing risks of inadvertent or malicious release. On 14 June 2019, the largest stock of the rinderpest virus was destroyed at the Pirbright Institute. Use as a biological weapon Rinderpest was one of more than a dozen agents the United States government researched as potential biological weapons before terminating its biological weapons program. Rinderpest is of concern as a biological weapon for the following reasons: The disease has high rates of morbidity and mortality. The disease is highly communicable and spreads rapidly once introduced into nonimmune herds. Cattle herds are no longer immunized against RPV, so are susceptible to infection. Rinderpest was also considered as a biological weapon in a United Kingdom government programme during World War II.
Biology and health sciences
Viral diseases
Health
501973
https://en.wikipedia.org/wiki/Hemostasis
Hemostasis
In biology, hemostasis or haemostasis is a process to prevent and stop bleeding, meaning to keep blood within a damaged blood vessel (the opposite of hemostasis is hemorrhage). It is the first stage of wound healing. Hemostasis involves three major steps: vasoconstriction temporary blockage of a hole in a damaged blood vessel by a platelet plug blood coagulation (formation of fibrin clots) Coagulation, the changing of blood from a liquid to a gel which forms the fibrin clots, is essential to hemostasis. Intact blood vessels moderate blood's tendency to form clots. The endothelial cells of intact vessels prevent blood clotting with a heparin-like molecule and thrombomodulin, and prevent platelet aggregation with nitric oxide and prostacyclin. When endothelium of a blood vessel is damaged, the endothelial cells stop secretion of coagulation and aggregation inhibitors and instead secrete von Willebrand factor, which initiates the maintenance of hemostasis after injury. These processes seal the injury or hole until tissues are healed. Etymology and pronunciation The word hemostasis (, sometimes ) uses the combining forms and , Neo-Latin from Ancient Greek (similar to ), meaning "blood", and , meaning "stasis", yielding "motionlessness or stopping of blood". Steps of mechanism Hemostasis occurs when blood is present outside of the body or blood vessels. It is the innate response for the body to stop bleeding and loss of blood. During hemostasis three steps occur in a rapid sequence. Vascular spasm is the first response as the blood vessels constrict to allow less blood to be lost. In the second step, platelet plug formation, platelets stick together to form a temporary seal to cover the break in the vessel wall. The third and last step is called coagulation or blood clotting. Coagulation reinforces the platelet plug with fibrin threads that act as a "molecular glue". Platelets are a large factor in the hemostatic process. They allow for the creation of the "platelet plug" that forms almost directly after a blood vessel has been ruptured. Within seconds of a blood vessel's epithelial wall being disrupted, platelets begin to adhere to the sub-endothelium surface. It takes approximately sixty seconds until the first fibrin strands begin to intersperse among the wound. After several minutes the platelet plug is completely formed by fibrin. Hemostasis is maintained in the body via three mechanisms: Vascular spasm: Vasoconstriction is produced by vascular smooth muscle cells, and is the blood vessel's first response to injury. The smooth muscle cells are controlled by vascular endothelium, which releases intravascular signals to control the contracting properties. When a blood vessel is damaged, there is an immediate reflex, initiated by local sympathetic pain receptors, which helps promote vasoconstriction. The damaged vessels will constrict (vasoconstrict) which reduces the amount of blood flow through the area and limits the amount of blood loss. Collagen is exposed at the site of injury, the collagen promotes platelets to adhere to the injury site. Platelets release cytoplasmic granules which contain serotonin, ADP and thromboxane A2, all of which increase the effect of vasoconstriction. The spasm response becomes more effective as the amount of damage is increased. Vascular spasm is much more effective in smaller blood vessels. Platelet plug formation: Platelets adhere to damaged endothelium to form a platelet plug (primary hemostasis) and then degranulate. This process is regulated through thromboregulation. Plug formation is activated by a glycoprotein called von Willebrand factor (vWF), which is found in plasma. Platelets play one of major roles in the hemostatic process. When platelets come across the injured endothelium cells, they change shape, release granules and ultimately become ‘sticky’. Platelets express certain receptors, some of which are used for the adhesion of platelets to collagen. When platelets are activated, they express glycoprotein receptors that interact with other platelets, producing aggregation and adhesion. Platelets release cytoplasmic granules such as adenosine diphosphate (ADP), serotonin and thromboxane A2. Adenosine diphosphate (ADP) attracts more platelets to the affected area, serotonin is a vasoconstrictor and thromboxane A2 assists in platelet aggregation, vasoconstriction and degranulation. As more chemicals are released more platelets stick and release their chemicals; creating a platelet plug and continuing the process in a positive feedback loop. Platelets alone are responsible for stopping the bleeding of unnoticed wear and tear of our skin on a daily basis. This is referred to as primary hemostasis. Clot formation: Once the platelet plug has been formed by the platelets, the clotting factors (a dozen proteins that travel along the blood plasma in an inactive state) are activated in a sequence of events known as 'coagulation cascade' which leads to the formation of fibrin from inactive fibrinogen plasma protein. Thus, a fibrin mesh is produced all around the platelet plug to hold it in place; this step is called secondary hemostasis. During this process some red and white blood cells are trapped in the mesh which causes the primary hemostasis plug to become harder: the resultant plug is called a thrombus or blood clot. The blood clot contains the secondary hemostasis plug with blood cells trapped in it. This is a necessary step for wound healing, but it has the ability to cause severe health problems if the thrombus becomes detached from the vessel wall and travels through the circulatory system; If it reaches the brain, heart or lungs it could lead to stroke, heart attack, or pulmonary embolism respectively. Types Hemostasis can be achieved in various other ways if the body cannot do it naturally (or needs help) during surgery or medical treatment. When the body is under shock and stress, hemostasis is harder to achieve. Though natural hemostasis is most desired, having other means of achieving this is vital for survival in many emergency settings. Without the ability to stimulate hemostasis the risk of hemorrhaging is great. During surgical procedures, the types of hemostasis listed below can be used to control bleeding while avoiding and reducing the risk of tissue destruction. Hemostasis can be achieved by chemical agent as well as mechanical or physical agents. Which hemostasis type used is determined based on the situation. Developmental Haemostasis refers to the differences in the haemostatic system between children and adults. In emergency medicine Debates by physicians and medical practitioners still continue to arise on the subject of hemostasis and how to handle situations with large injuries. If an individual acquires a large injury resulting in extreme blood loss, then a hemostatic agent alone would not be very effective. Medical professionals continue to debate on what the best ways are to assist a patient in a chronic state; however, it is universally accepted that hemostatic agents are the primary tool for smaller bleeding injuries. Some main types of hemostasis used in emergency medicine include: Chemical/topical – This is a topical agent often used in surgery settings to stop bleeding. Microfibrillar collagen is the most popular choice among surgeons [recent source?] because it attracts the patient's natural platelets and starts the blood clotting process when it comes in contact with the platelets. This topical agent requires the normal hemostatic pathway to be properly functional. Direct pressure or pressure dressing – This type of hemostasis approach is most commonly used in situations where proper medical attention is not available. Putting pressure and/or dressing to a bleeding wound slows the process of blood loss, allowing for more time to get to an emergency medical setting. Soldiers use this skill during combat when someone has been injured because this process allows for blood loss to be decreased, giving the system time to start coagulation. Sutures and ties – Sutures are often used to close an open wound, allowing for the injured area to stay free of pathogens and other unwanted debris to enter the site; however, it is also essential to the process of hemostasis. Sutures and ties allow for skin to be joined back together allowing for platelets to start the process of hemostasis at a quicker pace. Using sutures results in a quicker recovery period because the surface area of the wound has been decreased. Physical agents (gelatin sponge) – Gelatin sponges have been indicated as great hemostatic devices. Once applied to a bleeding area, a gelatin sponge quickly stops or reduces the amount of bleeding present. These physical agents are mostly used in surgical settings as well as after surgery treatments. These sponges absorb blood, allow for coagulation to occur faster, and give off chemical responses that decrease the time it takes for the hemostasis pathway to start. Disorders The body's hemostasis system requires careful regulation in order to work properly. If the blood does not clot sufficiently, it may be due to bleeding disorders such as hemophilia or immune thrombocytopenia; this requires careful investigation. Over-active clotting can also cause problems; thrombosis, where blood clots form abnormally, can potentially cause embolisms, where blood clots break off and subsequently become lodged in a vein or artery. Hemostasis disorders can develop for many different reasons. They may be congenital, due to a deficiency or defect in an individual's platelets or clotting factors. A number of disorders can be acquired as well, such as in HELLP syndrome, which is due to pregnancy, or Hemolytic-uremic syndrome (HUS), which is due to E. coli toxins. History of artificial hemostasis The process of preventing blood loss from a vessel or organ of the body is referred to as hemostasis. The term comes from the Ancient Greek roots "heme" meaning blood, and "stasis" meaning halting; Put together means the "halting of the blood". The origin of hemostasis dates back as far as ancient Greece; first referenced to being used in the Battle of Troy. It started with the realization that excessive bleeding inevitably equaled death. Vegetable and mineral styptics were used on large wounds by the Greeks and Romans until the takeover of Egypt around 332BC by Greece. At this time many more advances in the general medical field were developed through the study of Egyptian mummification practice, which led to greater knowledge of the hemostatic process. It was during this time that many of the veins and arteries running throughout the human body were found and the directions in which they traveled. Doctors of this time realized if these were plugged, blood could not continue to flow out of the body. Nevertheless, it took until the invention of the printing press during the fifteenth century for medical notes and ideas to travel westward, allowing for the idea and practice of hemostasis to be expanded. Research There is currently a great deal of research being conducted on hemostasis. The most current research is based on genetic factors of hemostasis and how it can be altered to reduce the cause of genetic disorders that alter the natural process hemostasis. Von Willebrand disease is associated with a defect in the ability of the body to create the platelet plug and the fibrin mesh that ultimately stops the bleeding. New research is concluding that the von Willebrand disease is much more common in adolescence. This disease negatively hinders the natural process of Hemostasis causing excessive bleeding to be a concern in patients with this disease. There are complex treatments that can be done including a combination of therapies, estrogen-progesterone preparations, desmopressin, and Von Willebrand factor concentrates. Current research is trying to find better ways to deal with this disease; however, much more research is needed in order to find out the effectiveness of the current treatments and if there are more operative ways to treat this disease.
Biology and health sciences
Basics
Biology
19908550
https://en.wikipedia.org/wiki/Diffusion
Diffusion
Diffusion is the net movement of anything (for example, atoms, ions, molecules, energy) generally from a region of higher concentration to a region of lower concentration. Diffusion is driven by a gradient in Gibbs free energy or chemical potential. It is possible to diffuse "uphill" from a region of lower concentration to a region of higher concentration, as in spinodal decomposition. Diffusion is a stochastic process due to the inherent randomness of the diffusing entity and can be used to model many real-life stochastic scenarios. Therefore, diffusion and the corresponding mathematical models are used in several fields beyond physics, such as statistics, probability theory, information theory, neural networks, finance, and marketing. The concept of diffusion is widely used in many fields, including physics (particle diffusion), chemistry, biology, sociology, economics, statistics, data science, and finance (diffusion of people, ideas, data and price values). The central idea of diffusion, however, is common to all of these: a substance or collection undergoing diffusion spreads out from a point or location at which there is a higher concentration of that substance or collection. A gradient is the change in the value of a quantity; for example, concentration, pressure, or temperature with the change in another variable, usually distance. A change in concentration over a distance is called a concentration gradient, a change in pressure over a distance is called a pressure gradient, and a change in temperature over a distance is called a temperature gradient. The word diffusion derives from the Latin word, diffundere, which means "to spread out". A distinguishing feature of diffusion is that it depends on particle random walk, and results in mixing or mass transport without requiring directed bulk motion. Bulk motion, or bulk flow, is the characteristic of advection. The term convection is used to describe the combination of both transport phenomena. If a diffusion process can be described by Fick's laws, it is called a normal diffusion (or Fickian diffusion); Otherwise, it is called an anomalous diffusion (or non-Fickian diffusion). When talking about the extent of diffusion, two length scales are used in two different scenarios: Brownian motion of an impulsive point source (for example, one single spray of perfume)—the square root of the mean squared displacement from this point. In Fickian diffusion, this is , where is the dimension of this Brownian motion; Constant concentration source in one dimension—the diffusion length. In Fickian diffusion, this is . Diffusion vs. bulk flow "Bulk flow" is the movement/flow of an entire body due to a pressure gradient (for example, water coming out of a tap). "Diffusion" is the gradual movement/dispersion of concentration within a body with no net movement of matter. An example of a process where both bulk motion and diffusion occur is human breathing. First, there is a "bulk flow" process. The lungs are located in the thoracic cavity, which expands as the first step in external respiration. This expansion leads to an increase in volume of the alveoli in the lungs, which causes a decrease in pressure in the alveoli. This creates a pressure gradient between the air outside the body at relatively high pressure and the alveoli at relatively low pressure. The air moves down the pressure gradient through the airways of the lungs and into the alveoli until the pressure of the air and that in the alveoli are equal, that is, the movement of air by bulk flow stops once there is no longer a pressure gradient. Second, there is a "diffusion" process. The air arriving in the alveoli has a higher concentration of oxygen than the "stale" air in the alveoli. The increase in oxygen concentration creates a concentration gradient for oxygen between the air in the alveoli and the blood in the capillaries that surround the alveoli. Oxygen then moves by diffusion, down the concentration gradient, into the blood. The other consequence of the air arriving in alveoli is that the concentration of carbon dioxide in the alveoli decreases. This creates a concentration gradient for carbon dioxide to diffuse from the blood into the alveoli, as fresh air has a very low concentration of carbon dioxide compared to the blood in the body. Third, there is another "bulk flow" process. The pumping action of the heart then transports the blood around the body. As the left ventricle of the heart contracts, the volume decreases, which increases the pressure in the ventricle. This creates a pressure gradient between the heart and the capillaries, and blood moves through blood vessels by bulk flow down the pressure gradient. Diffusion in the context of different disciplines There are two ways to introduce the notion of diffusion: either a phenomenological approach starting with Fick's laws of diffusion and their mathematical consequences, or a physical and atomistic one, by considering the random walk of the diffusing particles. In the phenomenological approach, diffusion is the movement of a substance from a region of high concentration to a region of low concentration without bulk motion. According to Fick's laws, the diffusion flux is proportional to the negative gradient of concentrations. It goes from regions of higher concentration to regions of lower concentration. Sometime later, various generalizations of Fick's laws were developed in the frame of thermodynamics and non-equilibrium thermodynamics. From the atomistic point of view, diffusion is considered as a result of the random walk of the diffusing particles. In molecular diffusion, the moving molecules in a gas, liquid, or solid are self-propelled by kinetic energy. Random walk of small particles in suspension in a fluid was discovered in 1827 by Robert Brown, who found that minute particle suspended in a liquid medium and just large enough to be visible under an optical microscope exhibit a rapid and continually irregular motion of particles known as Brownian movement. The theory of the Brownian motion and the atomistic backgrounds of diffusion were developed by Albert Einstein. The concept of diffusion is typically applied to any subject matter involving random walks in ensembles of individuals. In chemistry and materials science, diffusion also refers to the movement of fluid molecules in porous solids. Different types of diffusion are distinguished in porous solids. Molecular diffusion occurs when the collision with another molecule is more likely than the collision with the pore walls. Under such conditions, the diffusivity is similar to that in a non-confined space and is proportional to the mean free path. Knudsen diffusion occurs when the pore diameter is comparable to or smaller than the mean free path of the molecule diffusing through the pore. Under this condition, the collision with the pore walls becomes gradually more likely and the diffusivity is lower. Finally there is configurational diffusion, which happens if the molecules have comparable size to that of the pore. Under this condition, the diffusivity is much lower compared to molecular diffusion and small differences in the kinetic diameter of the molecule cause large differences in diffusivity. Biologists often use the terms "net movement" or "net diffusion" to describe the movement of ions or molecules by diffusion. For example, oxygen can diffuse through cell membranes so long as there is a higher concentration of oxygen outside the cell. However, because the movement of molecules is random, occasionally oxygen molecules move out of the cell (against the concentration gradient). Because there are more oxygen molecules outside the cell, the probability that oxygen molecules will enter the cell is higher than the probability that oxygen molecules will leave the cell. Therefore, the "net" movement of oxygen molecules (the difference between the number of molecules either entering or leaving the cell) is into the cell. In other words, there is a net movement of oxygen molecules down the concentration gradient. In astronomy, atomic diffusion is used to model the stellar atmospheres of chemically peculiar stars. Diffusion of the elements is critical in understanding the surface composition of degenerate white dwarf stars and their evolution over time. History of diffusion in physics In the scope of time, diffusion in solids was used long before the theory of diffusion was created. For example, Pliny the Elder had previously described the cementation process, which produces steel from the element iron (Fe) through carbon diffusion. Another example is well known for many centuries, the diffusion of colors of stained glass or earthenware and Chinese ceramics. In modern science, the first systematic experimental study of diffusion was performed by Thomas Graham. He studied diffusion in gases, and the main phenomenon was described by him in 1831–1833: "...gases of different nature, when brought into contact, do not arrange themselves according to their density, the heaviest undermost, and the lighter uppermost, but they spontaneously diffuse, mutually and equally, through each other, and so remain in the intimate state of mixture for any length of time." The measurements of Graham contributed to James Clerk Maxwell deriving, in 1867, the coefficient of diffusion for CO2 in the air. The error rate is less than 5%. In 1855, Adolf Fick, the 26-year-old anatomy demonstrator from Zürich, proposed his law of diffusion. He used Graham's research, stating his goal as "the development of a fundamental law, for the operation of diffusion in a single element of space". He asserted a deep analogy between diffusion and conduction of heat or electricity, creating a formalism similar to Fourier's law for heat conduction (1822) and Ohm's law for electric current (1827). Robert Boyle demonstrated diffusion in solids in the 17th century by penetration of zinc into a copper coin. Nevertheless, diffusion in solids was not systematically studied until the second part of the 19th century. William Chandler Roberts-Austen, the well-known British metallurgist and former assistant of Thomas Graham studied systematically solid state diffusion on the example of gold in lead in 1896. : "... My long connection with Graham's researches made it almost a duty to attempt to extend his work on liquid diffusion to metals." In 1858, Rudolf Clausius introduced the concept of the mean free path. In the same year, James Clerk Maxwell developed the first atomistic theory of transport processes in gases. The modern atomistic theory of diffusion and Brownian motion was developed by Albert Einstein, Marian Smoluchowski and Jean-Baptiste Perrin. Ludwig Boltzmann, in the development of the atomistic backgrounds of the macroscopic transport processes, introduced the Boltzmann equation, which has served mathematics and physics with a source of transport process ideas and concerns for more than 140 years. In 1920–1921, George de Hevesy measured self-diffusion using radioisotopes. He studied self-diffusion of radioactive isotopes of lead in the liquid and solid lead. Yakov Frenkel (sometimes, Jakov/Jacob Frenkel) proposed, and elaborated in 1926, the idea of diffusion in crystals through local defects (vacancies and interstitial atoms). He concluded, the diffusion process in condensed matter is an ensemble of elementary jumps and quasichemical interactions of particles and defects. He introduced several mechanisms of diffusion and found rate constants from experimental data. Sometime later, Carl Wagner and Walter H. Schottky developed Frenkel's ideas about mechanisms of diffusion further. Presently, it is universally recognized that atomic defects are necessary to mediate diffusion in crystals. Henry Eyring, with co-authors, applied his theory of absolute reaction rates to Frenkel's quasichemical model of diffusion. The analogy between reaction kinetics and diffusion leads to various nonlinear versions of Fick's law. Basic models of diffusion Definition of diffusion flux Each model of diffusion expresses the diffusion flux with the use of concentrations, densities and their derivatives. Flux is a vector representing the quantity and direction of transfer. Given a small area with normal , the transfer of a physical quantity through the area per time is where is the inner product and is the little-o notation. If we use the notation of vector area then The dimension of the diffusion flux is [flux] = [quantity]/([time]·[area]). The diffusing physical quantity may be the number of particles, mass, energy, electric charge, or any other scalar extensive quantity. For its density, , the diffusion equation has the form where is intensity of any local source of this quantity (for example, the rate of a chemical reaction). For the diffusion equation, the no-flux boundary conditions can be formulated as on the boundary, where is the normal to the boundary at point . Normal single component concentration gradient Fick's first law: The diffusion flux, , is proportional to the negative gradient of spatial concentration, : where D is the diffusion coefficient. The corresponding diffusion equation (Fick's second law) is In case the diffusion coefficient is independent of , Fick's second law can be simplified to where is the Laplace operator, Multicomponent diffusion and thermodiffusion Fick's law describes diffusion of an admixture in a medium. The concentration of this admixture should be small and the gradient of this concentration should be also small. The driving force of diffusion in Fick's law is the antigradient of concentration, . In 1931, Lars Onsager included the multicomponent transport processes in the general context of linear non-equilibrium thermodynamics. For multi-component transport, where is the flux of the th physical quantity (component), is the th thermodynamic force and is Onsager's matrix of kinetic transport coefficients. The thermodynamic forces for the transport processes were introduced by Onsager as the space gradients of the derivatives of the entropy density (he used the term "force" in quotation marks or "driving force"): where are the "thermodynamic coordinates". For the heat and mass transfer one can take (the density of internal energy) and is the concentration of the th component. The corresponding driving forces are the space vectors because where T is the absolute temperature and is the chemical potential of the th component. It should be stressed that the separate diffusion equations describe the mixing or mass transport without bulk motion. Therefore, the terms with variation of the total pressure are neglected. It is possible for diffusion of small admixtures and for small gradients. For the linear Onsager equations, we must take the thermodynamic forces in the linear approximation near equilibrium: where the derivatives of are calculated at equilibrium . The matrix of the kinetic coefficients should be symmetric (Onsager reciprocal relations) and positive definite (for the entropy growth). The transport equations are Here, all the indexes are related to the internal energy (0) and various components. The expression in the square brackets is the matrix of the diffusion (i,k > 0), thermodiffusion (i > 0, k = 0 or k > 0, i = 0) and thermal conductivity () coefficients. Under isothermal conditions T = constant. The relevant thermodynamic potential is the free energy (or the free entropy). The thermodynamic driving forces for the isothermal diffusion are antigradients of chemical potentials, , and the matrix of diffusion coefficients is (i,k > 0). There is intrinsic arbitrariness in the definition of the thermodynamic forces and kinetic coefficients because they are not measurable separately and only their combinations can be measured. For example, in the original work of Onsager the thermodynamic forces include additional multiplier T, whereas in the Course of Theoretical Physics this multiplier is omitted but the sign of the thermodynamic forces is opposite. All these changes are supplemented by the corresponding changes in the coefficients and do not affect the measurable quantities. Nondiagonal diffusion must be nonlinear The formalism of linear irreversible thermodynamics (Onsager) generates the systems of linear diffusion equations in the form If the matrix of diffusion coefficients is diagonal, then this system of equations is just a collection of decoupled Fick's equations for various components. Assume that diffusion is non-diagonal, for example, , and consider the state with . At this state, . If at some points, then becomes negative at these points in a short time. Therefore, linear non-diagonal diffusion does not preserve positivity of concentrations. Non-diagonal equations of multicomponent diffusion must be non-linear. Applied forces The Einstein relation (kinetic theory) connects the diffusion coefficient and the mobility (the ratio of the particle's terminal drift velocity to an applied force). For charged particles: where D is the diffusion constant, μ is the "mobility", kB is the Boltzmann constant, T is the absolute temperature, and q is the elementary charge, that is, the charge of one electron. Below, to combine in the same formula the chemical potential μ and the mobility, we use for mobility the notation . Diffusion across a membrane The mobility-based approach was further applied by T. Teorell. In 1935, he studied the diffusion of ions through a membrane. He formulated the essence of his approach in the formula: the flux is equal to mobility × concentration × force per gram-ion. This is the so-called Teorell formula. The term "gram-ion" ("gram-particle") is used for a quantity of a substance that contains the Avogadro number of ions (particles). The common modern term is mole. The force under isothermal conditions consists of two parts: Diffusion force caused by concentration gradient: . Electrostatic force caused by electric potential gradient: . Here R is the gas constant, T is the absolute temperature, n is the concentration, the equilibrium concentration is marked by a superscript "eq", q is the charge and φ is the electric potential. The simple but crucial difference between the Teorell formula and the Onsager laws is the concentration factor in the Teorell expression for the flux. In the Einstein–Teorell approach, if for the finite force the concentration tends to zero then the flux also tends to zero, whereas the Onsager equations violate this simple and physically obvious rule. The general formulation of the Teorell formula for non-perfect systems under isothermal conditions is where μ is the chemical potential, μ0 is the standard value of the chemical potential. The expression is the so-called activity. It measures the "effective concentration" of a species in a non-ideal mixture. In this notation, the Teorell formula for the flux has a very simple form The standard derivation of the activity includes a normalization factor and for small concentrations , where is the standard concentration. Therefore, this formula for the flux describes the flux of the normalized dimensionless quantity : Ballistic time scale The Einstein model neglects the inertia of the diffusing partial. The alternative Langevin equation starts with Newton's second law of motion: where x is the position. μ is the mobility of the particle in the fluid or gas, which can be calculated using the Einstein relation (kinetic theory). m is the mass of the particle. F is the random force applied to the particle. t is time. Solving this equation, one obtained the time-dependent diffusion constant in the long-time limit and when the particle is significantly denser than the surrounding fluid, where kB is the Boltzmann constant; T is the absolute temperature. μ is the mobility of the particle in the fluid or gas, which can be calculated using the Einstein relation (kinetic theory). m is the mass of the particle. t is time. At long time scales, Einstein's result is recovered, but short time scales, the ballistic regime are also explained. Moreover, unlike the Einstein approach, a velocity can be defined, leading to the Fluctuation-dissipation theorem, connecting the competition between friction and random forces in defining the temperature. Jumps on the surface and in solids Diffusion of reagents on the surface of a catalyst may play an important role in heterogeneous catalysis. The model of diffusion in the ideal monolayer is based on the jumps of the reagents on the nearest free places. This model was used for CO on Pt oxidation under low gas pressure. The system includes several reagents on the surface. Their surface concentrations are The surface is a lattice of the adsorption places. Each reagent molecule fills a place on the surface. Some of the places are free. The concentration of the free places is . The sum of all (including free places) is constant, the density of adsorption places b. The jump model gives for the diffusion flux of (i = 1, ..., n): The corresponding diffusion equation is: Due to the conservation law, and we have the system of m diffusion equations. For one component we get Fick's law and linear equations because . For two and more components the equations are nonlinear. If all particles can exchange their positions with their closest neighbours then a simple generalization gives where is a symmetric matrix of coefficients that characterize the intensities of jumps. The free places (vacancies) should be considered as special "particles" with concentration . Various versions of these jump models are also suitable for simple diffusion mechanisms in solids. Porous media For diffusion in porous media the basic equations are (if Φ is constant): where D is the diffusion coefficient, Φ is porosity, n is the concentration, m > 0 (usually m > 1, the case m = 1 corresponds to Fick's law). Care must be taken to properly account for the porosity (Φ) of the porous medium in both the flux terms and the accumulation terms. For example, as the porosity goes to zero, the molar flux in the porous medium goes to zero for a given concentration gradient. Upon applying the divergence of the flux, the porosity terms cancel out and the second equation above is formed. For diffusion of gases in porous media this equation is the formalization of Darcy's law: the volumetric flux of a gas in the porous media is where k is the permeability of the medium, μ is the viscosity and p is the pressure. The advective molar flux is given as J = nq and for Darcy's law gives the equation of diffusion in porous media with m = γ + 1. In porous media, the average linear velocity (ν), is related to the volumetric flux as: Combining the advective molar flux with the diffusive flux gives the advection dispersion equation For underground water infiltration, the Boussinesq approximation gives the same equation with m = 2. For plasma with the high level of radiation, the Zeldovich–Raizer equation gives m > 4 for the heat transfer. Diffusion in physics Diffusion coefficient in kinetic theory of gases The diffusion coefficient is the coefficient in the Fick's first law , where J is the diffusion flux (amount of substance) per unit area per unit time, n (for ideal mixtures) is the concentration, x is the position [length]. Consider two gases with molecules of the same diameter d and mass m (self-diffusion). In this case, the elementary mean free path theory of diffusion gives for the diffusion coefficient where kB is the Boltzmann constant, T is the temperature, P is the pressure, is the mean free path, and vT is the mean thermal speed: We can see that the diffusion coefficient in the mean free path approximation grows with T as T3/2 and decreases with P as 1/P. If we use for P the ideal gas law P = RnT with the total concentration n, then we can see that for given concentration n the diffusion coefficient grows with T as T1/2 and for given temperature it decreases with the total concentration as 1/n. For two different gases, A and B, with molecular masses mA, mB and molecular diameters dA, dB, the mean free path estimate of the diffusion coefficient of A in B and B in A is: The theory of diffusion in gases based on Boltzmann's equation In Boltzmann's kinetics of the mixture of gases, each gas has its own distribution function, , where t is the time moment, x is position and c is velocity of molecule of the ith component of the mixture. Each component has its mean velocity . If the velocities do not coincide then there exists diffusion. In the Chapman–Enskog approximation, all the distribution functions are expressed through the densities of the conserved quantities: individual concentrations of particles, (particles per volume), density of momentum (mi is the ith particle mass), density of kinetic energy The kinetic temperature T and pressure P are defined in 3D space as where is the total density. For two gases, the difference between velocities, is given by the expression: where is the force applied to the molecules of the ith component and is the thermodiffusion ratio. The coefficient D12 is positive. This is the diffusion coefficient. Four terms in the formula for C1−C2 describe four main effects in the diffusion of gases: describes the flux of the first component from the areas with the high ratio n1/n to the areas with lower values of this ratio (and, analogously the flux of the second component from high n2/n to low n2/n because n2/n = 1 – n1/n); describes the flux of the heavier molecules to the areas with higher pressure and the lighter molecules to the areas with lower pressure, this is barodiffusion; describes diffusion caused by the difference of the forces applied to molecules of different types. For example, in the Earth's gravitational field, the heavier molecules should go down, or in electric field the charged molecules should move, until this effect is not equilibrated by the sum of other terms. This effect should not be confused with barodiffusion caused by the pressure gradient. describes thermodiffusion, the diffusion flux caused by the temperature gradient. All these effects are called diffusion because they describe the differences between velocities of different components in the mixture. Therefore, these effects cannot be described as a bulk transport and differ from advection or convection. In the first approximation, for rigid spheres; for repulsing force The number is defined by quadratures (formulas (3.7), (3.9), Ch. 10 of the classical Chapman and Cowling book) We can see that the dependence on T for the rigid spheres is the same as for the simple mean free path theory but for the power repulsion laws the exponent is different. Dependence on a total concentration n for a given temperature has always the same character, 1/n. In applications to gas dynamics, the diffusion flux and the bulk flow should be joined in one system of transport equations. The bulk flow describes the mass transfer. Its velocity V is the mass average velocity. It is defined through the momentum density and the mass concentrations: where is the mass concentration of the ith species, is the mass density. By definition, the diffusion velocity of the ith component is , . The mass transfer of the ith component is described by the continuity equation where is the net mass production rate in chemical reactions, . In these equations, the term describes advection of the ith component and the term represents diffusion of this component. In 1948, Wendell H. Furry proposed to use the form of the diffusion rates found in kinetic theory as a framework for the new phenomenological approach to diffusion in gases. This approach was developed further by F.A. Williams and S.H. Lam. For the diffusion velocities in multicomponent gases (N components) they used Here, is the diffusion coefficient matrix, is the thermal diffusion coefficient, is the body force per unit mass acting on the ith species, is the partial pressure fraction of the ith species (and is the partial pressure), is the mass fraction of the ith species, and Diffusion of electrons in solids When the density of electrons in solids is not in equilibrium, diffusion of electrons occurs. For example, when a bias is applied to two ends of a chunk of semiconductor, or a light shines on one end (see right figure), electrons diffuse from high density regions (center) to low density regions (two ends), forming a gradient of electron density. This process generates current, referred to as diffusion current. Diffusion current can also be described by Fick's first law where J is the diffusion current density (amount of substance) per unit area per unit time, n (for ideal mixtures) is the electron density, x is the position [length]. Diffusion in geophysics Analytical and numerical models that solve the diffusion equation for different initial and boundary conditions have been popular for studying a wide variety of changes to the Earth's surface. Diffusion has been used extensively in erosion studies of hillslope retreat, bluff erosion, fault scarp degradation, wave-cut terrace/shoreline retreat, alluvial channel incision, coastal shelf retreat, and delta progradation. Although the Earth's surface is not literally diffusing in many of these cases, the process of diffusion effectively mimics the holistic changes that occur over decades to millennia. Diffusion models may also be used to solve inverse boundary value problems in which some information about the depositional environment is known from paleoenvironmental reconstruction and the diffusion equation is used to figure out the sediment influx and time series of landform changes. Dialysis Dialysis works on the principles of the diffusion of solutes and ultrafiltration of fluid across a semi-permeable membrane. Diffusion is a property of substances in water; substances in water tend to move from an area of high concentration to an area of low concentration. Blood flows by one side of a semi-permeable membrane, and a dialysate, or special dialysis fluid, flows by the opposite side. A semipermeable membrane is a thin layer of material that contains holes of various sizes, or pores. Smaller solutes and fluid pass through the membrane, but the membrane blocks the passage of larger substances (for example, red blood cells and large proteins). This replicates the filtering process that takes place in the kidneys when the blood enters the kidneys and the larger substances are separated from the smaller ones in the glomerulus. Random walk (random motion) One common misconception is that individual atoms, ions or molecules move randomly, which they do not. In the animation on the right, the ion in the left panel appears to have "random" motion in the absence of other ions. As the right panel shows, however, this motion is not random but is the result of "collisions" with other ions. As such, the movement of a single atom, ion, or molecule within a mixture just appears random when viewed in isolation. The movement of a substance within a mixture by "random walk" is governed by the kinetic energy within the system that can be affected by changes in concentration, pressure or temperature. (This is a classical description. At smaller scales, quantum effects will be non-negligible, in general. Thus, the study of the movement of a single atom becomes more subtle since particles at such small scales are described by probability amplitudes rather than deterministic measures of position and velocity.) Separation of diffusion from convection in gases While Brownian motion of multi-molecular mesoscopic particles (like pollen grains studied by Brown) is observable under an optical microscope, molecular diffusion can only be probed in carefully controlled experimental conditions. Since Graham experiments, it is well known that avoiding of convection is necessary and this may be a non-trivial task. Under normal conditions, molecular diffusion dominates only at lengths in the nanometre-to-millimetre range. On larger length scales, transport in liquids and gases is normally due to another transport phenomenon, convection. To separate diffusion in these cases, special efforts are needed. In contrast, heat conduction through solid media is an everyday occurrence (for example, a metal spoon partly immersed in a hot liquid). This explains why the diffusion of heat was explained mathematically before the diffusion of mass. Other types of diffusion Anisotropic diffusion, also known as the Perona–Malik equation, enhances high gradients Atomic diffusion, in solids Bohm diffusion, spread of plasma across magnetic fields Eddy diffusion, in coarse-grained description of turbulent flow Effusion of a gas through small holes Electronic diffusion, resulting in an electric current called the diffusion current Facilitated diffusion, present in some organisms Gaseous diffusion, used for isotope separation Heat equation, diffusion of thermal energy Itō diffusion, mathematisation of Brownian motion, continuous stochastic process. Knudsen diffusion of gas in long pores with frequent wall collisions Lévy flight Molecular diffusion, diffusion of molecules from more dense to less dense areas Momentum diffusion ex. the diffusion of the hydrodynamic velocity field Photon diffusion Plasma diffusion Random walk, model for diffusion Reverse diffusion, against the concentration gradient, in phase separation Rotational diffusion, random reorientation of molecules Spin diffusion, diffusion of spin magnetic moments in solids Surface diffusion, diffusion of adparticles on a surface Taxis is an animal's directional movement activity in response to a stimulus Kinesis is an animal's non-directional movement activity in response to a stimulus Trans-cultural diffusion, diffusion of cultural traits across geographical area Turbulent diffusion, transport of mass, heat, or momentum within a turbulent fluid
Physical sciences
Fluid mechanics
null
8404649
https://en.wikipedia.org/wiki/Pressure%E2%80%93volume%20diagram
Pressure–volume diagram
A pressure–volume diagram (or PV diagram, or volume–pressure loop) is used to describe corresponding changes in volume and pressure in a system. They are commonly used in thermodynamics, cardiovascular physiology, and respiratory physiology. PV diagrams, originally called indicator diagrams, were developed in the 18th century as tools for understanding the efficiency of steam engines. Description A PV diagram plots the change in pressure P with respect to volume V for some process or processes. Typically in thermodynamics, the set of processes forms a cycle, so that upon completion of the cycle there has been no net change in state of the system; i.e. the device returns to the starting pressure and volume. The figure shows the features of an idealized PV diagram. It shows a series of numbered states (1 through 4). The path between each state consists of some process (A through D) which alters the pressure or volume of the system (or both). A key feature of the diagram is that the amount of energy expended or received by the system as work can be measured because the net work is represented by the area enclosed by the four lines. In the figure, the processes 1-2-3 produce a work output, but processes from 3-4-1 require a smaller energy input to return to the starting position / state; so the net work is the difference between the two. This figure is highly idealized, in so far as all the lines are straight and the corners are right angles. A diagram showing the changes in pressure and volume in a real device will show a more complex shape enclosing the work cycle. (). History The PV diagram, then called an indicator diagram, was developed in 1796 by James Watt and his employee John Southern. Volume was traced by a plate moving with the piston, while pressure was traced by a pressure gauge whose indicator moved at right angles to the piston. A pencil was used to draw the diagram. Watt used the diagram to make radical improvements to steam engine performance. Applications Thermodynamics Specifically, the diagram records the pressure of steam versus the volume of steam in a cylinder, throughout a piston's cycle of motion in a steam engine. The diagram enables calculation of the work performed and thus can provide a measure of the power produced by the engine. To exactly calculate the work done by the system it is necessary to calculate the integral of the pressure with respect to volume. One can often quickly calculate this using the PV diagram as it is simply the area enclosed by the cycle. Note that in some cases specific volume will be plotted on the x-axis instead of volume, in which case the area under the curve represents work per unit mass of the working fluid (i.e. J/kg). Medicine In cardiovascular physiology, the diagram is often applied to the left ventricle, and it can be mapped to specific events of the cardiac cycle. PV loop studies are widely used in basic research and preclinical testing, to characterize the intact heart's performance under various situations (effect of drugs, disease, characterization of mouse strains) The sequence of events occurring in every heart cycle is as follows. The left figure shows a PV loop from a real experiment; letters refer to points. A is the end-diastolic point; this is the point where contraction begins. Pressure starts to increase, becomes rapidly higher than the atrial pressure, and the mitral valve closes. Since pressure is also lower than the aortic pressure, the aortic valve is closed as well. Segment AB is the contraction phase. Since both the mitral and aortic valves are closed, volume is constant. For this reason, this phase is called isovolumic contraction. At point B, pressure becomes higher than the aortic pressure and the aortic valve opens, initiating ejection. BC is the ejection phase, volume decreases. At the end of this phase, pressure lowers again and falls below aortic pressure. The aortic valve closes. Point C is the end-systolic point. Segment CD is the isovolumic relaxation. During this phase, pressure continues to fall. The mitral valve and aortic valve are both closed again so volume is constant. At point D pressure falls below the atrial pressure and the mitral valve opens, initiating ventricular filling. DA is the diastolic filling period. Blood flows from the left atrium to the left ventricle. Atrial contraction completes ventricular filling. As it can be seen, the PV loop forms a roughly rectangular shape and each loop is formed in an anti-clockwise direction. Very useful information can be derived by examination and analysis of individual loops or series of loops, for example: the horizontal distance between the top-left corner and the bottom-right corner of each loop is the stroke volume the line joining the top-left corner of several loops is the contractile or inotropic state. See external links for a much more precise representation.
Physical sciences
Thermodynamics
Physics
8406655
https://en.wikipedia.org/wiki/Introduction%20to%20genetics
Introduction to genetics
Genetics is the study of genes and tries to explain what they are and how they work. Genes are how living organisms inherit features or traits from their ancestors; for example, children usually look like their parents because they have inherited their parents' genes. Genetics tries to identify which traits are inherited and to explain how these traits are passed from generation to generation. Some traits are part of an organism's physical appearance, such as eye color or height. Other sorts of traits are not easily seen and include blood types or resistance to diseases. Some traits are inherited through genes, which is the reason why tall and thin people tend to have tall and thin children. Other traits come from interactions between genes and the environment, so a child who inherited the tendency of being tall will still be short if poorly nourished. The way our genes and environment interact to produce a trait can be complicated. For example, the chances of somebody dying of cancer or heart disease seems to depend on both their genes and their lifestyle. Genes are made from a long molecule called DNA, which is copied and inherited across generations. DNA is made of simple units that line up in a particular order within it, carrying genetic information. The language used by DNA is called genetic code, which lets organisms read the information in the genes. This information is the instructions for the construction and operation of a living organism. The information within a particular gene is not always exactly the same between one organism and another, so different copies of a gene do not always give exactly the same instructions. Each unique form of a single gene is called an allele. As an example, one allele for the gene for hair color could instruct the body to produce much pigment, producing black hair, while a different allele of the same gene might give garbled instructions that fail to produce any pigment, giving white hair. Mutations are random changes in genes and can create new alleles. Mutations can also produce new traits, such as when mutations to an allele for black hair produce a new allele for white hair. This appearance of new traits is important in evolution. Genes and inheritance Genes are pieces of DNA that contain information for the synthesis of ribonucleic acids (RNAs) or polypeptides. Genes are inherited as units, with two parents dividing out copies of their genes to their offspring. Humans have two copies of each of their genes, but each egg or sperm cell only gets one of those copies for each gene. An egg and sperm join to form a zygote with a complete set of genes. The resulting offspring has the same number of genes as their parents, but for any gene, one of their two copies comes from their father and one from their mother. Example of mixing The effects of mixing depend on the types (the alleles) of the gene. If the father has two copies of an allele for red hair, and the mother has two copies for brown hair, all their children get the two alleles that give different instructions, one for red hair and one for brown. The hair color of these children depends on how these alleles work together. If one allele dominates the instructions from another, it is called the dominant allele, and the allele that is overridden is called the recessive allele. In the case of a daughter with alleles for both red and brown hair, brown is dominant and she ends up with brown hair. Although the red color allele is still there in this brown-haired girl, it doesn't show. This is a difference between what is seen on the surface (the traits of an organism, called its phenotype) and the genes within the organism (its genotype). In this example, the allele for brown can be called "B" and the allele for red "b". (It is normal to write dominant alleles with capital letters and recessive ones with lower-case letters.) The brown hair daughter has the "brown hair phenotype" but her genotype is Bb, with one copy of the B allele, and one of the b allele. Now imagine that this woman grows up and has children with a brown-haired man who also has a Bb genotype. Her eggs will be a mixture of two types, one sort containing the B allele, and one sort the b allele. Similarly, her partner will produce a mix of two types of sperm containing one or the other of these two alleles. When the transmitted genes are joined up in their offspring, these children have a chance of getting either brown or red hair, since they could get a genotype of BB = brown hair, Bb = brown hair or bb = red hair. In this generation, there is, therefore, a chance of the recessive allele showing itself in the phenotype of the children—some of them may have red hair like their grandfather. Many traits are inherited in a more complicated way than the example above. This can happen when there are several genes involved, each contributing a small part to the result. Tall people tend to have tall children because their children get a package of many alleles that each contribute a bit to how much they grow. However, there are not clear groups of "short people" and "tall people", like there are groups of people with brown or red hair. This is because of the large number of genes involved; this makes the trait very variable and people are of many different heights. Despite a common misconception, the green/blue eye traits are also inherited in this complex inheritance model. Inheritance can also be complicated when the trait depends on the interaction between genetics and environment. For example, malnutrition does not change traits like eye color, but can stunt growth. How genes work Genes make proteins The function of genes is to provide the information needed to make molecules called proteins in cells. Cells are the smallest independent parts of organisms: the human body contains about 100 trillion cells, while very small organisms like bacteria are just a single cell. A cell is like a miniature and very complex factory that can make all the parts needed to produce a copy of itself, which happens when cells divide. There is a simple division of labor in cells—genes give instructions and proteins carry out these instructions, tasks like building a new copy of a cell, or repairing the damage. Each type of protein is a specialist that only does one job, so if a cell needs to do something new, it must make a new protein to do this job. Similarly, if a cell needs to do something faster or slower than before, it makes more or less of the protein responsible. Genes tell cells what to do by telling them which proteins to make and in what amounts. Proteins are made of a chain of 20 different types of amino acid molecules. This chain folds up into a compact shape, rather like an untidy ball of string. The shape of the protein is determined by the sequence of amino acids along its chain and it is this shape that, in turn, determines what the protein does. For example, some proteins have parts of their surface that perfectly match the shape of another molecule, allowing the protein to bind to this molecule very tightly. Other proteins are enzymes, which are like tiny machines that alter other molecules. The information in DNA is held in the sequence of the repeating units along the DNA chain. These units are four types of nucleotides (A, T, G and C) and the sequence of nucleotides stores information in an alphabet called the genetic code. When a gene is read by a cell the DNA sequence is copied into a very similar molecule called RNA (this process is called transcription). Transcription is controlled by other DNA sequences (such as promoters), which show a cell where genes are, and control how often they are copied. The RNA copy made from a gene is then fed through a structure called a ribosome, which translates the sequence of nucleotides in the RNA into the correct sequence of amino acids and joins these amino acids together to make a complete protein chain. The new protein then folds up into its active form. The process of moving information from the language of RNA into the language of amino acids is called translation. If the sequence of the nucleotides in a gene changes, the sequence of the amino acids in the protein it produces may also change—if part of a gene is deleted, the protein produced is shorter and may not work anymore. This is the reason why different alleles of a gene can have different effects on an organism. As an example, hair color depends on how much of a dark substance called melanin is put into the hair as it grows. If a person has a normal set of the genes involved in making melanin, they make all the proteins needed and they grow dark hair. However, if the alleles for a particular protein have different sequences and produce proteins that can't do their jobs, no melanin is produced and the person has white skin and hair (albinism). Genes are copied Genes are copied each time a cell divides into two new cells. The process that copies DNA is called DNA replication. It is through a similar process that a child inherits genes from its parents when a copy from the mother is mixed with a copy from the father. DNA can be copied very easily and accurately because each piece of DNA can direct the assembly of a new copy of its information. This is because DNA is made of two strands that pair together like the two sides of a zipper. The nucleotides are in the center, like the teeth in the zipper, and pair up to hold the two strands together. Importantly, the four different sorts of nucleotides are different shapes, so for the strands to close up properly, an A nucleotide must go opposite a T nucleotide, and a G opposite a C. This exact pairing is called base pairing. When DNA is copied, the two strands of the old DNA are pulled apart by enzymes; then they pair up with new nucleotides and then close. This produces two new pieces of DNA, each containing one strand from the old DNA and one newly made strand. This process is not predictably perfect as proteins attach to a nucleotide while they are building and cause a change in the sequence of that gene. These changes in the DNA sequence are called mutations. Mutations produce new alleles of genes. Sometimes these changes stop the functioning of that gene or make it serve another advantageous function, such as the melanin genes discussed above. These mutations and their effects on the traits of organisms are one of the causes of evolution. Genes and evolution A population of organisms evolves when an inherited trait becomes more common or less common over time. For instance, all the mice living on an island would be a single population of mice: some with white fur, some gray. If over generations, white mice became more frequent and gray mice less frequent, then the color of the fur in this population of mice would be evolving. In terms of genetics, this is called an increase in allele frequency. Alleles become more or less common either by chance in a process called genetic drift or by natural selection. In natural selection, if an allele makes it more likely for an organism to survive and reproduce, then over time this allele becomes more common. But if an allele is harmful, natural selection makes it less common. In the above example, if the island were getting colder each year and snow became present for much of the time, then the allele for white fur would favor survival since predators would be less likely to see them against the snow, and more likely to see the gray mice. Over time white mice would become more and more frequent, while gray mice less and less. Mutations create new alleles. These alleles have new DNA sequences and can produce proteins with new properties. So if an island was populated entirely by black mice, mutations could happen creating alleles for white fur. The combination of mutations creating new alleles at random, and natural selection picking out those that are useful, causes an adaptation. This is when organisms change in ways that help them to survive and reproduce. Many such changes, studied in evolutionary developmental biology, affect the way the embryo develops into an adult body. Inherited diseases Some diseases are hereditary and run in families; others, such as infectious diseases, are caused by the environment. Other diseases come from a combination of genes and the environment. Genetic disorders are diseases that are caused by a single allele of a gene and are inherited in families. These include Huntington's disease, cystic fibrosis or Duchenne muscular dystrophy. Cystic fibrosis, for example, is caused by mutations in a single gene called CFTR and is inherited as a recessive trait. Other diseases are influenced by genetics, but the genes a person gets from their parents only change their risk of getting a disease. Most of these diseases are inherited in a complex way, with either multiple genes involved, or coming from both genes and the environment. As an example, the risk of breast cancer is 50 times higher in the families most at risk, compared to the families least at risk. This variation is probably due to a large number of alleles, each changing the risk a little bit. Several of the genes have been identified, such as BRCA1 and BRCA2, but not all of them. However, although some of the risks are genetic, the risk of this cancer is also increased by being overweight, heavy alcohol consumption and not exercising. A woman's risk of breast cancer, therefore, comes from a large number of alleles interacting with her environment, so it is very hard to predict. Genetic engineering Since traits come from the genes in a cell, putting a new piece of DNA into a cell can produce a new trait. This is how genetic engineering works. For example, rice can be given genes from a maize and a soil bacteria so the rice produces beta-carotene, which the body converts to vitamin A. This can help children with Vitamin A deficiency. Another gene being put into some crops comes from the bacterium Bacillus thuringiensis; the gene makes a protein that is an insecticide. The insecticide kills insects that eat the plants but is harmless to people. In these plants, the new genes are put into the plant before it is grown, so the genes are in every part of the plant, including its seeds. The plant's offspring inherit the new genes, which has led to concern about the spread of new traits into wild plants. The kind of technology used in genetic engineering is also being developed to treat people with genetic disorders in an experimental medical technique called gene therapy. However, here the new, properly working gene is put in targeted cells, not altering the chance of future children inheriting the disease causing alleles.
Biology and health sciences
Genetics
Biology
8412074
https://en.wikipedia.org/wiki/Acid%E2%80%93base%20homeostasis
Acid–base homeostasis
Acid–base homeostasis is the homeostatic regulation of the pH of the body's extracellular fluid (ECF). The proper balance between the acids and bases (i.e. the pH) in the ECF is crucial for the normal physiology of the body—and for cellular metabolism. The pH of the intracellular fluid and the extracellular fluid need to be maintained at a constant level. The three dimensional structures of many extracellular proteins, such as the plasma proteins and membrane proteins of the body's cells, are very sensitive to the extracellular pH. Stringent mechanisms therefore exist to maintain the pH within very narrow limits. Outside the acceptable range of pH, proteins are denatured (i.e. their 3D structure is disrupted), causing enzymes and ion channels (among others) to malfunction. An acid–base imbalance is known as acidemia when the pH is acidic, or alkalemia when the pH is alkaline. Lines of defense In humans and many other animals, acid–base homeostasis is maintained by multiple mechanisms involved in three lines of defense: Chemical: The first lines of defense are immediate, consisting of the various chemical buffers which minimize pH changes that would otherwise occur in their absence. These buffers include the bicarbonate buffer system, the phosphate buffer system, and the protein buffer system. Respiratory component: The second line of defense is rapid consisting of the control the carbonic acid (H2CO3) concentration in the ECF by changing the rate and depth of breathing by hyperventilation or hypoventilation. This blows off or retains carbon dioxide (and thus carbonic acid) in the blood plasma as required. Metabolic component: The third line of defense is slow, best measured by the base excess, and mostly depends on the renal system which can add or remove bicarbonate ions () to or from the ECF. Bicarbonate ions are derived from metabolic carbon dioxide which is enzymatically converted to carbonic acid in the renal tubular cells. There, carbonic acid spontaneously dissociates into hydrogen ions and bicarbonate ions. When the pH in the ECF falls, hydrogen ions are excreted into urine, while bicarbonate ions are secreted into blood plasma, causing the plasma pH to rise. The converse happens if the pH in the ECF tends to rise: bicarbonate ions are then excreted into the urine and hydrogen ions into the blood plasma. The second and third lines of defense operate by making changes to the buffers, each of which consists of two components: a weak acid and its conjugate base. It is the ratio concentration of the weak acid to its conjugate base that determines the pH of the solution. Thus, by manipulating firstly the concentration of the weak acid, and secondly that of its conjugate base, the pH of the extracellular fluid (ECF) can be adjusted very accurately to the correct value. The bicarbonate buffer, consisting of a mixture of carbonic acid (H2CO3) and a bicarbonate () salt in solution, is the most abundant buffer in the extracellular fluid, and it is also the buffer whose acid-to-base ratio can be changed very easily and rapidly. Acid–base balance The pH of the extracellular fluid, including the blood plasma, is normally tightly regulated between 7.32 and 7.42 by the chemical buffers, the respiratory system, and the renal system. The normal pH in the fetus differs from that in the adult. In the fetus, the pH in the umbilical vein pH is normally 7.25 to 7.45 and that in the umbilical artery is normally 7.18 to 7.38. Aqueous buffer solutions will react with strong acids or strong bases by absorbing excess ions, or ions, replacing the strong acids and bases with weak acids and weak bases. This has the effect of damping the effect of pH changes, or reducing the pH change that would otherwise have occurred. But buffers cannot correct abnormal pH levels in a solution, be that solution in a test tube or in the extracellular fluid. Buffers typically consist of a pair of compounds in solution, one of which is a weak acid and the other a weak base. The most abundant buffer in the ECF consists of a solution of carbonic acid (H2CO3), and the bicarbonate () salt of, usually, sodium (Na+). Thus, when there is an excess of ions in the solution carbonic acid partially neutralizes them by forming H2O and bicarbonate () ions. Similarly an excess of H+ ions is partially neutralized by the bicarbonate component of the buffer solution to form carbonic acid (H2CO3), which, because it is a weak acid, remains largely in the undissociated form, releasing far fewer H+ ions into the solution than the original strong acid would have done. The pH of a buffer solution depends solely on the ratio of the molar concentrations of the weak acid to the weak base. The higher the concentration of the weak acid in the solution (compared to the weak base) the lower the resulting pH of the solution. Similarly, if the weak base predominates the higher the resulting pH. This principle is exploited to regulate the pH of the extracellular fluids (rather than just buffering the pH). For the carbonic acid-bicarbonate buffer, a molar ratio of weak acid to weak base of 1:20 produces a pH of 7.4; and vice versa—when the pH of the extracellular fluids is 7.4 then the ratio of carbonic acid to bicarbonate ions in that fluid is 1:20. Henderson–Hasselbalch equation The Henderson–Hasselbalch equation, when applied to the carbonic acid-bicarbonate buffer system in the extracellular fluids, states that: where: is the negative logarithm (or cologarithm) of molar concentration of hydrogen ions in the extracellular fluid. is the cologarithm of the acid dissociation constant of carbonic acid. It is equal to 6.1. is the molar concentration of bicarbonate in the blood plasma. is the molar concentration of carbonic acid in the extracellular fluid. However, since the carbonic acid concentration is directly proportional to the partial pressure of carbon dioxide () in the extracellular fluid, the equation can be rewritten as follows: where: is the negative logarithm of molar concentration of hydrogen ions in the extracellular fluid. is the molar concentration of bicarbonate in the plasma. is the partial pressure of carbon dioxide in the blood plasma. The pH of the extracellular fluids can thus be controlled by the regulation of and the other metabolic acids. Homeostatic mechanisms Homeostatic control can change the and hence the pH of the arterial plasma within a few seconds. The partial pressure of carbon dioxide in the arterial blood is monitored by the central chemoreceptors of the medulla oblongata. These chemoreceptors are sensitive to the levels of carbon dioxide and pH in the cerebrospinal fluid. The central chemoreceptors send their information to the respiratory centers in the medulla oblongata and pons of the brainstem. The respiratory centres then determine the average rate of ventilation of the alveoli of the lungs, to keep the in the arterial blood constant. The respiratory center does so via motor neurons which activate the muscles of respiration (in particular, the diaphragm). A rise in the in the arterial blood plasma above reflexly causes an increase in the rate and depth of breathing. Normal breathing is resumed when the partial pressure of carbon dioxide has returned to 5.3 kPa. The converse happens if the partial pressure of carbon dioxide falls below the normal range. Breathing may be temporally halted, or slowed down to allow carbon dioxide to accumulate once more in the lungs and arterial blood. The sensor for the plasma HCO concentration is not known for certain. It is very probable that the renal tubular cells of the distal convoluted tubules are themselves sensitive to the pH of the plasma. The metabolism of these cells produces CO2, which is rapidly converted to H+ and HCO through the action of carbonic anhydrase. When the extracellular fluids tend towards acidity, the renal tubular cells secrete the H+ ions into the tubular fluid from where they exit the body via the urine. The HCO ions are simultaneously secreted into the blood plasma, thus raising the bicarbonate ion concentration in the plasma, lowering the carbonic acid/bicarbonate ion ratio, and consequently raising the pH of the plasma. The converse happens when the plasma pH rises above normal: bicarbonate ions are excreted into the urine, and hydrogen ions into the plasma. These combine with the bicarbonate ions in the plasma to form carbonic acid (H+ + HCO H2CO3), thus raising the carbonic acid:bicarbonate ratio in the extracellular fluids, and returning its pH to normal. In general, metabolism produces more waste acids than bases. Urine produced is generally acidic and is partially neutralized by the ammonia (NH3) that is excreted into the urine when glutamate and glutamine (carriers of excess, no longer needed, amino groups) are deaminated by the distal renal tubular epithelial cells. Thus some of the "acid content" of the urine resides in the resulting ammonium ion (NH4+) content of the urine, though this has no effect on pH homeostasis of the extracellular fluids. Imbalance Acid–base imbalance occurs when a significant insult causes the blood pH to shift out of the normal range (7.32 to 7.42). An abnormally low pH in the extracellular fluid is called an acidemia and an abnormally high pH is called an alkalemia. Acidemia and alkalemia unambiguously refer to the actual change in the pH of the extracellular fluid (ECF). Two other similar sounding terms are acidosis and alkalosis. They refer to the customary effect of a component, respiratory or metabolic. Acidosis would cause an acidemia on its own (i.e. if left "uncompensated" by an alkalosis). Similarly, an alkalosis would cause an alkalemia on its own. In medical terminology, the terms acidosis and alkalosis should always be qualified by an adjective to indicate the etiology of the disturbance: respiratory (indicating a change in the partial pressure of carbon dioxide), or metabolic (indicating a change in the Base Excess of the ECF). There are therefore four different acid-base problems: metabolic acidosis, respiratory acidosis, metabolic alkalosis, and respiratory alkalosis. One or a combination of these conditions may occur simultaneously. For instance, a metabolic acidosis (as in uncontrolled diabetes mellitus) is almost always partially compensated by a respiratory alkalosis (hyperventilation). Similarly, a respiratory acidosis can be completely or partially corrected by a metabolic alkalosis.
Biology and health sciences
Basics
Biology
13062770
https://en.wikipedia.org/wiki/Atmosphere%20of%20Uranus
Atmosphere of Uranus
The atmosphere of Uranus is composed primarily of hydrogen and helium. At depth, it is significantly enriched in volatiles (dubbed "ices") such as water, ammonia, and methane. The opposite is true for the upper atmosphere, which contains very few gases heavier than hydrogen and helium due to its low temperature. Uranus's atmosphere is the coldest of all the planets, with its temperature reaching as low as 49 K. The Uranian atmosphere can be divided into three main layers: the troposphere, between altitudes of −300 and 50 km and pressures from 100 to 0.1 bar; the stratosphere, spanning altitudes between 50 and 4000 km and pressures of between and the hot thermosphere (and exosphere) extending from an altitude of 4,000 km to several Uranian radii from the nominal surface at 1 bar pressure. Unlike Earth's, Uranus's atmosphere has no mesosphere. The troposphere hosts four cloud layers: methane clouds at about 1.2 bar, hydrogen sulfide and ammonia clouds at 3–10 bar, ammonium hydrosulfide clouds at 20–40 bar, and finally water clouds below 50 bar. Only the upper two cloud layers have been observed directly—the deeper clouds remain speculative. Above the clouds lie several tenuous layers of photochemical haze. Discrete bright tropospheric clouds are rare on Uranus, probably due to sluggish convection in the planet's interior. Nevertheless, observations of such clouds were used to measure the planet's zonal winds, which are remarkably fast with speeds up to 240 m/s. Little is known about the Uranian atmosphere. To date, only one spacecraft, Voyager 2, which passed by the planet in 1986, obtained some valuable compositional data. The Uranus Orbiter and Probe is scheduled to launch in 2031, arriving at Uranus in 2044. Its primary science objectives include a detailed study of Uranus' atmosphere. Observation and exploration Although there is no well-defined solid surface within Uranus's interior, the outermost part of Uranus's gaseous envelope (the region accessible to remote sensing) is called its atmosphere. Remote sensing capability extends down to roughly 300 km below the 1 bar level, with a corresponding pressure of around 100 bar and temperature of 320 K. The observational history of the Uranian atmosphere is long and full of error and frustration. Uranus is a relatively faint object, and its visible angular diameter is smaller than 5″. The first spectra of Uranus were observed through a prism in 1869 and 1871 by Angelo Secchi and William Huggins, who found a number of broad dark bands, which they were unable to identify. They also failed to detect any solar Fraunhofer lines—the fact later interpreted by Norman Lockyer as indicating that Uranus emitted its own light as opposed to reflecting light from the Sun. In 1889 however, astronomers observed solar Fraunhofer lines in photographic ultraviolet spectra of the planet, proving once and for all that Uranus was shining by reflected light. The nature of the broad dark bands in its visible spectrum remained unknown until the fourth decade of the twentieth century. Although Uranus is presently largely blank in appearance, it has been historically shown to have occasional features, such as in March and April 1884, when astronomers Henri Joseph Perrotin, Norman Lockyer, and Charles Trépied observed a bright, elongated spot (presumably a storm) circling the equator of the planet. The key to deciphering Uranus's spectrum was found in the 1930s by Rupert Wildt and Vesto Slipher, who found that the dark bands at 543, 619, 925, 865 and 890 nm belonged to gaseous methane. They had never been observed before because they were very weak and required a long path length to be detected. This meant that the atmosphere of Uranus was transparent to a much greater depth compared to those of other giant planets. In 1950, Gerard Kuiper noticed another diffuse dark band in the spectrum of Uranus at 827 nm, which he failed to identify. In 1952 Gerhard Herzberg, a future Nobel Prize winner, showed that this band was caused by the weak quadrupole absorption of molecular hydrogen, which thus became the second compound detected on Uranus. Until 1986 only two gases, methane and hydrogen, were known in the Uranian atmosphere. The far-infrared spectroscopic observation beginning from 1967 consistently showed the atmosphere of Uranus was in approximate thermal balance with incoming solar radiation (in other words, it radiated as much heat as it received from the Sun), and no internal heat source was required to explain observed temperatures. No discrete features had been observed on Uranus prior to the Voyager 2 visit in 1986. In January 1986, the Voyager 2 spacecraft flew by Uranus at a minimal distance of 107,100 km providing the first close-up images and spectra of its atmosphere. They generally confirmed that the atmosphere was made of mainly hydrogen and helium with around 2% methane. The atmosphere appeared highly transparent and lacking thick stratospheric and tropospheric hazes. Only a limited number of discrete clouds were observed. In the 1990s and 2000s, observations by the Hubble Space Telescope and by ground-based telescopes equipped with adaptive optics systems (the Keck telescope and NASA Infrared Telescope Facility, for instance) made it possible for the first time to observe discrete cloud features from Earth. Tracking them has allowed astronomers to re-measure wind speeds on Uranus, known before only from the Voyager 2 observations, and to study the dynamics of the Uranian atmosphere. Composition The composition of the Uranian atmosphere is different from that of Uranus as a whole, consisting mainly of molecular hydrogen and helium. The helium molar fraction, i.e. the number of helium atoms per molecule of hydrogen/helium, was determined from the analysis of Voyager 2 far infrared and radio occultation observations. The currently accepted value is in the upper troposphere, which corresponds to a mass fraction . This value is very close to the protosolar helium mass fraction of , indicating that helium has not settled towards the centre of the planet as it has in the gas giants. The third most abundant constituent of the Uranian atmosphere is methane , the presence of which has been known for some time as a result of the ground-based spectroscopic observations. Methane possesses prominent absorption bands in the visible and near-infrared, making Uranus aquamarine or cyan in color. Below the methane cloud deck at 1.3 bar methane molecules account for about 2.3% of the atmosphere by molar fraction; about 10 to 30 times that found in the Sun. The mixing ratio is much lower in the upper atmosphere due to the extremely low temperature at the tropopause, which lowers the saturation level and causes excess methane to freeze out. Methane appears to be undersaturated in the upper troposphere above the clouds having a partial pressure of only 30% of the saturated vapor pressure there. The concentration of less volatile compounds such as ammonia, water and hydrogen sulfide in the deep atmosphere is poorly known. However, as with methane, their abundances are probably greater than solar values by a factor of at least 20 to 30, and possibly by a factor of a few hundred. Knowledge of the isotopic composition of Uranus's atmosphere is very limited. To date the only known isotope abundance ratio is that of deuterium to light hydrogen: , which was measured by the Infrared Space Observatory (ISO) in the 1990s. It appears to be higher than the protosolar value of measured in Jupiter. The deuterium is found almost exclusively in hydrogen deuteride molecules which it forms with normal hydrogen atoms. Infrared spectroscopy, including measurements with Spitzer Space Telescope (SST), and UV occultation observations, found trace amounts of complex hydrocarbons in the stratosphere of Uranus, which are thought to be produced from methane by photolysis induced by solar UV radiation. They include ethane , acetylene , methylacetylene , diacetylene . Infrared spectroscopy also uncovered traces of water vapour, carbon monoxide and carbon dioxide in the stratosphere, which are likely to originate from an external source such as infalling dust and comets. Structure The Uranian atmosphere can be divided into three main layers: the troposphere, between altitudes of −300 and 50 km and pressures from 100 to 0.1 bar; the stratosphere, spanning altitudes between 50 and 4000 km and pressures between and the thermosphere/exosphere extending from 4000 km to as high as a few Uranus radii from the surface. There is no mesosphere. Troposphere The troposphere is the lowest and densest part of the atmosphere and is characterised by a decrease in temperature with altitude. The temperature falls from about 320 K at the base of the troposphere at −300 km to about 53 K at 50 km. The temperature at the cold upper boundary of the troposphere (the tropopause) actually varies in the range between 49 and 57 K depending on planetary latitude, with the lowest temperature reached near 25° southern latitude. The troposphere holds almost all of the mass of the atmosphere, and the tropopause region is also responsible for the vast majority of the planet's thermal far infrared emissions, thus determining its effective temperature of . The troposphere is believed to possess a highly complex cloud structure; water clouds are hypothesised to lie in the pressure range of , ammonium hydrosulfide clouds in the range of , ammonia or hydrogen sulfide clouds at between 3 and 10 bar and finally thin methane clouds at . Although Voyager 2 directly detected methane clouds, all other cloud layers remain speculative. The existence of a hydrogen sulfide cloud layer is only possible if the ratio of sulfur and nitrogen abundances (S/N ratio) is significantly greater than its solar value of 0.16. Otherwise all hydrogen sulfide would react with ammonia, producing ammonium hydrosulfide, and the ammonia clouds would appear instead in the pressure range 3–10 bar. The elevated S/N ratio implies depletion of ammonia in the pressure range 20–40 bar, where the ammonium hydrosulfide clouds form. These can result from the dissolution of ammonia in water droplets within water clouds or in the deep water-ammonia ionic ocean. The exact location of the upper two cloud layers is somewhat controversial. Methane clouds were directly detected by Voyager 2 at 1.2–1.3 bar by radio occultation. This result was later confirmed by an analysis of the Voyager 2 limb images. The top of the deeper ammonia/hydrogen sulfide clouds were determined to be at 3 bar based on the spectroscopic data in the visible and near-infra spectral ranges (0.5–1 μm). However a recent analysis of the spectroscopic data in the wavelength range 1–2.3 μm placed the methane cloudtops at 2 bar, and the top of the lower clouds at 6 bar. This contradiction may be resolved when new data on methane absorption in Uranus's atmosphere are available. The optical depth of the two upper cloud layers varies with latitude: both become thinner at the poles as compared to the equator, though in 2007 the methane cloud layer's optical depth had a local maximum at 45°S, where the southern polar collar is located (see below). The troposphere is very dynamic, exhibiting strong zonal winds, bright methane clouds, dark spots and noticeable seasonal changes. (see below) Stratosphere The stratosphere is the middle layer of the Uranian atmosphere, in which temperature generally increases with altitude from 53 K in the tropopause to between 800 and 850 K at the base thermosphere. The heating of the stratosphere is caused by the downward heat conduction from the hot thermosphere as well as by absorption of solar UV and IR radiation by methane and the complex hydrocarbons formed as a result of methane photolysis. The methane enters the stratosphere through the cold tropopause, where its mixing ratio relative to molecular hydrogen is about 3, three times below saturation. It decreases further to about 10−7 at the altitude corresponding to pressure of 0.1 mbar. Hydrocarbons heavier than methane are present in a relatively narrow layer between 160 and 320 km in altitude, corresponding to the pressure range from 10 to 0.1 mbar and temperatures from 100 to 130 K. The most abundant stratospheric hydrocarbons after methane are acetylene and ethane, with mixing ratios of around 10−7. Heavier hydrocarbons like methylacetylene and diacetylene have mixing ratios of about 10−10—three orders of magnitude lower. The temperature and hydrocarbon mixing ratios in the stratosphere vary with time and latitude. Complex hydrocarbons are responsible for the cooling of the stratosphere, especially acetylene, having a strong emission line at the wavelength of 13.7 μm. In addition to hydrocarbons, the stratosphere contains carbon monoxide, as well as traces of water vapor and carbon dioxide. The mixing ratio of carbon monoxide—3—is very similar to that of the hydrocarbons, while the mixing ratios of carbon dioxide and water are about 10−11 and 8, respectively. These three compounds are distributed relatively homogeneously in the stratosphere and are not confined to a narrow layer like hydrocarbons. Ethane, acetylene and diacetylene condense in the colder lower part of stratosphere forming haze layers with an optical depth of about 0.01 in visible light. Condensation occurs at approximately 14, 2.5 and 0.1 mbar for ethane, acetylene and diacetylene, respectively. The concentration of hydrocarbons in the Uranian stratosphere is significantly lower than in the stratospheres of the other giant planets—the upper atmosphere of Uranus is very clean and transparent above the haze layers. This depletion is caused by weak vertical mixing, and makes Uranus's stratosphere less opaque and, as a result, colder than those of other giant planets. The hazes, like their parent hydrocarbons, are distributed unevenly across Uranus; at the solstice of 1986, when Voyager 2 passed by the planet, they were concentrated near the sunlit pole, making it dark in ultraviolet light. Thermosphere and ionosphere The outermost layer of the Uranian atmosphere, extending for thousands of kilometres, is the thermosphere/exosphere, which has a uniform temperature of around 800 to 850 K. This is much higher than, for instance, the 420 K observed in the thermosphere of Saturn. The heat sources necessary to sustain such high temperatures are not understood, since neither solar FUV/EUV radiation nor auroral activity can provide the necessary energy. The weak cooling efficiency due to the depletion of hydrocarbons in the stratosphere may contribute to this phenomenon. In addition to molecular hydrogen, the thermosphere contains a large proportion of free hydrogen atoms, while helium is thought to be absent here, because it separates diffusively at lower altitudes. The thermosphere and upper part of the stratosphere contain a large concentration of ions and electrons, forming the ionosphere of Uranus. Radio occultation observations by the Voyager 2 spacecraft showed that the ionosphere lies between 1,000 and 10,000 km altitude and may include several narrow and dense layers between 1,000 and 3,500 km. The electron density in the Uranian ionosphere is on average , reaching to as high as in the narrow layers in the stratosphere. The ionosphere is mainly sustained by solar UV radiation and its density depends on the solar activity. The auroral activity on Uranus is not as powerful as at Jupiter and Saturn and contributes little to the ionization. The high electron density may be in part caused by the low concentration of hydrocarbons in the stratosphere. One of the sources of information about the ionosphere and thermosphere comes from ground-based measurements of the intense mid-infrared (3–4 μm) emissions of the trihydrogen cation (). The total emitted power is 1–2 W—an order of magnitude higher than that the near-infrared hydrogen quadrupole emissions. Trihydrogen cation functions as one of main coolers of the ionosphere. The upper atmosphere of Uranus is the source of the far ultraviolet (90–140 nm) emissions known as dayglow or electroglow, which, like the IR radiation, emanates exclusively from the sunlit part of the planet. This phenomenon, which occurs in the thermospheres of all giant planets and was mysterious for a time after its discovery, is interpreted as a UV fluorescence of atomic and molecular hydrogen excited by solar radiation or by photoelectrons. Hydrogen corona The upper part of the thermosphere, where the mean free path of the molecules exceeds the scale height, is called the exosphere. The lower boundary of the Uranian exosphere, the exobase, is located at a height of about 6,500 km, or 1/4 of the planetary radius, above the surface. The exosphere is unusually extended, reaching as far as several Uranian radii from the planet. It is made mainly of hydrogen atoms and is often called the hydrogen corona of Uranus. The high temperature and relatively high pressure at the base of the thermosphere explain in part why Uranus's exosphere is so vast. The number density of atomic hydrogen in the corona falls slowly with the distance from the planet, remaining as high a few hundred atoms per cm3 at a few radii from Uranus. The effects of this bloated exosphere include a drag on small particles orbiting Uranus, causing a general depletion of dust in the Uranian rings. The infalling dust in turn contaminates the upper atmosphere of the planet. Dynamics Uranus has a relatively bland appearance, lacking broad colorful bands and large clouds prevalent on Jupiter and Saturn. Discrete features were only once observed in Uranus's atmosphere before 1986. The most conspicuous features on Uranus observed by Voyager 2 were the dark low latitude region between −40° and −20° and bright southern polar cap. The northern boundary of the cap was located at about −45° of latitude. The brightest zonal band was located near the edge of the cap at −50° to −45° and was then called a polar collar. The southern polar cap, which existed at the time of the solstice in 1986, faded away in 1990s. After the equinox in 2007, the southern polar collar started to fade away as well, while the northern polar collar located at 45° to 50° latitude (first appeared in 2007) have grown more conspicuous since then. The atmosphere of Uranus is calm compared to those of other giant planets. Only a limited number of small bright clouds at middle latitudes in both hemispheres and one Uranus Dark Spot have been observed since 1986. One of those bright cloud features, located at −34° of latitude and called Berg, probably existed continuously since at least 1986. Nevertheless, the Uranian atmosphere has rather strong zonal winds blowing in the retrograde (counter to the rotation) direction near the equator, but switching to the prograde direction poleward of ±20° latitude. The wind speeds are from −50 to −100 m/s at the equator increasing up to 240 m/s near 50° latitude. The wind profile measured before the equinox of 2007 was slightly asymmetric with winds stronger in the southern hemisphere, although it turned out to be a seasonal effect as this hemisphere was continuously illuminated by the Sun before 2007. After 2007 winds in the northern hemisphere accelerated while those in the southern one slowed down. Uranus exhibits a considerable seasonal variation over its 84-year orbit. It is generally brighter near solstices and dimmer at equinoxes. The variations are largely caused by changes in the viewing geometry: a bright polar region comes into view near solstices, while the dark equator is visible near equinoxes. Still there exist some intrinsic variations of the reflectivity of the atmosphere: periodically fading and brightening polar caps as well as appearing and disappearing polar collars.
Physical sciences
Solar System
Astronomy
1303134
https://en.wikipedia.org/wiki/Ethyl%20acetate
Ethyl acetate
Ethyl acetate (systematically ethyl ethanoate, commonly abbreviated EtOAc, ETAC or EA) is the organic compound with the formula , simplified to . This flammable, colorless liquid has a characteristic sweet smell (similar to pear drops) and is used in glues, nail polish removers, and the decaffeination process of tea and coffee. Ethyl acetate is the ester of ethanol and acetic acid; it is manufactured on a large scale for use as a solvent. Production and synthesis Ethyl acetate was first synthesized by the Count de Lauraguais in 1759 by distilling a mixture of ethanol and acetic acid. In 2004, an estimated 1.3 million tonnes were produced worldwide. The combined annual production in 1985 of Japan, North America, and Europe was about 400,000 tonnes. The global ethyl acetate market was valued at $3.3 billion in 2018. Ethyl acetate is synthesized in industry mainly via the classic Fischer esterification reaction of ethanol and acetic acid. This mixture converts to the ester in about 65% yield at room temperature: The reaction can be accelerated by acid catalysis and the equilibrium can be shifted to the right by removal of water. It is also prepared in industry using the Tishchenko reaction, by combining two equivalents of acetaldehyde in the presence of an alkoxide catalyst: Silicotungstic acid is used to manufacture ethyl acetate by the alkylation of acetic acid by ethylene: Uses Ethyl acetate is used primarily as a solvent and diluent, being favored because of its low cost, low toxicity, and agreeable odor. For example, it is commonly used to clean circuit boards and in some nail varnish removers (acetone is also used). Coffee beans and tea leaves are decaffeinated with this solvent. It is also used in paints as an activator or hardener. Ethyl acetate is present in confectionery, perfumes, and fruits. In perfumes it evaporates quickly, leaving the scent of the perfume on the skin. Ethyl acetate is an asphyxiant for use in insect collecting and study. In a killing jar charged with ethyl acetate, the vapors will kill the collected insect quickly without destroying it. Because it is not hygroscopic, ethyl acetate also keeps the insect soft enough to allow proper mounting suitable for a collection. However, ethyl acetate is regarded as potentially doing damage to insect DNA, making specimens processed this way less than ideal for subsequent DNA sequencing. Laboratory uses In the laboratory, mixtures containing ethyl acetate are commonly used in column chromatography and extractions. Ethyl acetate is rarely selected as a reaction solvent because it is prone to hydrolysis, transesterification, and condensations. Occurrence in wines Ethyl acetate is the most common ester in wine, being the product of the most common volatile organic acid – acetic acid, and the ethyl alcohol generated during the fermentation. The aroma of ethyl acetate is most vivid in younger wines and contributes towards the general perception of "fruitiness" in the wine. Sensitivity varies, with most people having a perception threshold around 120 mg/L. Excessive amounts of ethyl acetate are considered a wine fault. Reactions Ethyl acetate is only weakly Lewis basic, like a typical carboxylic acid ester. Ethyl acetate hydrolyses to give acetic acid and ethanol. Bases accelerate the hydrolysis, which is subject to the Fischer equilibrium mentioned above. In the laboratory, and usually for illustrative purposes only, ethyl esters are typically hydrolyzed in a two-step process starting with a stoichiometric amount of a strong base, such as sodium hydroxide. This reaction gives ethanol and sodium acetate, which is unreactive toward ethanol: In the Claisen condensation, anhydrous ethyl acetate and strong bases react to give ethyl acetoacetate: Properties Physical properties Under normal conditions, ethyl acetate exists as a colorless, low-viscosity, and flammable liquid. Its melting point is −83 °C, with a melting enthalpy of 10.48 kJ/mol. At atmospheric pressure, the compound boils at 77 °C. The vaporization enthalpy at the boiling point is 31.94 kJ/mol. The vapor pressure function follows the Antoine equation where is the vapor pressure in bars, is the absolute temperature in kelvins, and , , are constants. This function is valid within the temperature range of . The enthalpy of vaporization in kJ/mol is calculated according to the empirical equation by Majer and Svoboda where is the reduced temperature, and = 523.2 K is the critical temperature. = 54.26 kJ/mol and = 0.2982 are constants. The following table summarizes the most important thermodynamic properties of ethyl acetate under various conditions. Safety The for rats is 5620 mg/kg, indicating low acute toxicity. Given that the chemical is naturally present in many organisms, there is little risk of toxicity. Overexposure to ethyl acetate may cause irritation of the eyes, nose, and throat. Severe overexposure may cause weakness, drowsiness, and unconsciousness. Humans exposed to a concentration of 400 ppm in 1.4 mg/L ethyl acetate for a short time were affected by nose and throat irritation. Ethyl acetate is an irritant of the conjunctiva and mucous membrane of the respiratory tract. Animal experiments have shown that, at very high concentrations, the ester has central nervous system depressant and lethal effects; at concentrations of 20,000 to 43,000 ppm (2.0–4.3%), there may be pulmonary edema with hemorrhages, symptoms of central nervous system depression, secondary anemia and liver damage. In humans, concentrations of 400 ppm cause irritation of the nose and pharynx; cases have also been known of irritation of the conjunctiva with temporary opacity of the cornea. In rare cases exposure may cause sensitization of the mucous membrane and eruptions of the skin. The irritant effect of ethyl acetate is weaker than that of propyl acetate or butyl acetate.
Physical sciences
Esters and ethers
Chemistry
1305071
https://en.wikipedia.org/wiki/Bridge%20%28graph%20theory%29
Bridge (graph theory)
In graph theory, a bridge, isthmus, cut-edge, or cut arc is an edge of a graph whose deletion increases the graph's number of connected components. Equivalently, an edge is a bridge if and only if it is not contained in any cycle. For a connected graph, a bridge can uniquely determine a cut. A graph is said to be bridgeless or isthmus-free if it contains no bridges. This type of bridge should be distinguished from an unrelated meaning of "bridge" in graph theory, a subgraph separated from the rest of the graph by a specified subset of vertices; see bridge in the Glossary of graph theory. Trees and forests A graph with nodes can contain at most bridges, since adding additional edges must create a cycle. The graphs with exactly bridges are exactly the trees, and the graphs in which every edge is a bridge are exactly the forests. In every undirected graph, there is an equivalence relation on the vertices according to which two vertices are related to each other whenever there are two edge-disjoint paths connecting them. (Every vertex is related to itself via two length-zero paths, which are identical but nevertheless edge-disjoint.) The equivalence classes of this relation are called 2-edge-connected components, and the bridges of the graph are exactly the edges whose endpoints belong to different components. The bridge-block tree of the graph has a vertex for every nontrivial component and an edge for every bridge. Relation to vertex connectivity Bridges are closely related to the concept of articulation vertices, vertices that belong to every path between some pair of other vertices. The two endpoints of a bridge are articulation vertices unless they have a degree of 1, although it may also be possible for a non-bridge edge to have two articulation vertices as endpoints. Analogously to bridgeless graphs being 2-edge-connected, graphs without articulation vertices are 2-vertex-connected. In a cubic graph, every cut vertex is an endpoint of at least one bridge. Bridgeless graphs A bridgeless graph is a graph that does not have any bridges. Equivalent conditions are that each connected component of the graph has an open ear decomposition, that each connected component is 2-edge-connected, or (by Robbins' theorem) that every connected component has a strong orientation. An important open problem involving bridges is the cycle double cover conjecture, due to Seymour and Szekeres (1978 and 1979, independently), which states that every bridgeless graph admits a multi-set of simple cycles which contains each edge exactly twice. Tarjan's bridge-finding algorithm The first linear time algorithm (linear in the number of edges) for finding the bridges in a graph was described by Robert Tarjan in 1974. It performs the following steps: Find a spanning forest of Create a Rooted forest from the spanning forest Traverse the forest in preorder and number the nodes. Parent nodes in the forest now have lower numbers than child nodes. For each node in preorder (denoting each node using its preorder number), do: Compute the number of forest descendants for this node, by adding one to the sum of its children's descendants. Compute , the lowest preorder label reachable from by a path for which all but the last edge stays within the subtree rooted at . This is the minimum of the set consisting of the preorder label of , of the values of at child nodes of and of the preorder labels of nodes reachable from by edges that do not belong to . Similarly, compute , the highest preorder label reachable by a path for which all but the last edge stays within the subtree rooted at . This is the maximum of the set consisting of the preorder label of , of the values of at child nodes of and of the preorder labels of nodes reachable from by edges that do not belong to . For each node with parent node , if and then the edge from to is a bridge. Bridge-finding with chain decompositions A very simple bridge-finding algorithm uses chain decompositions. Chain decompositions do not only allow to compute all bridges of a graph, they also allow to read off every cut vertex of G (and the block-cut tree of G), giving a general framework for testing 2-edge- and 2-vertex-connectivity (which extends to linear-time 3-edge- and 3-vertex-connectivity tests). Chain decompositions are special ear decompositions depending on a DFS-tree T of G and can be computed very simply: Let every vertex be marked as unvisited. For each vertex v in ascending DFS-numbers 1...n, traverse every backedge (i.e. every edge not in the DFS tree) that is incident to v and follow the path of tree-edges back to the root of T, stopping at the first vertex that is marked as visited. During such a traversal, every traversed vertex is marked as visited. Thus, a traversal stops at the latest at v and forms either a directed path or cycle, beginning with v; we call this path or cycle a chain. The ith chain found by this procedure is referred to as Ci. C=C1,C2,... is then a chain decomposition of G. The following characterizations then allow to read off several properties of G from C efficiently, including all bridges of G. Let C be a chain decomposition of a simple connected graph G=(V,E). G is 2-edge-connected if and only if the chains in C partition E. An edge e in G is a bridge if and only if e is not contained in any chain in C. If G is 2-edge-connected, C is an ear decomposition. G is 2-vertex-connected if and only if G has minimum degree 2 and C1 is the only cycle in C. A vertex v in a 2-edge-connected graph G is a cut vertex if and only if v is the first vertex of a cycle in C - C1. If G is 2-vertex-connected, C is an open ear decomposition.
Mathematics
Graph theory
null
1305947
https://en.wikipedia.org/wiki/3D%20printing
3D printing
3D printing, or additive manufacturing, is the construction of a three-dimensional object from a CAD model or a digital 3D model. It can be done in a variety of processes in which material is deposited, joined or solidified under computer control, with the material being added together (such as plastics, liquids or powder grains being fused), typically layer by layer. In the 1980s, 3D printing techniques were considered suitable only for the production of functional or aesthetic prototypes, and a more appropriate term for it at the time was rapid prototyping. , the precision, repeatability, and material range of 3D printing have increased to the point that some 3D printing processes are considered viable as an industrial-production technology; in this context, the term additive manufacturing can be used synonymously with 3D printing. One of the key advantages of 3D printing is the ability to produce very complex shapes or geometries that would be otherwise infeasible to construct by hand, including hollow parts or parts with internal truss structures to reduce weight while creating less material waste. Fused deposition modeling (FDM), which uses a continuous filament of a thermoplastic material, is the most common 3D printing process in use . Terminology The umbrella term additive manufacturing (AM) gained popularity in the 2000s, inspired by the theme of material being added together (in any of various ways). In contrast, the term subtractive manufacturing appeared as a retronym for the large family of machining processes with material removal as their common process. The term 3D printing still referred only to the polymer technologies in most minds, and the term AM was more likely to be used in metalworking and end-use part production contexts than among polymer, inkjet, or stereolithography enthusiasts. By the early 2010s, the terms 3D printing and additive manufacturing evolved senses in which they were alternate umbrella terms for additive technologies, one being used in popular language by consumer-maker communities and the media, and the other used more formally by industrial end-use part producers, machine manufacturers, and global technical standards organizations. Until recently, the term 3D printing has been associated with machines low in price or capability. 3D printing and additive manufacturing reflect that the technologies share the theme of material addition or joining throughout a 3D work envelope under automated control. Peter Zelinski, the editor-in-chief of Additive Manufacturing magazine, pointed out in 2017 that the terms are still often synonymous in casual usage, but some manufacturing industry experts are trying to make a distinction whereby additive manufacturing comprises 3D printing plus other technologies or other aspects of a manufacturing process. Other terms that have been used as synonyms or hypernyms have included desktop manufacturing, rapid manufacturing (as the logical production-level successor to rapid prototyping), and on-demand manufacturing (which echoes on-demand printing in the 2D sense of printing). The fact that the application of the adjectives rapid and on-demand to the noun manufacturing was novel in the 2000s reveals the long-prevailing mental model of the previous industrial era during which almost all production manufacturing had involved long lead times for laborious tooling development. Today, the term subtractive has not replaced the term machining, instead complementing it when a term that covers any removal method is needed. Agile tooling is the use of modular means to design tooling that is produced by additive manufacturing or 3D printing methods to enable quick prototyping and responses to tooling and fixture needs. Agile tooling uses a cost-effective and high-quality method to quickly respond to customer and market needs, and it can be used in hydroforming, stamping, injection molding and other manufacturing processes. History 1940s and 1950s The general concept of and procedure to be used in 3D-printing was first described by Murray Leinster in his 1945 short story "Things Pass By": "But this constructor is both efficient and flexible. I feed magnetronic plastics — the stuff they make houses and ships of nowadays — into this moving arm. It makes drawings in the air following drawings it scans with photo-cells. But plastic comes out of the end of the drawing arm and hardens as it comes ... following drawings only" It was also described by Raymond F. Jones in his story, "Tools of the Trade", published in the November 1950 issue of Astounding Science Fiction magazine. He referred to it as a "molecular spray" in that story. 1970s In 1971, Johannes F Gottwald patented the Liquid Metal Recorder, U.S. patent 3596285A, a continuous inkjet metal material device to form a removable metal fabrication on a reusable surface for immediate use or salvaged for printing again by remelting. This appears to be the first patent describing 3D printing with rapid prototyping and controlled on-demand manufacturing of patterns. The patent states: In 1974, David E. H. Jones laid out the concept of 3D printing in his regular column Ariadne in the journal New Scientist. 1980s Early additive manufacturing equipment and materials were developed in the 1980s. In April 1980, Hideo Kodama of Nagoya Municipal Industrial Research Institute invented two additive methods for fabricating three-dimensional plastic models with photo-hardening thermoset polymer, where the UV exposure area is controlled by a mask pattern or a scanning fiber transmitter. He filed a patent for this XYZ plotter, which was published on 10 November 1981. (JP S56-144478). His research results as journal papers were published in April and November 1981. However, there was no reaction to the series of his publications. His device was not highly evaluated in the laboratory and his boss did not show any interest. His research budget was just 60,000 yen or $545 a year. Acquiring the patent rights for the XYZ plotter was abandoned, and the project was terminated. A US 4323756 patent, method of fabricating articles by sequential deposition, granted on 6 April 1982 to Raytheon Technologies Corp describes using hundreds or thousands of "layers" of powdered metal and a laser energy source and represents an early reference to forming "layers" and the fabrication of articles on a substrate. On 2 July 1984, American entrepreneur Bill Masters filed a patent for his computer automated manufacturing process and system (US 4665492). This filing is on record at the USPTO as the first 3D printing patent in history; it was the first of three patents belonging to Masters that laid the foundation for the 3D printing systems used today. On 16 July 1984, Alain Le Méhauté, Olivier de Witte, and Jean Claude André filed their patent for the stereolithography process. The application of the French inventors was abandoned by the French General Electric Company (now Alcatel-Alsthom) and CILAS (The Laser Consortium). The claimed reason was "for lack of business perspective". In 1983, Robert Howard started R.H. Research, later named Howtek, Inc. in Feb 1984 to develop a color inkjet 2D printer, Pixelmaster, commercialized in 1986, using Thermoplastic (hot-melt) plastic ink. A team was put together, 6 members from Exxon Office Systems, Danbury Systems Division, an inkjet printer startup and some members of Howtek, Inc group who became popular figures in the 3D printing industry. One Howtek member, Richard Helinski (patent US5136515A, Method and Means for constructing three-dimensional articles by particle deposition, application 11/07/1989 granted 8/04/1992) formed a New Hampshire company C.A.D-Cast, Inc, name later changed to Visual Impact Corporation (VIC) on 8/22/1991. A prototype of the VIC 3D printer for this company is available with a video presentation showing a 3D model printed with a single nozzle inkjet. Another employee Herbert Menhennett formed a New Hampshire company HM Research in 1991 and introduced the Howtek, Inc, inkjet technology and thermoplastic materials to Royden Sanders of SDI and Bill Masters of Ballistic Particle Manufacturing (BPM) where he worked for a number of years. Both BPM 3D printers and SPI 3D printers use Howtek, Inc style Inkjets and Howtek, Inc style materials. Royden Sanders licensed the Helinksi patent prior to manufacturing the Modelmaker 6 Pro at Sanders prototype, Inc (SPI) in 1993. James K. McMahon who was hired by Howtek, Inc to help develop the inkjet, later worked at Sanders Prototype and now operates Layer Grown Model Technology, a 3D service provider specializing in Howtek single nozzle inkjet and SDI printer support. James K. McMahon worked with Steven Zoltan, 1972 drop-on-demand inkjet inventor, at Exxon and has a patent in 1978 that expanded the understanding of the single nozzle design inkjets (Alpha jets) and helped perfect the Howtek, Inc hot-melt inkjets. This Howtek hot-melt thermoplastic technology is popular with metal investment casting, especially in the 3D printing jewelry industry. Sanders (SDI) first Modelmaker 6Pro customer was Hitchner Corporations, Metal Casting Technology, Inc in Milford, NH a mile from the SDI facility in late 1993-1995 casting golf clubs and auto engine parts. On 8 August 1984 a patent, US4575330, assigned to UVP, Inc., later assigned to Chuck Hull of 3D Systems Corporation was filed, his own patent for a stereolithography fabrication system, in which individual laminae or layers are added by curing photopolymers with impinging radiation, particle bombardment, chemical reaction or just ultraviolet light lasers. Hull defined the process as a "system for generating three-dimensional objects by creating a cross-sectional pattern of the object to be formed". Hull's contribution was the STL (Stereolithography) file format and the digital slicing and infill strategies common to many processes today. In 1986, Charles "Chuck" Hull was granted a patent for this system, and his company, 3D Systems Corporation was formed and it released the first commercial 3D printer, the SLA-1, later in 1987 or 1988. The technology used by most 3D printers to date—especially hobbyist and consumer-oriented models—is fused deposition modeling, a special application of plastic extrusion, developed in 1988 by S. Scott Crump and commercialized by his company Stratasys, which marketed its first FDM machine in 1992. Owning a 3D printer in the 1980s cost upwards of $300,000 ($650,000 in 2016 dollars). 1990s AM processes for metal sintering or melting (such as selective laser sintering, direct metal laser sintering, and selective laser melting) usually went by their own individual names in the 1980s and 1990s. At the time, all metalworking was done by processes that are now called non-additive (casting, fabrication, stamping, and machining); although plenty of automation was applied to those technologies (such as by robot welding and CNC), the idea of a tool or head moving through a 3D work envelope transforming a mass of raw material into a desired shape with a toolpath was associated in metalworking only with processes that removed metal (rather than adding it), such as CNC milling, CNC EDM, and many others. However, the automated techniques that added metal, which would later be called additive manufacturing, were beginning to challenge that assumption. By the mid-1990s, new techniques for material deposition were developed at Stanford and Carnegie Mellon University, including microcasting and sprayed materials. Sacrificial and support materials had also become more common, enabling new object geometries. The term 3D printing originally referred to a powder bed process employing standard and custom inkjet print heads, developed at MIT by Emanuel Sachs in 1993 and commercialized by Soligen Technologies, Extrude Hone Corporation, and Z Corporation. The year 1993 also saw the start of an inkjet 3D printer company initially named Sanders Prototype, Inc and later named Solidscape, introducing a high-precision polymer jet fabrication system with soluble support structures, (categorized as a "dot-on-dot" technique). In 1995 the Fraunhofer Society developed the selective laser melting process. 2000s In the early 2000s 3D printers were still largely being used just in the manufacturing and research industries, as the technology was still relatively young and was too expensive for most consumers to be able to get their hands on. The 2000s was when larger scale use of the technology began being seen in industry, most often in the architecture and medical industries, though it was typically used for low accuracy modeling and testing, rather than the production of common manufactured goods or heavy prototyping. In 2005 users began to design and distribute plans for 3D printers that could print around 70% of their own parts, the original plans of which were designed by Adrian Bowyer at the University of Bath in 2004, with the name of the project being RepRap (Replicating Rapid-prototyper). Similarly, in 2006 the Fab@Home project was started by Evan Malone and Hod Lipson, another project whose purpose was to design a low-cost and open source fabrication system that users could develop on their own and post feedback on, making the project very collaborative. Much of the software for 3D printing available to the public at the time was open source, and as such was quickly distributed and improved upon by many individual users. In 2009 the Fused Deposition Modeling (FDM) printing process patents expired. This opened the door to a new wave of startup companies, many of which were established by major contributors of these open source initiatives, with the goal of many of them being to start developing commercial FDM 3D printers that were more accessible to the general public. 2010s As the various additive processes matured, it became clear that soon metal removal would no longer be the only metalworking process done through a tool or head moving through a 3D work envelope, transforming a mass of raw material into a desired shape layer by layer. The 2010s were the first decade in which metal end-use parts such as engine brackets and large nuts would be grown (either before or instead of machining) in job production rather than obligately being machined from bar stock or plate. It is still the case that casting, fabrication, stamping, and machining are more prevalent than additive manufacturing in metalworking, but AM is now beginning to make significant inroads, and with the advantages of design for additive manufacturing, it is clear to engineers that much more is to come. One place that AM is making a significant inroad is in the aviation industry. With nearly 3.8 billion air travelers in 2016, the demand for fuel efficient and easily produced jet engines has never been higher. For large OEMs (original equipment manufacturers) like Pratt and Whitney (PW) and General Electric (GE) this means looking towards AM as a way to reduce cost, reduce the number of nonconforming parts, reduce weight in the engines to increase fuel efficiency and find new, highly complex shapes that would not be feasible with the antiquated manufacturing methods. One example of AM integration with aerospace was in 2016 when Airbus delivered the first of GE's LEAP engines. This engine has integrated 3D-printed fuel nozzles, reducing parts from 20 to 1, a 25% weight reduction, and reduced assembly times. A fuel nozzle is the perfect inroad for additive manufacturing in a jet engine since it allows for optimized design of the complex internals and it is a low-stress, non-rotating part. Similarly, in 2015, PW delivered their first AM parts in the PurePower PW1500G to Bombardier. Sticking to low-stress, non-rotating parts, PW selected the compressor stators and synch ring brackets to roll out this new manufacturing technology for the first time. While AM is still playing a small role in the total number of parts in the jet engine manufacturing process, the return on investment can already be seen by the reduction in parts, the rapid production capabilities and the "optimized design in terms of performance and cost". As technology matured, several authors began to speculate that 3D printing could aid in sustainable development in the developing world. In 2012, Filabot developed a system for closing the loop with plastic and allows for any FDM or FFF 3D printer to be able to print with a wider range of plastics. In 2014, Benjamin S. Cook and Manos M. Tentzeris demonstrated the first multi-material, vertically integrated printed electronics additive manufacturing platform (VIPRE) which enabled 3D printing of functional electronics operating up to 40 GHz. As the price of printers started to drop people interested in this technology had more access and freedom to make what they wanted. As of 2014, the price for commercial printers was still high with the cost being over $2,000. The term "3D printing" originally referred to a process that deposits a binder material onto a powder bed with inkjet printer heads layer by layer. More recently, the popular vernacular has started using the term to encompass a wider variety of additive-manufacturing techniques such as electron-beam additive manufacturing and selective laser melting. The United States and global technical standards use the official term additive manufacturing for this broader sense. The most commonly used 3D printing process (46% ) is a material extrusion technique called fused deposition modeling, or FDM. While FDM technology was invented after the other two most popular technologies, stereolithography (SLA) and selective laser sintering (SLS), FDM is typically the most inexpensive of the three by a large margin, which lends to the popularity of the process. 2020s As of 2020, 3D printers have reached the level of quality and price that allows most people to enter the world of 3D printing. In 2020 decent quality printers can be found for less than US$200 for entry-level machines. These more affordable printers are usually fused deposition modeling (FDM) printers. In November 2021 a British patient named Steve Verze received the world's first fully 3D-printed prosthetic eye from the Moorfields Eye Hospital in London. In April 2024, the world's largest 3D printer, the Factory of the Future 1.0 was revealed at the University of Maine. It is able to make objects 96 feet long, or 29 meters. In 2024, researchers used machine learning to improve the construction of synthetic bone and set a record for shock absorption. In July 2024, researchers published a paper in Advanced Materials Technologies describing the development of artificial blood vessels using 3D-printing technology, which are as strong and durable as natural blood vessels. The process involved using a rotating spindle integrated into a 3D printer to create grafts from a water-based gel, which were then coated in biodegradable polyester molecules. Benefits of 3D printing Additive manufacturing or 3D printing has rapidly gained importance in the field of engineering due to its many benefits. The vision of 3D printing is design freedom, individualization, decentralization and executing processes that were previously impossible through alternative methods. Some of these benefits include enabling faster prototyping, reducing manufacturing costs, increasing product customization, and improving product quality. Furthermore, the capabilities of 3D printing have extended beyond traditional manufacturing, like lightweight construction, or repair and maintenance with applications in prosthetics, bioprinting, food industry, rocket building, design and art and renewable energy systems. 3D printing technology can be used to produce battery energy storage systems, which are essential for sustainable energy generation and distribution. Another benefit of 3D printing is the technology's ability to produce complex geometries with high precision and accuracy. This is particularly relevant in the field of microwave engineering, where 3D printing can be used to produce components with unique properties that are difficult to achieve using traditional manufacturing methods. Additive Manufacturing processes generate minimal waste by adding material only where needed, unlike traditional methods that cut away excess material. This reduces both material costs and environmental impact. This reduction in waste also lowers energy consumption for material production and disposal, contributing to a smaller carbon footprint. General principles Modeling 3D printable models may be created with a computer-aided design (CAD) package, via a 3D scanner, or by a plain digital camera and photogrammetry software. 3D printed models created with CAD result in relatively fewer errors than other methods. Errors in 3D printable models can be identified and corrected before printing. The manual modeling process of preparing geometric data for 3D computer graphics is similar to plastic arts such as sculpting. 3D scanning is a process of collecting digital data on the shape and appearance of a real object, and creating a digital model based on it. CAD models can be saved in the stereolithography file format (STL), a de facto CAD file format for additive manufacturing that stores data based on triangulations of the surface of CAD models. STL is not tailored for additive manufacturing because it generates large file sizes of topology-optimized parts and lattice structures due to the large number of surfaces involved. A newer CAD file format, the additive manufacturing file format (AMF), was introduced in 2011 to solve this problem. It stores information using curved triangulations. Printing Before printing a 3D model from an STL file, it must first be examined for errors. Most CAD applications produce errors in output STL files, of the following types: holes faces normals self-intersections noise shells manifold errors overhang issues A step in the STL generation known as "repair" fixes such problems in the original model. Generally, STLs that have been produced from a model obtained through 3D scanning often have more of these errors as 3D scanning is often achieved by point to point acquisition/mapping. 3D reconstruction often includes errors. Once completed, the STL file needs to be processed by a piece of software called a "slicer", which converts the model into a series of thin layers and produces a G-code file containing instructions tailored to a specific type of 3D printer (FDM printers). This G-code file can then be printed with 3D printing client software (which loads the G-code and uses it to instruct the 3D printer during the 3D printing process). Printer resolution describes layer thickness and X–Y resolution in dots per inch (dpi) or micrometers (μm). Typical layer thickness is around , although some machines can print layers as thin as . X–Y resolution is comparable to that of laser printers. The particles (3D dots) are around in diameter. For that printer resolution, specifying a mesh resolution of and a chord length generates an optimal STL output file for a given model input file. Specifying higher resolution results in larger files without increase in print quality. Construction of a model with contemporary methods can take anywhere from several hours to several days, depending on the method used and the size and complexity of the model. Additive systems can typically reduce this time to a few hours, although it varies widely depending on the type of machine used and the size and number of models being produced simultaneously. Finishing Though the printer-produced resolution and surface finish are sufficient for some applications, post-processing and finishing methods allow for benefits such as greater dimensional accuracy, smoother surfaces, and other modifications such as coloration. The surface finish of a 3D-printed part can improved using subtractive methods such as sanding and bead blasting. When smoothing parts that require dimensional accuracy, it is important to take into account the volume of the material being removed. Some printable polymers, such as acrylonitrile butadiene styrene (ABS), allow the surface finish to be smoothed and improved using chemical vapor processes based on acetone or similar solvents. Some additive manufacturing techniques can benefit from annealing as a post-processing step. Annealing a 3D-printed part allows for better internal layer bonding due to recrystallization of the part. It allows for an increase in mechanical properties, some of which are fracture toughness, flexural strength, impact resistance, and heat resistance. Annealing a component may not be suitable for applications where dimensional accuracy is required, as it can introduce warpage or shrinkage due to heating and cooling. Additive or subtractive hybrid manufacturing (ASHM) is a method that involves producing a 3D printed part and using machining (subtractive manufacturing) to remove material. Machining operations can be completed after each layer, or after the entire 3D print has been completed depending on the application requirements. These hybrid methods allow for 3D-printed parts to achieve better surface finishes and dimensional accuracy. The layered structure of traditional additive manufacturing processes leads to a stair-stepping effect on part-surfaces that are curved or tilted with respect to the building platform. The effect strongly depends on the layer height used, as well as the orientation of a part surface inside the building process. This effect can be minimized using "variable layer heights" or "adaptive layer heights". These methods decrease the layer height in places where higher quality is needed. Painting a 3D-printed part offers a range of finishes and appearances that may not be achievable through most 3D printing techniques. The process typically involves several steps, such as surface preparation, priming, and painting. These steps help prepare the surface of the part and ensuring the paint adheres properly. Some additive manufacturing techniques are capable of using multiple materials simultaneously. These techniques are able to print in multiple colors and color combinations simultaneously and can produce parts that may not necessarily require painting. Some printing techniques require internal supports to be built to support overhanging features during construction. These supports must be mechanically removed or dissolved if using a water-soluble support material such as PVA after completing a print. Some commercial metal 3D printers involve cutting the metal component off the metal substrate after deposition. A new process for the GMAW 3D printing allows for substrate surface modifications to remove aluminium or steel. Materials Traditionally, 3D printing focused on polymers for printing, due to the ease of manufacturing and handling polymeric materials. However, the method has rapidly evolved to not only print various polymers but also metals and ceramics, making 3D printing a versatile option for manufacturing. Layer-by-layer fabrication of three-dimensional physical models is a modern concept that "stems from the ever-growing CAD industry, more specifically the solid modeling side of CAD. Before solid modeling was introduced in the late 1980s, three-dimensional models were created with wire frames and surfaces." but in all cases the layers of materials are controlled by the printer and the material properties. The three-dimensional material layer is controlled by the deposition rate as set by the printer operator and stored in a computer file. The earliest printed patented material was a hot melt type ink for printing patterns using a heated metal alloy. Charles Hull filed the first patent on August 8, 1984, to use a UV-cured acrylic resin using a UV-masked light source at UVP Corp to build a simple model. The SLA-1 was the first SL product announced by 3D Systems at Autofact Exposition, Detroit, November 1978. The SLA-1 Beta shipped in Jan 1988 to Baxter Healthcare, Pratt and Whitney, General Motors and AMP. The first production SLA-1 shipped to Precision Castparts in April 1988. The UV resin material changed over quickly to an epoxy-based material resin. In both cases, SLA-1 models needed UV oven curing after being rinsed in a solvent cleaner to remove uncured boundary resin. A post cure apparatus (PCA) was sold with all systems. The early resin printers required a blade to move fresh resin over the model on each layer. The layer thickness was 0.006 inches and the HeCd laser model of the SLA-1 was 12 watts and swept across the surface at 30 in per second. UVP was acquired by 3D Systems in January 1990. A review of the history shows that a number of materials (resins, plastic powder, plastic filament and hot-melt plastic ink) were used in the 1980s for patents in the rapid prototyping field. Masked lamp UV-cured resin was also introduced by Cubital's Itzchak Pomerantz in the Soldier 5600, Carl Deckard's (DTM) laser sintered thermoplastic powders, and adhesive-laser cut paper (LOM) stacked to form objects by Michael Feygin before 3D Systems made its first announcement. Scott Crump was also working with extruded "melted" plastic filament modeling (FDM) and drop deposition had been patented by William E Masters a week after Hull's patent in 1984, but he had to discover thermoplastic inkjets, introduced by Visual Impact Corporation 3D printer in 1992, using inkjets from Howtek, Inc., before he formed BPM to bring out his own 3D printer product in 1994. Multi-material 3D printing Efforts to achieve multi-material 3D printing range from enhanced FDM-like processes like VoxelJet to novel voxel-based printing technologies like layered assembly. A drawback of many existing 3D printing technologies is that they only allow one material to be printed at a time, limiting many potential applications that require the integration of different materials in the same object. Multi-material 3D printing solves this problem by allowing objects of complex and heterogeneous arrangements of materials to be manufactured using a single printer. Here, a material must be specified for each voxel (or 3D printing pixel element) inside the final object volume. The process can be fraught with complications, however, due to the isolated and monolithic algorithms. Some commercial devices have sought to solve these issues, such as building a Spec2Fab translator, but the progress is still very limited. Nonetheless, in the medical industry, a concept of 3D-printed pills and vaccines has been presented. With this new concept, multiple medications can be combined, which is expected to decrease many risks. With more and more applications of multi-material 3D printing, the costs of daily life and high technology development will become inevitably lower. Metallographic materials of 3D printing is also being researched. By classifying each material, CIMP-3D can systematically perform 3D printing with multiple materials. 4D printing Using 3D printing and multi-material structures in additive manufacturing has allowed for the design and creation of what is called 4D printing. 4D printing is an additive manufacturing process in which the printed object changes shape with time, temperature, or some other type of stimulation. 4D printing allows for the creation of dynamic structures with adjustable shapes, properties or functionality. The smart/stimulus-responsive materials that are created using 4D printing can be activated to create calculated responses such as self-assembly, self-repair, multi-functionality, reconfiguration and shape-shifting. This allows for customized printing of shape-changing and shape-memory materials. 4D printing has the potential to find new applications and uses for materials (plastics, composites, metals, etc.) and has the potential to create new alloys and composites that were not viable before. The versatility of this technology and materials can lead to advances in multiple fields of industry, including space, commercial and medical fields. The repeatability, precision, and material range for 4D printing must increase to allow the process to become more practical throughout these industries.  To become a viable industrial production option, there are a few challenges that 4D printing must overcome. The challenges of 4D printing include the fact that the microstructures of these printed smart materials must be close to or better than the parts obtained through traditional machining processes. New and customizable materials need to be developed that have the ability to consistently respond to varying external stimuli and change to their desired shape. There is also a need to design new software for the various technique types of 4D printing. The 4D printing software will need to take into consideration the base smart material, printing technique, and structural and geometric requirements of the design. Processes and printers ISO/ASTM52900-15 defines seven categories of additive manufacturing (AM) processes within its meaning. They are: Vat photopolymerization Material jetting Binder jetting Powder bed fusion Material extrusion Directed energy deposition Sheet lamination The main differences between processes are in the way layers are deposited to create parts and in the materials that are used. Each method has its own advantages and drawbacks, which is why some companies offer a choice of powder and polymer for the material used to build the object. Others sometimes use standard, off-the-shelf business paper as the build material to produce a durable prototype. The main considerations in choosing a machine are generally speed, costs of the 3D printer, of the printed prototype, choice and cost of the materials, and color capabilities. Printers that work directly with metals are generally expensive. However, less expensive printers can be used to make a mold, which is then used to make metal parts. Material jetting The first process where three-dimensional material is deposited to form an object was done with material jetting or as it was originally called particle deposition. Particle deposition by inkjet first started with continuous inkjet technology (CIT) (1950s) and later with drop-on-demand inkjet technology (1970s) using hot-melt inks. Wax inks were the first three-dimensional materials jetted and later low-temperature alloy metal was jetted with CIT. Wax and thermoplastic hot melts were jetted next by DOD. Objects were very small and started with text characters and numerals for signage. An object must have form and can be handled. Wax characters tumbled off paper documents and inspired a liquid metal recorder patent to make metal characters for signage in 1971. Thermoplastic color inks (CMYK) were printed with layers of each color to form the first digitally formed layered objects in 1984. The idea of investment casting with Solid-Ink jetted images or patterns in 1984 led to the first patent to form articles from particle deposition in 1989, issued in 1992. Material extrusion Some methods melt or soften the material to produce the layers. In fused filament fabrication, also known as fused deposition modeling (FDM), the model or part is produced by extruding small beads or streams of material that harden immediately to form layers. A filament of thermoplastic, metal wire, or other material is fed into an extrusion nozzle head (3D printer extruder), which heats the material and turns the flow on and off. FDM is somewhat restricted in the variation of shapes that may be fabricated. Another technique fuses parts of the layer and then moves upward in the working area, adding another layer of granules and repeating the process until the piece has built up. This process uses the unfused media to support overhangs and thin walls in the part being produced, which reduces the need for temporary auxiliary supports for the piece. Recently, FFF/FDM has expanded to 3-D print directly from pellets to avoid the conversion to filament. This process is called fused particle fabrication (FPF) (or fused granular fabrication (FGF) and has the potential to use more recycled materials. Powder bed fusion Powder bed fusion techniques, or PBF, include several processes such as DMLS, SLS, SLM, MJF and EBM. Powder bed fusion processes can be used with an array of materials and their flexibility allows for geometrically complex structures, making it a good choice for many 3D printing projects. These techniques include selective laser sintering, with both metals and polymers and direct metal laser sintering. Selective laser melting does not use sintering for the fusion of powder granules but will completely melt the powder using a high-energy laser to create fully dense materials in a layer-wise method that has mechanical properties similar to those of conventional manufactured metals. Electron beam melting is a similar type of additive manufacturing technology for metal parts (e.g. titanium alloys). EBM manufactures parts by melting metal powder layer by layer with an electron beam in a high vacuum. Another method consists of an inkjet 3D printing system, which creates the model one layer at a time by spreading a layer of powder (plaster or resins) and printing a binder in the cross-section of the part using an inkjet-like process. With laminated object manufacturing, thin layers are cut to shape and joined. In addition to the previously mentioned methods, HP has developed the Multi Jet Fusion (MJF) which is a powder base technique, though no lasers are involved. An inkjet array applies fusing and detailing agents which are then combined by heating to create a solid layer. Binder jetting The binder jetting 3D printing technique involves the deposition of a binding adhesive agent onto layers of material, usually powdered, and then this "green" state part may be cured and even sintered. The materials can be ceramic-based, metal or plastic. This method is also known as inkjet 3D printing. To produce a part, the printer builds the model using a head that moves over the platform base to spread or deposit alternating layers of powder (plaster and resins) and binder. Most modern binder jet printers also cure each layer of binder. These steps are repeated until all layers have been printed. This green part is usually cured in an oven to off-gas most of the binder before being sintered in a kiln with a specific time-temperature curve for the given material(s). This technology allows the printing of full-color prototypes, overhangs, and elastomer parts. The strength of bonded powder prints can be enhanced by impregnating in the spaces between the necked or sintered matrix of powder with other compatible materials depending on the powder material, like wax, thermoset polymer, or even bronze. Stereolithography Other methods cure liquid materials using different sophisticated technologies, such as stereolithography. Photopolymerization is primarily used in stereolithography to produce a solid part from a liquid. Inkjet printer systems like the Objet PolyJet system spray photopolymer materials onto a build tray in ultra-thin layers (between 16 and 30 μm) until the part is completed. Each photopolymer layer is cured with UV light after it is jetted, producing fully cured models that can be handled and used immediately, without post-curing. Ultra-small features can be made with the 3D micro-fabrication technique used in multiphoton photopolymerisation. Due to the nonlinear nature of photo excitation, the gel is cured to a solid only in the places where the laser was focused while the remaining gel is then washed away. Feature sizes of under 100 nm are easily produced, as well as complex structures with moving and interlocked parts. Yet another approach uses a synthetic resin that is solidified using LEDs. In Mask-image-projection-based stereolithography, a 3D digital model is sliced by a set of horizontal planes. Each slice is converted into a two-dimensional mask image. The mask image is then projected onto a photocurable liquid resin surface and light is projected onto the resin to cure it in the shape of the layer. Continuous liquid interface production begins with a pool of liquid photopolymer resin. Part of the pool bottom is transparent to ultraviolet light (the "window"), which causes the resin to solidify. The object rises slowly enough to allow the resin to flow under and maintain contact with the bottom of the object. In powder-fed directed-energy deposition, a high-power laser is used to melt metal powder supplied to the focus of the laser beam. The powder-fed directed energy process is similar to selective laser sintering, but the metal powder is applied only where material is being added to the part at that moment. Computed axial lithography Computed axial lithography is a method for 3D printing based on computerised tomography scans to create prints in photo-curable resin. It was developed by a collaboration between the University of California, Berkeley with Lawrence Livermore National Laboratory. Unlike other methods of 3D printing it does not build models through depositing layers of material like fused deposition modelling and stereolithography, instead it creates objects using a series of 2D images projected onto a cylinder of resin. It is notable for its ability to build an object much more quickly than other methods using resins and the ability to embed objects within the prints. Liquid additive manufacturing Liquid additive manufacturing (LAM) is a 3D printing technique that deposits a liquid or high viscose material (e.g. liquid silicone rubber) onto a build surface to create an object which then is vulcanised using heat to harden the object. The process was originally created by Adrian Bowyer and was then built upon by German RepRap. A technique called programmable tooling uses 3D printing to create a temporary mold, which is then filled via a conventional injection molding process and then immediately dissolved. Lamination In some printers, paper can be used as the build material, resulting in a lower cost to print. During the 1990s some companies marketed printers that cut cross-sections out of special adhesive coated paper using a carbon dioxide laser and then laminated them together. In 2005 Mcor Technologies Ltd developed a different process using ordinary sheets of office paper, a tungsten carbide blade to cut the shape, and selective deposition of adhesive and pressure to bond the prototype. Directed-energy deposition (DED) Powder-fed directed-energy deposition In powder-fed directed-energy deposition (also known as laser metal deposition), a high-power laser is used to melt metal powder supplied to the focus of the laser beam. The laser beam typically travels through the center of the deposition head and is focused on a small spot by one or more lenses. The build occurs on an X-Y table which is driven by a tool path created from a digital model to fabricate an object layer by layer. The deposition head is moved up vertically as each layer is completed. Some systems even make use of 5-axis or 6-axis systems (i.e. articulated arms) capable of delivering material on the substrate (a printing bed, or a pre-existing part) with few to no spatial access restrictions. Metal powder is delivered and distributed around the circumference of the head or can be split by an internal manifold and delivered through nozzles arranged in various configurations around the deposition head. A hermetically sealed chamber filled with inert gas or a local inert shroud gas (sometimes both combined) is often used to shield the melt pool from atmospheric oxygen, to limit oxidation and to better control the material properties. The powder-fed directed-energy process is similar to selective laser sintering, but the metal powder is projected only where the material is being added to the part at that moment. The laser beam is used to heat up and create a "melt pool" on the substrate, in which the new powder is injected quasi-simultaneously. The process supports a wide range of materials including titanium, stainless steel, aluminium, tungsten, and other specialty materials as well as composites and functionally graded materials. The process can not only fully build new metal parts but can also add material to existing parts for example for coatings, repair, and hybrid manufacturing applications. Laser engineered net shaping (LENS), which was developed by Sandia National Labs, is one example of the powder-fed directed-energy deposition process for 3D printing or restoring metal parts. Metal wire processes Laser-based wire-feed systems, such as laser metal deposition-wire (LMD-w), feed the wire through a nozzle that is melted by a laser using inert gas shielding in either an open environment (gas surrounding the laser) or in a sealed chamber. Electron beam freeform fabrication uses an electron beam heat source inside a vacuum chamber. It is also possible to use conventional gas metal arc welding attached to a 3D stage to 3-D print metals such as steel, bronze and aluminium. Low-cost open source RepRap-style 3-D printers have been outfitted with Arduino-based sensors and demonstrated reasonable metallurgical properties from conventional welding wire as feedstock. Selective powder deposition (SPD) In selective powder deposition, build and support powders are selectively deposited into a crucible, such that the build powder takes the shape of the desired object and support powder fills the rest of the volume in the crucible. Then an infill material is applied, such that it comes in contact with the build powder. Then the crucible is fired up in a kiln at the temperature above the melting point of the infill but below the melting points of the powders. When the infill melts, it soaks the build powder. But it does not soak the support powder, because the support powder is chosen to be such that it is not wettable by the infill. If at the firing temperature, the atoms of the infill material and the build powder are mutually defusable, such as in the case of copper powder and zinc infill, then the resulting material will be a uniform mixture of those atoms, in this case, bronze. But if the atoms are not mutually defusable, such as in the case of tungsten and copper at 1100 °C, then the resulting material will be a composite. To prevent shape distortion, the firing temperature must be below the solidus temperature of the resulting alloy. Cryogenic 3D printing Cryogenic 3D printing is a collection of techniques that forms solid structures by freezing liquid materials while they are deposited. As each liquid layer is applied, it is cooled by the low temperature of the previous layer and printing environment which results in solidification. Unlike other 3D printing techniques, cryogenic 3D printing requires a controlled printing environment. The ambient temperature must be below the material's freezing point to ensure the structure remains solid during manufacturing and the humidity must remain low to prevent frost formation between the application of layers. Materials typically include water and water-based solutions, such as brine, slurry, and hydrogels. Cryogenic 3D printing techniques include rapid freezing prototype (RFP), low-temperature deposition manufacturing (LDM), and freeze-form extrusion fabrication (FEF). Applications 3D printing or additive manufacturing has been used in manufacturing, medical, industry and sociocultural sectors (e.g. cultural heritage) to create successful commercial technology. More recently, 3D printing has also been used in the humanitarian and development sector to produce a range of medical items, prosthetics, spares and repairs. The earliest application of additive manufacturing was on the toolroom end of the manufacturing spectrum. For example, rapid prototyping was one of the earliest additive variants, and its mission was to reduce the lead time and cost of developing prototypes of new parts and devices, which was earlier only done with subtractive toolroom methods such as CNC milling, turning, and precision grinding. In the 2010s, additive manufacturing entered production to a much greater extent. Food Additive manufacturing of food is being developed by squeezing out food, layer by layer, into three-dimensional objects. A large variety of foods are appropriate candidates, such as chocolate and candy, and flat foods such as crackers, pasta, and pizza. NASA is looking into the technology in order to create 3D-printed food to limit food waste and to make food that is designed to fit an astronaut's dietary needs. In 2018, Italian bioengineer Giuseppe Scionti developed a technology allowing the production of fibrous plant-based meat analogues using a custom 3D bioprinter, mimicking meat texture and nutritional values. Fashion 3D printing has entered the world of clothing, with fashion designers experimenting with 3D-printed bikinis, shoes, and dresses. In commercial production, Nike used 3D printing to prototype and manufacture the 2012 Vapor Laser Talon football shoe for players of American football, and New Balance has 3D manufactured custom-fit shoes for athletes. 3D printing has come to the point where companies are printing consumer-grade eyewear with on-demand custom fit and styling (although they cannot print the lenses). On-demand customization of glasses is possible with rapid prototyping. Transportation In cars, trucks, and aircraft, additive manufacturing is beginning to transform both unibody and fuselage design and production, and powertrain design and production. For example, General Electric uses high-end 3D printers to build parts for turbines. Many of these systems are used for rapid prototyping before mass production methods are employed. Other prominent examples include: In early 2014, Swedish supercar manufacturer Koenigsegg announced the One:1, a supercar that utilizes many components that were 3D printed. Urbee is the first car produced using 3D printing (the bodywork and car windows were "printed"). In 2014, Local Motors debuted Strati, a functioning vehicle that was entirely 3D printed using ABS plastic and carbon fiber, except the powertrain. In May 2015 Airbus announced that its new Airbus A350 XWB included over 1000 components manufactured by 3D printing. In 2015, a Royal Air Force Eurofighter Typhoon fighter jet flew with printed parts. The United States Air Force has begun to work with 3D printers, and the Israeli Air Force has also purchased a 3D printer to print spare parts. In 2017, GE Aviation revealed that it had used design for additive manufacturing to create a helicopter engine with 16 parts instead of 900, with great potential impact on reducing the complexity of supply chains. Firearms AM's impact on firearms involves two dimensions: new manufacturing methods for established companies, and new possibilities for the making of do-it-yourself firearms. In 2012, the US-based group Defense Distributed disclosed plans to design a working plastic 3D-printed firearm "that could be downloaded and reproduced by anybody with a 3D printer". After Defense Distributed released their plans, questions were raised regarding the effects that 3D printing and widespread consumer-level CNC machining may have on gun control effectiveness. Moreover, armor-design strategies can be enhanced by taking inspiration from nature and prototyping those designs easily, using AM. Health Surgical uses of 3D printing-centric therapies began in the mid-1990s with anatomical modeling for bony reconstructive surgery planning. Patient-matched implants were a natural extension of this work, leading to truly personalized implants that fit one unique individual. Virtual planning of surgery and guidance using 3D printed, personalized instruments have been applied to many areas of surgery including total joint replacement and craniomaxillofacial reconstruction with great success. One example of this is the bioresorbable trachial splint to treat newborns with tracheobronchomalacia developed at the University of Michigan. The use of additive manufacturing for serialized production of orthopedic implants (metals) is also increasing due to the ability to efficiently create porous surface structures that facilitate osseointegration. The hearing aid and dental industries are expected to be the biggest areas of future development using custom 3D printing technology. 3D printing is not just limited to inorganic materials; there have been a number of biomedical advancements made possible by 3D printing. , 3D bio-printing technology has been studied by biotechnology firms and academia for possible use in tissue engineering applications in which organs and body parts are built using inkjet printing techniques. In this process, layers of living cells are deposited onto a gel medium or sugar matrix and slowly built up to form three-dimensional structures including vascular systems. 3D printing has been considered as a method of implanting stem cells capable of generating new tissues and organs in living humans. In 2018, 3D printing technology was used for the first time to create a matrix for cell immobilization in fermentation. Propionic acid production by Propionibacterium acidipropionici immobilized on 3D-printed nylon beads was chosen as a model study. It was shown that those 3D-printed beads were capable of promoting high-density cell attachment and propionic acid production, which could be adapted to other fermentation bioprocesses. 3D printing has also been employed by researchers in the pharmaceutical field. During the last few years, there has been a surge in academic interest regarding drug delivery with the aid of AM techniques. This technology offers a unique way for materials to be utilized in novel formulations. AM manufacturing allows for the usage of materials and compounds in the development of formulations, in ways that are not possible with conventional/traditional techniques in the pharmaceutical field, e.g. tableting, cast-molding, etc. Moreover, one of the major advantages of 3D printing, especially in the case of fused deposition modelling (FDM), is the personalization of the dosage form that can be achieved, thus, targeting the patient's specific needs. In the not-so-distant future, 3D printers are expected to reach hospitals and pharmacies in order to provide on-demand production of personalized formulations according to the patients' needs. 3D printing has also been used for medical equipment. During the COVID-19 pandemic 3D printers were used to supplement the strained supply of PPE through volunteers using their personally owned printers to produce various pieces of personal protective equipment (i.e. frames for face shields). Education 3D printing, and open source 3D printers, in particular, are the latest technologies making inroads into the classroom. Higher education has proven to be a major buyer of desktop and professional 3D printers which industry experts generally view as a positive indicator. Some authors have claimed that 3D printers offer an unprecedented "revolution" in STEM education. The evidence for such claims comes from both the low-cost ability for rapid prototyping in the classroom by students, but also the fabrication of low-cost high-quality scientific equipment from open hardware designs forming open-source labs. Additionally, Libraries around the world have also become locations to house smaller 3D printers for educational and community access. Future applications for 3D printing might include creating open-source scientific equipment. Replicating archeological artifacts In the 2010s, 3D printing became intensively used in the cultural heritage field for preservation, restoration and dissemination purposes. Many Europeans and North American Museums have purchased 3D printers and actively recreate missing pieces of their relics and archaeological monuments such as Tiwanaku in Bolivia. The Metropolitan Museum of Art and the British Museum have started using their 3D printers to create museum souvenirs that are available in the museum shops. Other museums, like the National Museum of Military History and Varna Historical Museum, have gone further and sell through the online platform Threeding digital models of their artifacts, created using Artec 3D scanners, in 3D printing friendly file format, which everyone can 3D print at home. Morehshin Allahyari, an Iranian-born U.S. artist, considers her use of 3D sculpting processes of re-constructing Iranian cultural treasures as feminist activism. Allahyari uses a 3D modeling software to reconstruct a series of cultural artifacts that were demolished by ISIS militants in 2014. Replicating historic buildings and architectural structures The application of 3D printing for the representation of architectural assets has many challenges. In 2018, the structure of Iran National Bank was traditionally surveyed and modeled in computer graphics software (specifically, Cinema4D) and was optimized for 3D printing. The team tested the technique for the construction of the part and it was successful. After testing the procedure, the modellers reconstructed the structure in Cinema4D and exported the front part of the model to Netfabb. The entrance of the building was chosen due to the 3D printing limitations and the budget of the project for producing the maquette. 3D printing was only one of the capabilities enabled by the produced 3D model of the bank, but due to the project's limited scope, the team did not continue modelling for the virtual representation or other applications. In 2021, Parsinejad et al. comprehensively compared the hand surveying method for 3D reconstruction ready for 3D printing with digital recording (adoption of photogrammetry method). The world's first 3D-printed steel bridge was unveiled in Amsterdam in July 2021. Spanning 12 meters over the Oudezijds Achterburgwal canal, the bridge was created using robotic arms that printed over 4,500 kilograms of stainless steel. It took six months to complete. Soft actuators 3D printed soft actuators is a growing application of 3D printing technology that has found its place in the 3D printing applications. These soft actuators are being developed to deal with soft structures and organs, especially in biomedical sectors and where the interaction between humans and robots is inevitable. The majority of the existing soft actuators are fabricated by conventional methods that require manual fabrication of devices, post-processing/assembly, and lengthy iterations until the maturity of the fabrication is achieved. Instead of the tedious and time-consuming aspects of the current fabrication processes, researchers are exploring an appropriate manufacturing approach for the effective fabrication of soft actuators. Thus, 3D-printed soft actuators are introduced to revolutionize the design and fabrication of soft actuators with custom geometrical, functional, and control properties in a faster and inexpensive approach. They also enable incorporation of all actuator components into a single structure eliminating the need to use external joints, adhesives, and fasteners. Circuit boards Circuit board manufacturing involves multiple steps which include imaging, drilling, plating, solder mask coating, nomenclature printing and surface finishes. These steps include many chemicals such as harsh solvents and acids. 3D printing circuit boards remove the need for many of these steps while still producing complex designs. Polymer ink is used to create the layers of the build while silver polymer is used for creating the traces and holes used to allow electricity to flow. Current circuit board manufacturing can be a tedious process depending on the design. Specified materials are gathered and sent into inner layer processing where images are printed, developed and etched. The etch cores are typically punched to add lamination tooling. The cores are then prepared for lamination. The stack-up, the buildup of a circuit board, is built and sent into lamination where the layers are bonded. The boards are then measured and drilled. Many steps may differ from this stage however for simple designs, the material goes through a plating process to plate the holes and surface. The outer image is then printed, developed and etched. After the image is defined, the material must get coated with a solder mask for later soldering. Nomenclature is then added so components can be identified later. Then the surface finish is added. The boards are routed out of panel form into their singular or array form and then electrically tested. Aside from the paperwork that must be completed which proves the boards meet specifications, the boards are then packed and shipped. The benefits of 3D printing would be that the final outline is defined from the beginning, no imaging, punching or lamination is required and electrical connections are made with the silver polymer which eliminates drilling and plating. The final paperwork would also be greatly reduced due to the lack of materials required to build the circuit board. Complex designs which may take weeks to complete through normal processing can be 3D printed, greatly reducing manufacturing time. Hobbyists In 2005, academic journals began to report on the possible artistic applications of 3D printing technology. Off-the-shelf machines were increasingly capable of producing practical household applications, for example, ornamental objects. Some practical examples include a working clock and gears printed for home woodworking machines among other purposes. Websites associated with home 3D printing tended to include backscratchers, coat hooks, door knobs, etc. As of 2017, domestic 3D printing was reaching a consumer audience beyond hobbyists and enthusiasts. Several projects and companies are making efforts to develop affordable 3D printers for home desktop use. Much of this work has been driven by and targeted at DIY/maker/enthusiast/early adopter communities, with additional ties to the academic and hacker communities. Sped on by decreases in price and increases in quality, an estimated 2 million people worldwide have purchased a 3D printer for hobby use. Legal aspects Intellectual property 3D printing has existed for decades within certain manufacturing industries where many legal regimes, including patents, industrial design rights, copyrights, and trademarks may apply. However, there is not much jurisprudence to say how these laws will apply if 3D printers become mainstream and individuals or hobbyist communities begin manufacturing items for personal use, for non-profit distribution, or for sale. Any of the mentioned legal regimes may prohibit the distribution of the designs used in 3D printing or the distribution or sale of the printed item. To be allowed to do these things, where active intellectual property was involved, a person would have to contact the owner and ask for a licence, which may come with conditions and a price. However, many patent, design and copyright laws contain a standard limitation or exception for "private" or "non-commercial" use of inventions, designs or works of art protected under intellectual property (IP). That standard limitation or exception may leave such private, non-commercial uses outside the scope of IP rights. Patents cover inventions including processes, machines, manufacturing, and compositions of matter and have a finite duration which varies between countries, but generally 20 years from the date of application. Therefore, if a type of wheel is patented, printing, using, or selling such a wheel could be an infringement of the patent. Copyright covers an expression in a tangible, fixed medium and often lasts for the life of the author plus 70 years thereafter. For example, a sculptor retains copyright over a statue, such that other people cannot then legally distribute designs to print an identical or similar statue without paying royalties, waiting for the copyright to expire, or working within a fair use exception. When a feature has both artistic (copyrightable) and functional (patentable) merits when the question has appeared in US court, the courts have often held the feature is not copyrightable unless it can be separated from the functional aspects of the item. In other countries the law and the courts may apply a different approach allowing, for example, the design of a useful device to be registered (as a whole) as an industrial design on the understanding that, in case of unauthorized copying, only the non-functional features may be claimed under design law whereas any technical features could only be claimed if covered by a valid patent. Gun legislation and administration The US Department of Homeland Security and the Joint Regional Intelligence Center released a memo stating that "significant advances in three-dimensional (3D) printing capabilities, availability of free digital 3D printable files for firearms components, and difficulty regulating file sharing may present public safety risks from unqualified gun seekers who obtain or manufacture 3D printed guns" and that "proposed legislation to ban 3D printing of weapons may deter, but cannot completely prevent their production. Even if the practice is prohibited by new legislation, online distribution of these 3D printable files will be as difficult to control as any other illegally traded music, movie or software files." Attempting to restrict the distribution of gun plans via the Internet has been likened to the futility of preventing the widespread distribution of DeCSS, which enabled DVD ripping. After the US government had Defense Distributed take down the plans, they were still widely available via the Pirate Bay and other file sharing sites. Downloads of the plans from the UK, Germany, Spain, and Brazil were heavy. Some US legislators have proposed regulations on 3D printers to prevent them from being used for printing guns. 3D printing advocates have suggested that such regulations would be futile, could cripple the 3D printing industry and could infringe on free speech rights, with early pioneers of 3D printing professor Hod Lipson suggesting that gunpowder could be controlled instead. Internationally, where gun controls are generally stricter than in the United States, some commentators have said the impact may be more strongly felt since alternative firearms are not as easily obtainable. Officials in the United Kingdom have noted that producing a 3D-printed gun would be illegal under their gun control laws. Europol stated that criminals have access to other sources of weapons but noted that as technology improves, the risks of an effect would increase. Aerospace regulation In the United States, the FAA has anticipated a desire to use additive manufacturing techniques and has been considering how best to regulate this process. The FAA has jurisdiction over such fabrication because all aircraft parts must be made under FAA production approval or under other FAA regulatory categories. In December 2016, the FAA approved the production of a 3D-printed fuel nozzle for the GE LEAP engine. Aviation attorney Jason Dickstein has suggested that additive manufacturing is merely a production method, and should be regulated like any other production method. He has suggested that the FAA's focus should be on guidance to explain compliance, rather than on changing the existing rules, and that existing regulations and guidance permit a company "to develop a robust quality system that adequately reflects regulatory needs for quality assurance". Health and safety Polymer feedstock materials can release ultrafine particles and volatile organic compounds (VOCs) if sufficiently heated, which in combination have been associated with adverse respiratory and cardiovascular health effects. In addition, temperatures of 190 °C to 260 °C are typically reached by an FFF extrusion nozzle, which can cause skin burns. Vat photopolymerization stereolithography printers use high-powered lasers that present a skin and eye hazard, although they are considered nonhazardous during printing because the laser is enclosed within the printing chamber. 3D printers also contain many moving parts that include stepper motors, pulleys, threaded rods, carriages, and small fans, which generally do not have enough power to cause serious injuries but can still trap a user's finger, long hair, or loose clothing. Most desktop FFF 3D printers do not have any added electrical safety features beyond regular internal fuses or external transformers, although the voltages in the exposed parts of 3D printers usually do not exceed 12V to 24V, which is generally considered safe. Research on the health and safety concerns of 3D printing is new and in development due to the recent proliferation of 3D printing devices. In 2017, the European Agency for Safety and Health at Work published a discussion paper on the processes and materials involved in 3D printing, the potential implications of this technology for occupational safety and health and avenues for controlling potential hazards. Noise level is measured in decibels (dB), and can vary greatly in home printers from 15 dB to 75 dB. Some main sources of noise in filament printers are fans, motors and bearings, while in resin printers the fans usually are responsible for most of the noise. Some methods for dampening the noise from a printer may be to install vibration isolation, use larger diameter fans, perform regular maintenance and lubrication, or use a soundproofing enclosure. Impact Additive manufacturing, starting with today's infancy period, requires manufacturing firms to be flexible, ever-improving users of all available technologies to remain competitive. Advocates of additive manufacturing also predict that this arc of technological development will counter globalization, as end users will do much of their own manufacturing rather than engage in trade to buy products from other people and corporations. The real integration of the newer additive technologies into commercial production, however, is more a matter of complementing traditional subtractive methods rather than displacing them entirely. The futurologist Jeremy Rifkin claimed that 3D printing signals the beginning of a third industrial revolution, succeeding the production line assembly that dominated manufacturing starting in the late 19th century. Social change Since the 1950s, a number of writers and social commentators have speculated in some depth about the social and cultural changes that might result from the advent of commercially affordable additive manufacturing technology. In recent years, 3D printing has created a significant impact in the humanitarian and development sector. Its potential to facilitate distributed manufacturing is resulting in supply chain and logistics benefits, by reducing the need for transportation, warehousing and wastage. Furthermore, social and economic development is being advanced through the creation of local production economies. Others have suggested that as more and more 3D printers start to enter people's homes, the conventional relationship between the home and the workplace might get further eroded. Likewise, it has also been suggested that, as it becomes easier for businesses to transmit designs for new objects around the globe, so the need for high-speed freight services might also become less. Finally, given the ease with which certain objects can now be replicated, it remains to be seen whether changes will be made to current copyright legislation so as to protect intellectual property rights with the new technology widely available. Some call attention to the conjunction of commons-based peer production with 3D printing and other low-cost manufacturing techniques. The self-reinforced fantasy of a system of eternal growth can be overcome with the development of economies of scope, and here, society can play an important role contributing to the raising of the whole productive structure to a higher plateau of more sustainable and customized productivity. Further, it is true that many issues, problems, and threats arise due to the democratization of the means of production, and especially regarding the physical ones. For instance, the recyclability of advanced nanomaterials is still questioned; weapons manufacturing could become easier; not to mention the implications for counterfeiting and on intellectual property. It might be maintained that in contrast to the industrial paradigm whose competitive dynamics were about economies of scale, commons-based peer production 3D printing could develop economies of scope. While the advantages of scale rest on cheap global transportation, the economies of scope share infrastructure costs (intangible and tangible productive resources), taking advantage of the capabilities of the fabrication tools. And following Neil Gershenfeld in that "some of the least developed parts of the world need some of the most advanced technologies", commons-based peer production and 3D printing may offer the necessary tools for thinking globally but acting locally in response to certain needs. Larry Summers wrote about the "devastating consequences" of 3D printing and other technologies (robots, artificial intelligence, etc.) for those who perform routine tasks. In his view, "already there are more American men on disability insurance than doing production work in manufacturing. And the trends are all in the wrong direction, particularly for the less skilled, as the capacity of capital embodying artificial intelligence to replace white-collar as well as blue-collar work will increase rapidly in the years ahead." Summers recommends more vigorous cooperative efforts to address the "myriad devices" (e.g., tax havens, bank secrecy, money laundering, and regulatory arbitrage) enabling the holders of great wealth to "a paying" income and estate taxes, and to make it more difficult to accumulate great fortunes without requiring "great social contributions" in return, including: more vigorous enforcement of anti-monopoly laws, reductions in "excessive" protection for intellectual property, greater encouragement of profit-sharing schemes that may benefit workers and give them a stake in wealth accumulation, strengthening of collective bargaining arrangements, improvements in corporate governance, strengthening of financial regulation to eliminate subsidies to financial activity, easing of land-use restrictions that may cause the real estate of the rich to keep rising in value, better training for young people and retraining for displaced workers, and increased public and private investment in infrastructure development—e.g., in energy production and transportation. Michael Spence wrote that "Now comes a ... powerful, wave of digital technology that is replacing labor in increasingly complex tasks. This process of labor substitution and disintermediation has been underway for some time in service sectors—think of ATMs, online banking, enterprise resource planning, customer relationship management, mobile payment systems, and much more. This revolution is spreading to the production of goods, where robots and 3D printing are displacing labor." In his view, the vast majority of the cost of digital technologies comes at the start, in the design of hardware (e.g. 3D printers) and, more importantly, in creating the software that enables machines to carry out various tasks. "Once this is achieved, the marginal cost of the hardware is relatively low (and declines as scale rises), and the marginal cost of replicating the software is essentially zero. With a huge potential global market to amortize the upfront fixed costs of design and testing, the incentives to invest [in digital technologies] are compelling." Spence believes that, unlike prior digital technologies, which drove firms to deploy underutilized pools of valuable labor around the world, the motivating force in the current wave of digital technologies "is cost reduction via the replacement of labor". For example, as the cost of 3D printing technology declines, it is "easy to imagine" that production may become "extremely" local and customized. Moreover, production may occur in response to actual demand, not anticipated or forecast demand. Spence believes that labor, no matter how inexpensive, will become a less important asset for growth and employment expansion, with labor-intensive, process-oriented manufacturing becoming less effective, and that re-localization will appear in both developed and developing countries. In his view, production will not disappear, but it will be less labor-intensive, and all countries will eventually need to rebuild their growth models around digital technologies and the human capital supporting their deployment and expansion. Spence writes that "the world we are entering is one in which the most powerful global flows will be ideas and digital capital, not goods, services, and traditional capital. Adapting to this will require shifts in mindsets, policies, investments (especially in human capital), and quite possibly models of employment and distribution." Naomi Wu regards the usage of 3D printing in the Chinese classroom (where rote memorization is standard) to teach design principles and creativity as the most exciting recent development of the technology, and more generally regards 3D printing as being the next desktop publishing revolution. A printer was donated to the Juan Fernandez Women's Group in 2024, to support women in the remote community to be able to create parts to fix broken equipment, without having to wait for a ship to import the needed compenents. Environmental change The growth of additive manufacturing could have a large impact on the environment. Traditional subtractive manufacturing methods such as CNC milling create products by cutting away material from a larger block. In contrast, additive manufacturing creates products layer-by layer, using the minimum required materials to create the product. This has the benefit of reducing material waste, which further contributes to energy savings by avoiding raw material production. Life-cycle assessment of additive manufacturing has estimated that adopting the technology could further lower carbon dioxide emissions since 3D printing creates localized production, thus reducing the need to transport products and the emissions associated. AM could also allow consumers to create their own replacement parts to fix purchased products to extend the lifespan of purchased products. By making only the bare structural necessities of products, additive manufacturing also has the potential to make profound contributions to lightweighting. The use of these lightweight components would allow for reductions in the energy consumption and greenhouse gas emissions of vehicles and other forms of transportation. A case study on an airplane component made using additive manufacturing, for example, found that the use of the component saves 63% of relevant energy and carbon dioxide emissions over the course of the product's lifetime. However, the adoption of additive manufacturing also has environmental disadvantages. Firstly, AM has a high energy consumption compared to traditional processes. This is due to its use of processes such as lasers and high temperatures for product creation. Secondly, despite additive manufacturing reducing up to 90% of waste compared to subtractive manufacturing, AM can generate waste that is non-recyclable. For example, there are issues with the recyclability of materials in metal AM as some highly regulated industries such as aerospace often insist on using virgin powder in the creation of safety critical components. Additive manufacturing has not yet reached its theoretical material efficiency potential of 97%, but it may get closer as the technology continues to increase productivity. Despite the drawbacks, research and industry are making further strides to support AM's sustainability. Some large FDM printers that melt high-density polyethylene (HDPE) pellets may also accept sufficiently clean recycled material such as chipped milk bottles. In addition, these printers can use shredded material from faulty builds or unsuccessful prototype versions, thus reducing overall project wastage and materials handling and storage. The concept has been explored in the RecycleBot. There are also industrial efforts to produce metal powder from recycled metals.
Technology
Industry: General
null
1306157
https://en.wikipedia.org/wiki/Launch%20vehicle
Launch vehicle
A launch vehicle is typically a rocket-powered vehicle designed to carry a payload (a crewed spacecraft or satellites) from Earth's surface or lower atmosphere to outer space. The most common form is the ballistic missile-shaped multistage rocket, but the term is more general and also encompasses vehicles like the Space Shuttle. Most launch vehicles operate from a launch pad, supported by a launch control center and systems such as vehicle assembly and fueling. Launch vehicles are engineered with advanced aerodynamics and technologies, which contribute to high operating costs. An orbital launch vehicle must lift its payload at least to the boundary of space, approximately and accelerate it to a horizontal velocity of at least . Suborbital vehicles launch their payloads to lower velocity or are launched at elevation angles greater than horizontal. Practical orbital launch vehicles use chemical propellants such as solid fuel, liquid hydrogen, kerosene, liquid oxygen, or hypergolic propellants. Launch vehicles are classified by their orbital payload capacity, ranging from small-, medium-, heavy- to super-heavy lift. History Mass to orbit Launch vehicles are classed by NASA according to low Earth orbit payload capability: Small-lift launch vehicle: < - e.g. Vega Medium-lift launch vehicle: - e.g. Soyuz ST Heavy-lift launch vehicle: > - e.g. Ariane 5 Super-heavy lift vehicle: > - e.g. Saturn V Sounding rockets are similar to small-lift launch vehicles, however they are usually even smaller and do not place payloads into orbit. A modified SS-520 sounding rocket was used to place a 4-kilogram payload (TRICOM-1R) into orbit in 2018. General information Orbital spaceflight requires a satellite or spacecraft payload to be accelerated to very high velocity. In the vacuum of space, reaction forces must be provided by the ejection of mass, resulting in the rocket equation. The physics of spaceflight are such that rocket stages are typically required to achieve the desired orbit. Expendable launch vehicles are designed for one-time use, with boosters that usually separate from their payload and disintegrate during atmospheric reentry or on contact with the ground. In contrast, reusable launch vehicles are designed to be recovered intact and launched again. The Falcon 9 is an example of a reusable launch vehicle. As of 2023, all reusable launch vehicles that were ever operational have been partially reusable, meaning some components are recovered and others are not. This usually means the recovery of specific stages, usually just the first stage, but sometimes specific components of a rocket stage may be recovered while others are not. The Space Shuttle, for example, recovered and reused its solid rocket boosters, the Space Shuttle orbiter that also acted as a second stage, and the engines used by the core stage (the RS-25, which was located at the back of the orbiter), however the fuel tank that the engines sourced fuel from, which was separate from the engines, was not reused. For example, the European Space Agency is responsible for the Ariane V, and the United Launch Alliance manufactures and launches the Delta IV and Atlas V rockets. Launch platform locations Launchpads can be located on land (spaceport), on a fixed ocean platform (San Marco), on a mobile ocean platform (Sea Launch), and on a submarine. Launch vehicles can also be launched from the air. Flight regimes A launch vehicle will start off with its payload at some location on the surface of the Earth. To reach orbit, the vehicle must travel vertically to leave the atmosphere and horizontally to prevent re-contacting the ground. The required velocity varies depending on the orbit but will always be extreme when compared to velocities encountered in normal life. Launch vehicles provide varying degrees of performance. For example, a satellite bound for Geostationary orbit (GEO) can either be directly inserted by the upper stage of the launch vehicle or launched to a geostationary transfer orbit (GTO). A direct insertion places greater demands on the launch vehicle, while GTO is more demanding of the spacecraft. Once in orbit, launch vehicle upper stages and satellites can have overlapping capabilities, although upper stages tend to have orbital lifetimes measured in hours or days while spacecraft can last decades. Distributed launch Distributed launch involves the accomplishment of a goal with multiple spacecraft launches. A large spacecraft such as the International Space Station can be constructed by assembling modules in orbit, or in-space propellant transfer conducted to greatly increase the delta-V capabilities of a cislunar or deep space vehicle. Distributed launch enables space missions that are not possible with single launch architectures. Mission architectures for distributed launch were explored in the 2000s and launch vehicles with integrated distributed launch capability built in began development in 2017 with the Starship design. The standard Starship launch architecture is to refuel the spacecraft in low Earth orbit to enable the craft to send high-mass payloads on much more energetic missions. Return to launch site After 1980, but before the 2010s, two orbital launch vehicles developed the capability to return to the launch site (RTLS). Both the US Space Shuttle—with one of its abort modes—and the Soviet Buran had a designed-in capability to return a part of the launch vehicle to the launch site via the mechanism of horizontal-landing of the spaceplane portion of the launch vehicle. In both cases, the main vehicle thrust structure and the large propellant tank were expendable, as had been the standard procedure for all orbital launch vehicles flown prior to that time. Both were subsequently demonstrated on actual orbital nominal flights, although both also had an abort mode during launch that could conceivably allow the crew to land the spaceplane following an off-nominal launch. In the 2000s, both SpaceX and Blue Origin have privately developed a set of technologies to support vertical landing of the booster stage of a launch vehicle. After 2010, SpaceX undertook a development program to acquire the ability to bring back and vertically land a part of the Falcon 9 orbital launch vehicle: the first stage. The first successful landing was done in December 2015, since 2017 rocket stages routinely land either at a landing pad adjacent to the launch site or on a landing platform at sea, some distance away from the launch site. The Falcon Heavy is similarly designed to reuse the three cores comprising its first stage. On its first flight in February 2018, the two outer cores successfully returned to the launch site landing pads while the center core targeted the landing platform at sea but did not successfully land on it. Blue Origin developed similar technologies for bringing back and landing their suborbital New Shepard, and successfully demonstrated return in 2015, and successfully reused the same booster on a second suborbital flight in January 2016. By October 2016, Blue had reflown, and landed successfully, that same launch vehicle a total of five times. The launch trajectories of both vehicles are very different, with New Shepard going straight up and down, whereas Falcon 9 has to cancel substantial horizontal velocity and return from a significant distance downrange. Both Blue Origin and SpaceX also have additional reusable launch vehicles under development. Blue is developing the first stage of the orbital New Glenn LV to be reusable, with first flight planned for no earlier than 2024. SpaceX has a new super-heavy launch vehicle under development for missions to interplanetary space. The SpaceX Starship is designed to support RTLS, vertical-landing and full reuse of both the booster stage and the integrated second-stage/large-spacecraft that are designed for use with Starship. Its first launch attempt took place in April 2023; however, both stages were lost during ascent. The fifth launch attempt ended with Booster 12 being caught by the launch tower, and Ship 30, the upper stage, successfully landing in the Indian Ocean.
Technology
Basics_10
null
1306608
https://en.wikipedia.org/wiki/Triceps
Triceps
The triceps, or triceps brachii (Latin for "three-headed muscle of the arm"), is a large muscle on the back of the upper limb of many vertebrates. It consists of three parts: the medial, lateral, and long head. It is the muscle principally responsible for extension of the elbow joint (straightening of the arm). Structure The long head arises from the infraglenoid tubercle of the scapula. It extends distally anterior to the teres minor and posterior to the teres major. The medial head arises proximally in the humerus, just inferior to the groove of the radial nerve; from the dorsal (back) surface of the humerus; from the medial intermuscular septum; and its distal part also arises from the lateral intermuscular septum. The medial head is mostly covered by the lateral and long heads and is only visible distally on the humerus. The lateral head arises from the dorsal surface of the humerus, lateral and proximal to the groove of the radial nerve, from the greater tubercle down to the region of the lateral intermuscular septum. Each of the three fascicles has its own motorneuron subnucleus in the motor column in the spinal cord. The medial head is formed predominantly by small type I fibers and motor units, the lateral head of large type IIb fibers and motor units, and the long head of a mixture of fiber types and motor units. It has been suggested that each fascicle "may be considered an independent muscle with specific functional roles." The fibers converge to a single tendon to insert onto the olecranon process of the ulna (though some research indicates that there may be more than one tendon) and to the posterior wall of the capsule of the elbow joint where bursae (cushion sacks) are often found. Parts of the common tendon radiate into the fascia of the forearm and can almost cover the anconeus muscle. Innervation All three heads of the triceps brachii are classically believed to be innervated by the radial nerve. However, more recent studies observed that in around 14% of individuals, the long head of the triceps brachii was innervated by the axillary nerve, and in 3% it received dual innervation from both the radial nerve and axillary nerve. Variation A tendinous arch is frequently the origin of the long head and the tendon of latissimus dorsi. In rare cases, the long head can originate from the lateral margin of the scapula and from the capsule of the shoulder joint. Function The triceps is an extensor muscle of the elbow joint and an antagonist of the biceps and brachialis muscles. It can also fixate the elbow joint when the forearm and hand are used for fine movements, e.g., when writing. It has been suggested that the long head fascicle is employed when sustained force generation is demanded, or when there is a need for a synergistic control of the shoulder and elbow or both. The lateral head is used for movements requiring occasional high-intensity force, while the medial fascicle enables more precise, low-force movements. With its origin on the scapula, the long head also acts on the shoulder joint and is also involved in retroversion and adduction of the arm. It helps stabilise the shoulder joint at the top of the humerus. Training The triceps can be worked through either isolation or compound elbow extension movements and can contract statically to keep the arm straightened against resistance. Isolation movements include cable push-downs, lying triceps extensions, and arm extensions behind the back. Examples of compound elbow extension include pressing movements like the push up, bench press, close grip bench press (flat, incline or decline), military press and dips. A closer grip targets the triceps more than wider grip movements. Static contraction movements include pullovers, straight-arm pulldowns, and bent-over lateral raises, which are also used to build the deltoids and latissimus dorsi. Ruptures of the triceps muscle are rare, and typically only occur in anabolic steroid users. Clinical significance The triceps reflex, elicited by hitting the triceps, is often used to test the function of the nerves of the arm. This tests spinal nerves C6 and C7, predominantly C7. History Etymology It is sometimes called a three-headed muscle (Latin literally three-headed, tri - three, and ceps, from caput - head), because there are three bundles of muscles, each of different origins, joining at the elbow. Though a similarly named muscle, the triceps surae, is found on the lower leg, the triceps brachii is commonly called the triceps. Historically, the plural form of triceps was tricipites, a form not in general use today; instead, triceps is both singular and plural (i.e., when referring to both arms). Other animals In the horse, 84%, 15%, and 3% of the total triceps muscle weight correspond to the long, lateral, and medial heads, respectively. Many mammals, such as dogs, cattle, and pigs, have a fourth head, the accessory head. It lies between the lateral and medial heads. In humans, the anconeus is sometimes loosely called "the fourth head of the triceps brachii". Additional images
Biology and health sciences
Human anatomy
Health
1306907
https://en.wikipedia.org/wiki/American%20mink
American mink
The American mink (Neogale vison) is a semiaquatic species of mustelid native to North America, though human introduction has expanded its range to many parts of Europe, Asia, and South America. Because of range expansion, the American mink is classed as a least-concern species by the IUCN. The American mink was formerly thought to be the only extant member of the genus Neovison following the extinction of the sea mink (N. macrodon), but recent studies, followed by taxonomic authorities, have reclassified it and the sea mink within the genus Neogale, which also contains a few New World weasel species. The American mink is a carnivore that feeds on rodents, fish, crustaceans, frogs, and birds. In its introduced range in Europe it has been classified as an invasive species linked to declines in European mink, Pyrenean desman, and water vole populations. It is the animal most frequently farmed for its fur, exceeding the silver fox, sable, marten, and skunk in economic importance. Evolution As a species, the American mink represents a more specialized form than the European mink in the direction of carnivory, as indicated by the more developed structure of the skull. Fossil records of the American mink go back as far as the Irvingtonian, though the species is uncommon among Pleistocene animals. Its fossil range corresponds with the species' current natural range. The American minks of the Pleistocene did not differ much in size or morphology from modern populations, though a slight trend toward increased size is apparent from the Irvingtonian through to the Illinoian and Wisconsinan periods. Although superficially similar to the European mink, studies indicate the European mink's closest relative is the Siberian weasel (kolonok) of Asia. The American mink has been recorded to hybridize with European minks and polecats in captivity, though the hybrid embryos of the American and European minks are usually reabsorbed. Subspecies , 15 subspecies are recognised: Description Build The American mink differs from members of the genus Mustela (stoats and weasels), as well as the other members of Neogale, by its larger size and stouter form, which closely approach those of martens. It shares with martens a uniformly enlarged, bushy and somewhat tapering tail, rather than a slender, cylindrical tail with an enlarged bushy tip, as is the case in stoats. The American mink is similar in build to the European mink, but the tail is longer (constituting 38–51% of its body length). The American mink has a long body, which allows the species to enter the burrows of prey. Its streamlined shape helps it to reduce water resistance while swimming. The skull is similar to that of the European mink, but is more massive, narrower, and less elongated, with more strongly developed projections and a wider, shorter cranium. The upper molars are larger and more massive than those of the European mink. The dental formula is . Domestic mink, which are bred in fur farms and are substandard genetically, have 19.6% smaller brains, 8.1% smaller hearts, and 28.2% smaller spleens than wild mink. The feet are broad, with webbed digits. They generally have eight nipples, with one pair of inguinal teats and three pairs of abdominal teats. The adult male's penis is long, and is covered by a sheath. The baculum is well-developed, being triangular in cross section and curved at the tip. Males measure in body length, while females measure . The tail measures in males and in females. Weights vary with sex and season, with males being heavier than females. In winter, males weigh and females . Maximum heaviness occurs in autumn. Fur The American mink's winter fur is denser, longer, softer, and more close-fitting than that of the European mink. The winter fur's tone is generally very dark blackish-tawny to light-tawny. Colour is evenly distributed over all the body, with the under side being only slightly lighter than the back. The guard hairs are bright and dark-tawny, often approaching black on the spine. The underfur on the back is very wavy and greyish-tawny with a bluish tint. The tail is darker than the trunk and sometimes becomes pure black on the tip. The chin and lower lip are white. Captive individuals tend to develop irregular white patches on the lower surface of their bodies, though escaped individuals from Tartaria gradually lost these patches. The summer fur is generally shorter, sparser and duller than the winter fur. The thick underfur and oily guard hairs render the pelage water-resistant, with the length of the guard hairs being intermediate between those of otters and polecats, thus indicating the American mink is incompletely adapted to an aquatic life. It moults twice a year, during spring and autumn. It does not turn white in winter. A variety of different colour mutations have arisen from experimental breeding on fur farms. Locomotion On land, the American mink moves by a bounding gait, with speeds of up to . It also climbs trees and swims well. During swimming, the mink propels itself primarily through undulating movements of the trunk. When diving, it undergoes bradycardia, which is likely an adaptation to conserve oxygen. In warm water (), the American mink can swim for 3 hours without stopping, but in cold water it can die within 27 minutes. It generally dives to depths of for 10 seconds, though depths of 3 m lasting 60 seconds have been recorded. It typically catches fish after five- to 20-second chases. Senses and scent glands The American mink relies heavily on sight when foraging. Its eyesight is clearer on land than underwater. Its auditory perception is high enough to detect the ultrasonic vocalisations (1–16 kHz) of rodent prey. Its sense of smell is comparatively weak. Its two anal glands are used for scent marking, either through defecation or by rubbing the anal region on the ground. The secretions of the anal glands are composed of 2,2-dimethyl thietane, 2-ethylthietane, cyclic disulfide, 3,3-dimethyl-1,2-dithiacyclopentane, and indole. When stressed, the American mink can expel the contents of its anal glands at a distance of . Scent glands may also be located on the throat and chest. The smell produced by these scent glands was described by Clinton Hart Merriam as more unbearable than that produced by skunks, and added it was "one of the few substances, of animal, vegetable, or mineral origin, that has, on land or sea, rendered me aware of the existence of the abominable sensation called nausea". Behavior Social and territorial behaviours American mink territories are held by individual animals with minimal intrasex overlap, but with extensive overlap between animals of the opposite sex. Most territories are in undisturbed, rocky coastal habitats with broad littoral zones and dense cover. Some are on estuaries, rivers and canals near urban areas. Home ranges are typically long, with male territories larger than females'. As long as it is close to water, the American mink is not fussy about its choice of den. Mink dens typically consist of long burrows in river banks, holes under logs, tree stumps, or roots and hollow trees, though dens located in rock crevices, drains, and nooks under stone piles and bridges are occasionally selected. The burrows they dig themselves are typically about four inches in diameter and may continue along for at a depth of . The American mink may nest in burrows dug previously by muskrats, badgers and skunks, and may also dig dens in old ant hills. The nesting chamber is at the end of a four-inch tunnel, and is about a foot in diameter. It is warm, dry, and lined with straw and feathers. The American mink's dens are characterized by a large number of entrances and twisting passages. The number of exits varies from one to eight. The American mink normally only vocalises during close encounters with other minks or predators. The sounds it emits include piercing shrieks and hisses when threatened and muffled chuckling sounds when mating. Kits squeak repeatedly when separated from their mothers. Ernest Thompson Seton reported hearing minks growl and snarl when confronting a threat. During aggressive interactions, mink assert their dominance by arching their backs, puffing up, lashing out their tails, and stomping and scraping the ground with their feet, while also opening their mouth in a threat-gape. Should this be unsuccessful, fights may result, with injuries to the head and neck. Reproduction and development The American mink is a promiscuous animal that does not form pair bonds. The start of mating season ranges from February in its southern range to April in the north. In its introduced range, the American mink breeds one month earlier than the European mink. Males commonly fight during the mating season, which may result in the formation of loose, temporary dominance hierarchies governing access to receptive females. The mating season lasts for three weeks, with ovulation being induced by the presence of males. The mating process is violent, with the male typically biting the female on the nape of the neck and pinning her with his forefeet. Mating lasts from 10 minutes to four hours. Females are receptive for seven- to 10-day intervals during the three-week breeding season, and can mate with multiple males. Along with the striped skunk, the American mink is among the only mammals to mate in spring that have a short delay before implantation. This delayed implantation allows pregnant minks to keep track of environmental conditions and select an ideal time and place for parturition. The gestation period lasts from 40 to 75 days, with actual embryonic development taking place for 30–32 days, indicating implantation delay can last from eight to 45 days. The young are born from April to June, in litters consisting of four kits on average. The litters are often multiply sired. Exceptionally large litters of 11 kits have been recorded in Tartaria and 16 in the United States. The kits are blind at birth, weighing six grams and possessing a short coat of fine, silver-white hairs. The kits are dependent on their mother's milk, which contains 3.8% lipids, 6.2% protein, 4.6% lactose and 10.66% mineral salts. Their eyes open after 25 days, with weaning occurring after five weeks. The kits begin hunting after 8 weeks of age, but stay close to their mother until autumn, when they become independent. Sexual maturity is attained during the kit's first spring, when they are about 10 months old. Diet The American mink is a carnivorous animal that feeds on rodents, fish, crustaceans, amphibians, and birds. It kills vertebrate prey by biting the back of the head or neck, leaving canine puncture marks apart. The American mink often kills birds, including larger species like seagulls and cormorants, by drowning. In its natural range, fish are its primary prey. Although inferior to the North American river otter in hunting fish, Audubon and Bachman once reported seeing a mink carrying a foot-long trout. Mink inhabiting the prairie sloughs primarily target frogs, tadpoles, and mice. It is a formidable predator of muskrats, which are chased underwater and killed in their own burrows. Among the rodents killed by the American mink in its native range are rats and mice of the genera Hesperomys, Microtus, Sigmodon, and Neotoma. Marsh rabbits are frequently taken in marshy or swampy tracts. In Tartaria, the American mink's most important food items are voles, fish, crustaceans, frogs, and aquatic insects. In winter, aquatic foods predominate, while land-based prey increases in importance during the spring. Within the Altai Mountains, the American mink feeds predominantly on mammals such as rodents, shrews, and moles, as well as birds, reptiles, amphibians, and fish. Among the 11 different bird species preyed upon by minks in Altai are dippers and pine grosbeaks. Among fish, small species predominate in the diet of minks in Altai, and include minnows, gudgeons, and wide-headed sculpins. In the Sverdlovsk and Irkutsk Oblasts, mouse-like rodents are their most important foods, followed by birds, fish and insects. In the Russian Far East, where crustaceans are scarce, the American mink feeds extensively on amphipods. In the British Isles, dietary composition varies seasonally and regionally. European rabbits are the most commonly taken prey in areas where they are common, especially in summer. A range of small rodents and insectivores are preyed upon, but to a lesser degree. European hares are occasionally attacked. Minks in Britain prey on several bird species, with ducks, moorhens, and coots being most frequently targeted on lakes and rivers, while gulls are taken in coastal habitats. Marine species preyed upon in Britain include European eels, rock-pool fish such as blenny, shore crabs and crayfish. American minks have been implicated in the decline of the water vole in the United Kingdom and linked to the decline of waterfowl across their range in Europe. They are now considered vermin in much of Europe and are hunted for the purpose of wildlife management. In South America's Cape Horn Biosphere Reserve, mammals, including both native and exotic rodents, are the American mink's main prey throughout the year, though birds are of equal importance during their summer nesting periods. The American mink may pose a threat to poultry. According to Clinton Hart Merriam and Ernest Thompson Seton, although the American mink is a potential poultry thief, it is overall less damaging than the stoat. Unlike the stoat, which often engages in surplus killing, the mink usually limits itself to killing and eating one fowl during each attack. Studies in Britain indicate poultry and game birds only constitute 1% of the animals' overall diets; small mammals, especially rabbits, tend to dominate, followed by fish and birds, especially moorhens and coots. Relationships with other predators The American mink replaces and sometimes kills the European mink wherever their ranges overlap. The decline of European mink populations seems to coincide with the spread of the American mink, but "The early declines in Central Europe and later in Finland took place before the spread of the American mink." The diets of the American mink and European otter overlap to a great extent. In areas where these two species are sympatric, competition with the otter for fish causes the American mink to hunt land-based prey more frequently. Large birds of prey, such as bald eagles (Haliaeetus leucocephalus) and great horned owls (Bubo virginianus) occasionally hunt American mink. In Finland, white-tailed eagles (Haliaeetus albicilla) have become the main natural control and may inhibit the mink from breeding via heavy predation. Eurasian eagle-owls (Bubo bubo) also can be predators of mink in the introduced regions. Intelligence An early behavioral study was performed in the 1960s to assess visual learning ability in minks, ferrets, skunks, and house cats. Animals were tested on their ability to recognize objects, learn their valences and make object selections from memory. Minks were found to outperform ferrets, skunks, and cats in this task, but this letter (short paper) fails to account for a possible conflation of a cognitive ability (decision making, associative learning) with a largely perceptual ability (invariant object recognition). Range Natural The species' natural range encompasses most of North America, from Alaska, through Canada and further into the United States except Arizona and the more arid areas of California, Nevada, Utah, New Mexico, and West Texas. Introduced Argentina The American mink was deliberately introduced for commercial fur production in several provinces of Patagonia in 1930. The animals escaped or were released from farms in Chubut Province and now occur in the Chubut and Río Negro Provinces and Tierra del Fuego. In Argentina the mink is one of the major menaces of the Hooded grebe, which risks becoming extinct. South Chile In Chile, American minks were introduced to Magallanes Region in the 1930s. Ever since minks were freed into nature during the crisis of the fur industry the mink has expanded its range across Chile. Actually it ranges from Araucanía Region in the north to Magallanes Region in the south. However, there are isolated territories in between where the mink is not found, probably due to biogeographic barriers. One of the latest areas where the mink has been found is Chiloé Archipelago,- minks were reported there for the first time 2013, making scientists suspect they may have arrived on a ship. Western Europe Feral American minks in Europe are thought to be of domesticated stock derived from the N. v. vison, N. v. melampeplus and N. v. ingens subspecies. The first specimens were imported to Europe in 1920 for fur-farming purposes. The American mink was introduced in Italy in the 1950s, and currently resides mostly in the northeastern part of the Italian Peninsula. The majority of these populations do not appear to be self-sufficient, though minks in the Monti Prenestini and Simbruini in Lazio have reproduced successfully. Escapees of fur farms established a self-sustaining and expanding population on the Iberian Peninsula by the second half of the 20th century. In 2013, the Spanish government announced a plan to eradicate the species, as a means to protect the falling populations of European mink and other endangered species affected such as the Pyrenean desman. The first mink farm in Norway was built in 1927, with escapees establishing wild populations within 30 years of its establishment. The first feral mink populations arose in 1930, establishing territories in southwestern Norway. These feral minks, augmented by further escapees, formed the basis of a strong population in Hordaland by the end of World War II. Feral mink colonised eastern Norway in 1930 and had become established in most southeastern counties in the early 1940s. By 1950, feral mink reached central Norway, with further populations occurring in the northern counties of Nordland and Troms. During the post-World War II period until 1965, mink had colonised most of the country. In modern times, the American mink occupies all of the Norwegian mainland, but is absent on some islands. The American mink was first imported to Great Britain in 1929, though a series of escapes and releases led to the establishment of a self-sufficient feral population in Devon by the late 1950s, and others by the early 1960s. In Ireland, the American mink was not farmed until the early 1950s, thus feral populations established themselves there much later. The species is now widespread in mainland Great Britain and Ireland, though some places remain uncolonised. It has established itself on a few islands, including Arran and Lewis and Harris. Until 2005, mink hunting with packs of hounds occurred in the UK. The total mink population in Great Britain is estimated at 110,000 (England: 46,750; Scotland: 52,250; Wales: 9,750). This population may be declining as European otter numbers increase. There are no estimates for the mink population in Ireland, but it is thought to be low, because of Ireland's strong otter population. Former USSR In 1933, American minks were released into the Voronezh Oblast in European Russia. Until 1963, more minks were introduced in various quantities in the Voronezh and Arkhangelsk Oblasts, Karelia, in Kalininsk, Gorkovsk, Volgograd and Chelyabinsk Oblasts, and into Tatarstan and Bashkir, as well as the Lithuanian and Byelorussian SSRs. Beyond the Urals, American minks were introduced in the Sverdlovsk, Tyumen, Omsk, Kemerovo, Novosibirsk, Chita and Irkutsk Oblasts, in the Altai and Krasnoyarsk Krai, in the Tuvan, Buryat and Yakut Autonomous Soviet Socialist Republics, into the Magadan, Kamchatka and Amur Oblasts, into the Khabarovsk and Primorsky Krai, into the Chukotka Autonomous Okrug and several other locations, including Sakhalin and Urup Island. In the Caucasus region, American minks were released into North Ossetia. In Central Asia they were released in the Tien Shan region. Originally, captive-bred minks were used, but wild specimens were later released to facilitate the species' acclimatisation within Soviet territories. Several years after the first release, introductions into the ranges already held by native European minks were discontinued, with most releases from then on taking place in Siberia and the Far East. Although considerable areas were occupied by the American mink by the early 1960s, the species' Soviet range was never continuous, as most released populations were isolated from one another. Iceland The species has been present in Iceland since the 1930s, and has become well established, despite it being heavily hunted since 1939. However, its population underwent a 42% decline during the years 2002–2006, which coincided with a decline in sand eel populations resulting in a drop in the seabird populations on which the minks feed. Asia American mink is considered as one of the most widespread invasive species in China, especially in Northeastern area. Minks are introduced in some prefectures Japan as well, especially being problematic in Hokkaido, and regulated by law. Diseases and parasites The American mink often carries light tick and flea infestations. Tick species known to infest minks include Ixodes hexagonus, Ixodes canisuga, Ixodes ricinus, and Ixodes acuminatus. Flea species known to infest minks include Palaeopsylla minor, Malaraeus penicilliger, Ctenopthalmus noblis, Megabothris walkeri, Typhloceras poppei, and Nosopsyllus fasciatus. Endoparasites include Skrjabingylus nasicola and Troglotrema acutum. Trematode Metorchis conjunctus can also infect American minks. Transmissible mink encephalopathy (TME) is a prion disease of mink, similar to BSE in cattle and scrapie in sheep. A 1985 outbreak of TME in Stetsonville, Wisconsin resulted in a 60% mortality rate for the minks. Further testing revealed this agent is transmissible between mink, cattle, and sheep. The Stetsonville outbreak may have been due to the mink being fed carcasses or otherwise consuming other infected animals. Toxoplasma gondii has been recorded in American minks in southern Chile, with 59% seroprevalence in the 73 mink analyzed in one study. Southern river otters of the area were also found to have high T. gondii seroprevalence in this study, showing a high exposure of semiaquatic mustelids to the parasite in this part of Chile. Decline of wild mink Because of numerous incidents of domestic mink escaping from fur farms and establishing themselves in the wild, concern has arisen among conservationists of the possible repercussions such escapes may have on natural wild mink populations. Domestic mink are larger than wild mink, which may cause problems with the ecosystem when they escape. Minks are solitary, territorial animals and are intolerant of other minks. In times of overpopulation, they control their own numbers by either killing each other through direct conflict or by causing weaker minks to be driven from territory until starvation sets in. When hundreds or thousands of released domestic minks flood an ecosystem, it causes a great disturbance for the wild minks, resulting in the deaths of the majority of the released mink and many of the wild ones from starvation or injuries incurred while fighting over territory. When a domestic mink survives long enough to reproduce, it may cause problems for the wild mink populations. The adding of weaker domestic mink genes into wild mink populations is believed by some to have contributed to the decline of mink populations in Canada. A 2006 study in Denmark concluded, due to frequent escapes from existing mink farms, "Closing mink farms may result in a crash of the free-ranging population, or alternatively it may result in the establishment of a better-adapted, truly feral population that may ultimately outnumber the population that was present before farm closures." The study reported more information would be necessary to determine the outcome. Another Danish study reported a significant majority of the "wild" mink were mink which had escaped from fur farms. About 47% had escaped within two months, 31% had escaped prior to two months, and 21% "may or may not have been born in nature." The survival rate for recently released minks is reportedly lower than for wild minks, but if feral minks survive at least two months, their survival rate is the same as for wild minks. The authors suggest this is due to the rapid behavioural adaptation of the animals. Relationships with humans Disease The American mink and European mink have both been found to transmit SARS-CoV-2, the virus that causes COVID-19. Fur use American minks are primarily used in manufacturing fur coats, jackets, and capes. Pelts that are not able to be converted into these items are made into trimming for cloth and fur coats. Mink scarves and stoles are also manufactured. Jackets and capes are mostly made from small to medium-sized specimens, usually females and young males, while trimming, scarves and stoles are made from adult males. The most valuable peltries come from eastern Canada which, although the smallest, are the silkiest and darkest. Trapping Although difficult to catch, the American mink, prior to being commercially farmed, was among the most frequently trapped furbearers as, unlike other furbearing mammals, it did not hibernate in winter, and could thus be caught on a nightly basis even in the far north. Minks were legally trapped from early November to early April, when their pelts were prime. Minks caught in traps cling to life with great tenacity, having been known to break their teeth in trying to extricate themselves from steel traps. Elliott Coues described a trapped mink thusly: One Native American method involved using a bait (usually a slit open chicken carcass filled with fish oil and oysters) tied to a rope and dragged around an area laden with traps. A mink would thus follow the trail into one of the traps. Another indigenous method involved placing traps scented with muskrat and female mink musk on top of disused muskrat dens by water bodies. Attracted by the scent of food and a female, the mink would get caught in the trap and drown. On the American prairies, only the steel trap was used, due to the lack of timber. Farming Breeding American minks for their fur began in the late 19th century, as increasing enthusiasm for mink pelts made the harvesting of wild minks insufficient to meet the new demands. American minks are easily kept in captivity, and breed readily. In 2005, the US ranked fourth in production behind Denmark, China and the Netherlands. Minks typically breed in March, and give birth to their litters in May. Farmers vaccinate the young kits for botulism, distemper, enteritis, and, if needed, pneumonia. They are harvested in late November and December. Methods for killing animals on fur farms, as on all farms, are detailed in the American Veterinary Medical Association's Report on Euthanasia which is used as a voluntary guideline for state departments of agriculture which have jurisdiction over all farms raising domesticated livestock, including minks. In the past, some mink farms successfully provided pools of water for the mink to swim; however, this practice is unheard-of in modern mink production. Minks are motivated to access swimming water, and the absence of water is a source of frustration on modern farms. The ideal diet for farm-bred minks consists of four to five ounces of horse meat and a quarter-pint of milk once daily. Colour mutations Selective breeding has produced a number of different colour shades in mink peltries, ranging from pure white, through beiges, browns, and greys, to a brown that is almost black. The two standard strains are brown and "black cross" which, when paired, produce numerous colour variations. When an albino mink is born, it is standard procedure in fur farms to breed it to other colour mutations to produce grey and light-brown pastel shades. The following graph is a simplification of the main colour strains: As pets Wild mink can be tamed if caught young, but can be difficult to handle and are usually not handled bare-handed. In the late 19th century, tame American minks were often reared for ratting, much as ferrets were used in Europe. Some modern ratters have revived this practice using farm-raised mink, sometimes alongside ratting dogs. They are sometimes more effective ratters than terriers, as they can enter rat holes and drive rats from their hiding places. When mink of wild stock are confined with tame ones, the wild invariably dominate the tamed mink. They have also been known to dominate cats in confrontations. Although intelligent, minks are not quick to learn tricks taught to them by their owners. Because of their fondness for bathing, captive American minks may enter kettles or other open water-containing vessels. Although domestic minks have been bred in captivity for almost a century, they have not been bred to be tame. Domestic minks have been bred for size, fur quality, and color. However, the U.S. Fur Commission claims "mink are truly domesticated animals", based on the number of years they have been kept on fur farms. Literature As an invasive species in the United Kingdom, minks have been the subject of at least two novels. Ewan Clarkson's 1968 Break for Freedom (published as Syla, the Mink in the US) tells the story of a female mink escaped from a fur farm in a realistic style. On the other hand, A.R. Lloyd's 1982 Kine is a heroic fantasy with the minks as villains and the weasels and other indigenous animals as heroes. Indigenous names Abenaki language Penobscot: or James Bay Cree: Dakelh Nadleh Whut'en: Nak'azdli: Dënesųłinë́: Halkomelem Hul'q'umi'num: Halqeméylem: Haíɫzaqv: Kwak̓wala: Lenape Lushootseed Northern Lushootseed: Southern Lushootseed: Nishnaabemwin: Nlaka'pamuctsin: Nuu-chah-nulth Ehattesaht: Tseshaht: Salish: Senćoŧen: Shashishalhem: Secwepemc: Sm'álgyax: Tse'khene: Witsuwitʼen: Wolastoqey-Passamaquoddy:
Biology and health sciences
Mustelidae
Animals
19916559
https://en.wikipedia.org/wiki/Atomic%20nucleus
Atomic nucleus
The atomic nucleus is the small, dense region consisting of protons and neutrons at the center of an atom, discovered in 1911 by Ernest Rutherford based on the 1909 Geiger–Marsden gold foil experiment. After the discovery of the neutron in 1932, models for a nucleus composed of protons and neutrons were quickly developed by Dmitri Ivanenko and Werner Heisenberg. An atom is composed of a positively charged nucleus, with a cloud of negatively charged electrons surrounding it, bound together by electrostatic force. Almost all of the mass of an atom is located in the nucleus, with a very small contribution from the electron cloud. Protons and neutrons are bound together to form a nucleus by the nuclear force. The diameter of the nucleus is in the range of () for hydrogen (the diameter of a single proton) to about for uranium. These dimensions are much smaller than the diameter of the atom itself (nucleus + electron cloud), by a factor of about 26,634 (uranium atomic radius is about ()) to about 60,250 (hydrogen atomic radius is about ). The branch of physics involved with the study and understanding of the atomic nucleus, including its composition and the forces that bind it together, is called nuclear physics. History The nucleus was discovered in 1911, as a result of Ernest Rutherford's efforts to test Thomson's "plum pudding model" of the atom. The electron had already been discovered by J. J. Thomson. Knowing that atoms are electrically neutral, J. J. Thomson postulated that there must be a positive charge as well. In his plum pudding model, Thomson suggested that an atom consisted of negative electrons randomly scattered within a sphere of positive charge. Ernest Rutherford later devised an experiment with his research partner Hans Geiger and with help of Ernest Marsden, that involved the deflection of alpha particles (helium nuclei) directed at a thin sheet of metal foil. He reasoned that if J. J. Thomson's model were correct, the positively charged alpha particles would easily pass through the foil with very little deviation in their paths, as the foil should act as electrically neutral if the negative and positive charges are so intimately mixed as to make it appear neutral. To his surprise, many of the particles were deflected at very large angles. Because the mass of an alpha particle is about 8000 times that of an electron, it became apparent that a very strong force must be present if it could deflect the massive and fast moving alpha particles. He realized that the plum pudding model could not be accurate and that the deflections of the alpha particles could only be explained if the positive and negative charges were separated from each other and that the mass of the atom was a concentrated point of positive charge. This justified the idea of a nuclear atom with a dense center of positive charge and mass. Etymology The term nucleus is from the Latin word , a diminutive of ('nut'), meaning 'the kernel' (i.e., the 'small nut') inside a watery type of fruit (like a peach). In 1844, Michael Faraday used the term to refer to the "central point of an atom". The modern atomic meaning was proposed by Ernest Rutherford in 1912. The adoption of the term "nucleus" to atomic theory, however, was not immediate. In 1916, for example, Gilbert N. Lewis stated, in his famous article The Atom and the Molecule, that "the atom is composed of the kernel and an outer atom or shell." Similarly, the term kern meaning kernel is used for nucleus in German and Dutch. Principles The nucleus of an atom consists of neutrons and protons, which in turn are the manifestation of more elementary particles, called quarks, that are held in association by the nuclear strong force in certain stable combinations of hadrons, called baryons. The nuclear strong force extends far enough from each baryon so as to bind the neutrons and protons together against the repulsive electrical force between the positively charged protons. The nuclear strong force has a very short range, and essentially drops to zero just beyond the edge of the nucleus. The collective action of the positively charged nucleus is to hold the electrically negative charged electrons in their orbits about the nucleus. The collection of negatively charged electrons orbiting the nucleus display an affinity for certain configurations and numbers of electrons that make their orbits stable. Which chemical element an atom represents is determined by the number of protons in the nucleus; the neutral atom will have an equal number of electrons orbiting that nucleus. Individual chemical elements can create more stable electron configurations by combining to share their electrons. It is that sharing of electrons to create stable electronic orbits about the nuclei that appears to us as the chemistry of our macro world. Protons define the entire charge of a nucleus, and hence its chemical identity. Neutrons are electrically neutral, but contribute to the mass of a nucleus to nearly the same extent as the protons. Neutrons can explain the phenomenon of isotopes (same atomic number with different atomic mass). The main role of neutrons is to reduce electrostatic repulsion inside the nucleus. Composition and shape Protons and neutrons are fermions, with different values of the strong isospin quantum number, so two protons and two neutrons can share the same space wave function since they are not identical quantum entities. They are sometimes viewed as two different quantum states of the same particle, the nucleon. Two fermions, such as two protons, or two neutrons, or a proton + neutron (the deuteron) can exhibit bosonic behavior when they become loosely bound in pairs, which have integer spin. In the rare case of a hypernucleus, a third baryon called a hyperon, containing one or more strange quarks and/or other unusual quark(s), can also share the wave function. However, this type of nucleus is extremely unstable and not found on Earth except in high-energy physics experiments. The neutron has a positively charged core of radius ≈ 0.3 fm surrounded by a compensating negative charge of radius between 0.3 fm and 2 fm. The proton has an approximately exponentially decaying positive charge distribution with a mean square radius of about 0.8 fm. The shape of the atomic nucleus can be spherical, rugby ball-shaped (prolate deformation), discus-shaped (oblate deformation), triaxial (a combination of oblate and prolate deformation) or pear-shaped. Forces Nuclei are bound together by the residual strong force (nuclear force). The residual strong force is a minor residuum of the strong interaction which binds quarks together to form protons and neutrons. This force is much weaker between neutrons and protons because it is mostly neutralized within them, in the same way that electromagnetic forces between neutral atoms (such as van der Waals forces that act between two inert gas atoms) are much weaker than the electromagnetic forces that hold the parts of the atoms together internally (for example, the forces that hold the electrons in an inert gas atom bound to its nucleus). The nuclear force is highly attractive at the distance of typical nucleon separation, and this overwhelms the repulsion between protons due to the electromagnetic force, thus allowing nuclei to exist. However, the residual strong force has a limited range because it decays quickly with distance (see Yukawa potential); thus only nuclei smaller than a certain size can be completely stable. The largest known completely stable nucleus (i.e. stable to alpha, beta, and gamma decay) is lead-208 which contains a total of 208 nucleons (126 neutrons and 82 protons). Nuclei larger than this maximum are unstable and tend to be increasingly short-lived with larger numbers of nucleons. However, bismuth-209 is also stable to beta decay and has the longest half-life to alpha decay of any known isotope, estimated at a billion times longer than the age of the universe. The residual strong force is effective over a very short range (usually only a few femtometres (fm); roughly one or two nucleon diameters) and causes an attraction between any pair of nucleons. For example, between a proton and a neutron to form a deuteron [NP], and also between protons and protons, and neutrons and neutrons. Halo nuclei and nuclear force range limits The effective absolute limit of the range of the nuclear force (also known as residual strong force) is represented by halo nuclei such as lithium-11 or boron-14, in which dineutrons, or other collections of neutrons, orbit at distances of about (roughly similar to the radius of the nucleus of uranium-238). These nuclei are not maximally dense. Halo nuclei form at the extreme edges of the chart of the nuclides—the neutron drip line and proton drip line—and are all unstable with short half-lives, measured in milliseconds; for example, lithium-11 has a half-life of . Halos in effect represent an excited state with nucleons in an outer quantum shell which has unfilled energy levels "below" it (both in terms of radius and energy). The halo may be made of either neutrons [NN, NNN] or protons [PP, PPP]. Nuclei which have a single neutron halo include 11Be and 19C. A two-neutron halo is exhibited by 6He, 11Li, 17B, 19B and 22C. Two-neutron halo nuclei break into three fragments, never two, and are called Borromean nuclei because of this behavior (referring to a system of three interlocked rings in which breaking any ring frees both of the others). 8He and 14Be both exhibit a four-neutron halo. Nuclei which have a proton halo include 8B and 26P. A two-proton halo is exhibited by 17Ne and 27S. Proton halos are expected to be more rare and unstable than the neutron examples, because of the repulsive electromagnetic forces of the halo proton(s). Nuclear models Although the standard model of physics is widely believed to completely describe the composition and behavior of the nucleus, generating predictions from theory is much more difficult than for most other areas of particle physics. This is due to two reasons: In principle, the physics within a nucleus can be derived entirely from quantum chromodynamics (QCD). In practice however, current computational and mathematical approaches for solving QCD in low-energy systems such as the nuclei are extremely limited. This is due to the phase transition that occurs between high-energy quark matter and low-energy hadronic matter, which renders perturbative techniques unusable, making it difficult to construct an accurate QCD-derived model of the forces between nucleons. Current approaches are limited to either phenomenological models such as the Argonne v18 potential or chiral effective field theory. Even if the nuclear force is well constrained, a significant amount of computational power is required to accurately compute the properties of nuclei ab initio. Developments in many-body theory have made this possible for many low mass and relatively stable nuclei, but further improvements in both computational power and mathematical approaches are required before heavy nuclei or highly unstable nuclei can be tackled. Historically, experiments have been compared to relatively crude models that are necessarily imperfect. None of these models can completely explain experimental data on nuclear structure. The nuclear radius (R) is considered to be one of the basic quantities that any model must predict. For stable nuclei (not halo nuclei or other unstable distorted nuclei) the nuclear radius is roughly proportional to the cube root of the mass number (A) of the nucleus, and particularly in nuclei containing many nucleons, as they arrange in more spherical configurations: The stable nucleus has approximately a constant density and therefore the nuclear radius R can be approximated by the following formula, where A = Atomic mass number (the number of protons Z, plus the number of neutrons N) and r0 = 1.25 fm = 1.25 × 10−15 m. In this equation, the "constant" r0 varies by 0.2 fm, depending on the nucleus in question, but this is less than 20% change from a constant. In other words, packing protons and neutrons in the nucleus gives approximately the same total size result as packing hard spheres of a constant size (like marbles) into a tight spherical or almost spherical bag (some stable nuclei are not quite spherical, but are known to be prolate). Models of nuclear structure include: Cluster model The cluster model describes the nucleus as a molecule-like collection of proton-neutron groups (e.g., alpha particles) with one or more valence neutrons occupying molecular orbitals. Liquid drop model Early models of the nucleus viewed the nucleus as a rotating liquid drop. In this model, the trade-off of long-range electromagnetic forces and relatively short-range nuclear forces, together cause behavior which resembled surface tension forces in liquid drops of different sizes. This formula is successful at explaining many important phenomena of nuclei, such as their changing amounts of binding energy as their size and composition changes (see semi-empirical mass formula), but it does not explain the special stability which occurs when nuclei have special "magic numbers" of protons or neutrons. The terms in the semi-empirical mass formula, which can be used to approximate the binding energy of many nuclei, are considered as the sum of five types of energies (see below). Then the picture of a nucleus as a drop of incompressible liquid roughly accounts for the observed variation of binding energy of the nucleus: Volume energy. When an assembly of nucleons of the same size is packed together into the smallest volume, each interior nucleon has a certain number of other nucleons in contact with it. So, this nuclear energy is proportional to the volume. Surface energy. A nucleon at the surface of a nucleus interacts with fewer other nucleons than one in the interior of the nucleus and hence its binding energy is less. This surface energy term takes that into account and is therefore negative and is proportional to the surface area. Coulomb energy. The electric repulsion between each pair of protons in a nucleus contributes toward decreasing its binding energy. Asymmetry energy (also called Pauli Energy). An energy associated with the Pauli exclusion principle. Were it not for the Coulomb energy, the most stable form of nuclear matter would have the same number of neutrons as protons, since unequal numbers of neutrons and protons imply filling higher energy levels for one type of particle, while leaving lower energy levels vacant for the other type. Pairing energy. An energy which is a correction term that arises from the tendency of proton pairs and neutron pairs to occur. An even number of particles is more stable than an odd number. Shell models and other quantum models A number of models for the nucleus have also been proposed in which nucleons occupy orbitals, much like the atomic orbitals in atomic physics theory. These wave models imagine nucleons to be either sizeless point particles in potential wells, or else probability waves as in the "optical model", frictionlessly orbiting at high speed in potential wells. In the above models, the nucleons may occupy orbitals in pairs, due to being fermions, which allows explanation of even/odd Z and N effects well known from experiments. The exact nature and capacity of nuclear shells differs from those of electrons in atomic orbitals, primarily because the potential well in which the nucleons move (especially in larger nuclei) is quite different from the central electromagnetic potential well which binds electrons in atoms. Some resemblance to atomic orbital models may be seen in a small atomic nucleus like that of helium-4, in which the two protons and two neutrons separately occupy 1s orbitals analogous to the 1s orbital for the two electrons in the helium atom, and achieve unusual stability for the same reason. Nuclei with 5 nucleons are all extremely unstable and short-lived, yet, helium-3, with 3 nucleons, is very stable even with lack of a closed 1s orbital shell. Another nucleus with 3 nucleons, the triton hydrogen-3 is unstable and will decay into helium-3 when isolated. Weak nuclear stability with 2 nucleons {NP} in the 1s orbital is found in the deuteron hydrogen-2, with only one nucleon in each of the proton and neutron potential wells. While each nucleon is a fermion, the {NP} deuteron is a boson and thus does not follow Pauli Exclusion for close packing within shells. Lithium-6 with 6 nucleons is highly stable without a closed second 1p shell orbital. For light nuclei with total nucleon numbers 1 to 6 only those with 5 do not show some evidence of stability. Observations of beta-stability of light nuclei outside closed shells indicate that nuclear stability is much more complex than simple closure of shell orbitals with magic numbers of protons and neutrons. For larger nuclei, the shells occupied by nucleons begin to differ significantly from electron shells, but nevertheless, present nuclear theory does predict the magic numbers of filled nuclear shells for both protons and neutrons. The closure of the stable shells predicts unusually stable configurations, analogous to the noble group of nearly-inert gases in chemistry. An example is the stability of the closed shell of 50 protons, which allows tin to have 10 stable isotopes, more than any other element. Similarly, the distance from shell-closure explains the unusual instability of isotopes which have far from stable numbers of these particles, such as the radioactive elements 43 (technetium) and 61 (promethium), each of which is preceded and followed by 17 or more stable elements. There are however problems with the shell model when an attempt is made to account for nuclear properties well away from closed shells. This has led to complex post hoc distortions of the shape of the potential well to fit experimental data, but the question remains whether these mathematical manipulations actually correspond to the spatial deformations in real nuclei. Problems with the shell model have led some to propose realistic two-body and three-body nuclear force effects involving nucleon clusters and then build the nucleus on this basis. Three such cluster models are the 1936 Resonating Group Structure model of John Wheeler, Close-Packed Spheron Model of Linus Pauling and the 2D Ising Model of MacGregor.
Physical sciences
Nuclear physics
null
19916594
https://en.wikipedia.org/wiki/Acetic%20acid
Acetic acid
Acetic acid , systematically named ethanoic acid , is an acidic, colourless liquid and organic compound with the chemical formula (also written as , , or ). Vinegar is at least 4% acetic acid by volume, making acetic acid the main component of vinegar apart from water. It has been used, as a component of vinegar, throughout history from at least the third century BC. Acetic acid is the second simplest carboxylic acid (after formic acid). It is an important chemical reagent and industrial chemical across various fields, used primarily in the production of cellulose acetate for photographic film, polyvinyl acetate for wood glue, and synthetic fibres and fabrics. In households, diluted acetic acid is often used in descaling agents. In the food industry, acetic acid is controlled by the food additive code E260 as an acidity regulator and as a condiment. In biochemistry, the acetyl group, derived from acetic acid, is fundamental to all forms of life. When bound to coenzyme A, it is central to the metabolism of carbohydrates and fats. The global demand for acetic acid as of 2023 is about 17.88 million metric tonnes per year (t/a). Most of the world's acetic acid is produced via the carbonylation of methanol. Its production and subsequent industrial use poses health hazards to workers, including incidental skin damage and chronic respiratory injuries from inhalation. Nomenclature The trivial name "acetic acid" is the most commonly used and preferred IUPAC name. The systematic name "ethanoic acid", a valid IUPAC name, is constructed according to the substitutive nomenclature. The name "acetic acid" derives from the Latin word for vinegar, "", which is related to the word "acid" itself. "Glacial acetic acid" is a name for water-free (anhydrous) acetic acid. Similar to the German name "Eisessig" ("ice vinegar"), the name comes from the solid ice-like crystals that form with agitation, slightly below room temperature at . Acetic acid can never be truly water-free in an atmosphere that contains water, so the presence of 0.1% water in glacial acetic acid lowers its melting point by 0.2 °C. A common symbol for acetic acid is AcOH (or HOAc), where Ac is the pseudoelement symbol representing the acetyl group ; the conjugate base, acetate (), is thus represented as . Acetate is the ion resulting from loss of from acetic acid. The name "acetate" can also refer to a salt containing this anion, or an ester of acetic acid. (The symbol Ac for the acetyl functional group is not to be confused with the symbol Ac for the element actinium; context prevents confusion among organic chemists). To better reflect its structure, acetic acid is often written as , , , and . In the context of acid–base reactions, the abbreviation HAc is sometimes used, where Ac in this case is a symbol for acetate (rather than acetyl). The carboxymethyl functional group derived from removing one hydrogen from the methyl group of acetic acid has the chemical formula . History Vinegar was known early in civilization as the natural result of exposure of beer and wine to air because acetic acid-producing bacteria are present globally. The use of acetic acid in alchemy extends into the third century BC, when the Greek philosopher Theophrastus described how vinegar acted on metals to produce pigments useful in art, including white lead (lead carbonate) and verdigris, a green mixture of copper salts including copper(II) acetate. Ancient Romans boiled soured wine to produce a highly sweet syrup called sapa. Sapa that was produced in lead pots was rich in lead acetate, a sweet substance also called sugar of lead or sugar of Saturn, which contributed to lead poisoning among the Roman aristocracy. In the 16th-century German alchemist Andreas Libavius described the production of acetone from the dry distillation of lead acetate, ketonic decarboxylation. The presence of water in vinegar has such a profound effect on acetic acid's properties that for centuries chemists believed that glacial acetic acid and the acid found in vinegar were two different substances. French chemist Pierre Adet proved them identical. In 1845 German chemist Hermann Kolbe synthesised acetic acid from inorganic compounds for the first time. This reaction sequence consisted of chlorination of carbon disulfide to carbon tetrachloride, followed by pyrolysis to tetrachloroethylene and aqueous chlorination to trichloroacetic acid, and concluded with electrolytic reduction to acetic acid. By 1910, most glacial acetic acid was obtained from the pyroligneous liquor, a product of the distillation of wood. The acetic acid was isolated by treatment with milk of lime, and the resulting calcium acetate was then acidified with sulfuric acid to recover acetic acid. At that time, Germany was producing 10,000 tons of glacial acetic acid, around 30% of which was used for the manufacture of indigo dye. Because both methanol and carbon monoxide are commodity raw materials, methanol carbonylation long appeared to be attractive precursors to acetic acid. Henri Dreyfus at British Celanese developed a methanol carbonylation pilot plant as early as 1925. However, a lack of practical materials that could contain the corrosive reaction mixture at the high pressures needed (200 atm or more) discouraged commercialization of these routes. The first commercial methanol carbonylation process, which used a cobalt catalyst, was developed by German chemical company BASF in 1963. In 1968, a rhodium-based catalyst (cis−) was discovered that could operate efficiently at lower pressure with almost no by-products. US chemical company Monsanto Company built the first plant using this catalyst in 1970, and rhodium-catalyzed methanol carbonylation became the dominant method of acetic acid production (see Monsanto process). In the late 1990s, BP Chemicals commercialised the Cativa catalyst (), which is promoted by iridium for greater efficiency. Known as the Cativa process, the iridium-catalyzed production of glacial acetic acid is greener, and has largely supplanted the Monsanto process, often in the same production plants. Interstellar medium Interstellar acetic acid was discovered in 1996 by a team led by David Mehringer using the former Berkeley-Illinois-Maryland Association array at the Hat Creek Radio Observatory and the former Millimeter Array located at the Owens Valley Radio Observatory. It was first detected in the Sagittarius B2 North molecular cloud (also known as the Sgr B2 Large Molecule Heimat source). Acetic acid has the distinction of being the first molecule discovered in the interstellar medium using solely radio interferometers; in all previous ISM molecular discoveries made in the millimetre and centimetre wavelength regimes, single dish radio telescopes were at least partly responsible for the detections. Properties Acidity The hydrogen centre in the carboxyl group (−COOH) in carboxylic acids such as acetic acid can separate from the molecule by ionization: Because of this release of the proton (), acetic acid has acidic character. Acetic acid is a weak monoprotic acid. In aqueous solution, it has a pKa value of 4.76. Its conjugate base is acetate (). A 1.0 M solution (about the concentration of domestic vinegar) has a pH of 2.4, indicating that merely 0.4% of the acetic acid molecules are dissociated. Structure In solid acetic acid, the molecules form chains of individual molecules interconnected by hydrogen bonds. In the vapour phase at , dimers can be detected. Dimers also occur in the liquid phase in dilute solutions with non-hydrogen-bonding solvents, and to a certain extent in pure acetic acid, but are disrupted by hydrogen-bonding solvents. The dissociation enthalpy of the dimer is estimated at 65.0–66.0 kJ/mol, and the dissociation entropy at 154–157 J mol−1 K−1. Other carboxylic acids engage in similar intermolecular hydrogen bonding interactions. Solvent properties Liquid acetic acid is a hydrophilic (polar) protic solvent, similar to ethanol and water. With a relative static permittivity (dielectric constant) of 6.2, it dissolves not only polar compounds such as inorganic salts and sugars, but also non-polar compounds such as oils as well as polar solutes. It is miscible with polar and non-polar solvents such as water, chloroform, and hexane. With higher alkanes (starting with octane), acetic acid is not miscible at all compositions, and solubility of acetic acid in alkanes declines with longer n-alkanes. The solvent and miscibility properties of acetic acid make it a useful industrial chemical, for example, as a solvent in the production of dimethyl terephthalate. Biochemistry At physiological pHs, acetic acid is usually fully ionised to acetate in aqueous solution. The acetyl group, formally derived from acetic acid, is fundamental to all forms of life. Typically, it is bound to coenzyme A by acetyl-CoA synthetase enzymes, where it is central to the metabolism of carbohydrates and fats. Unlike longer-chain carboxylic acids (the fatty acids), acetic acid does not occur in natural triglycerides. Most of the acetate generated in cells for use in acetyl-CoA is synthesized directly from ethanol or pyruvate. However, the artificial triglyceride triacetin (glycerine triacetate) is a common food additive and is found in cosmetics and topical medicines; this additive is metabolized to glycerol and acetic acid in the body. Acetic acid is produced and excreted by acetic acid bacteria, notably the genus Acetobacter and Clostridium acetobutylicum. These bacteria are found universally in foodstuffs, water, and soil, and acetic acid is produced naturally as fruits and other foods spoil. Acetic acid is also a component of the vaginal lubrication of humans and other primates, where it appears to serve as a mild antibacterial agent. Production Acetic acid is produced industrially both synthetically and by bacterial fermentation. About 75% of acetic acid made for use in the chemical industry is made by the carbonylation of methanol, explained below. The biological route accounts for only about 10% of world production, but it remains important for the production of vinegar because many food purity laws require vinegar used in foods to be of biological origin. Other processes are methyl formate isomerization, conversion of syngas to acetic acid, and gas phase oxidation of ethylene and ethanol. Acetic acid can be purified via fractional freezing using an ice bath. The water and other impurities will remain liquid while the acetic acid will precipitate out. As of 2003–2005, total worldwide production of virgin acetic acid was estimated at 5 Mt/a (million tonnes per year), approximately half of which was produced in the United States. European production was approximately 1 Mt/a and declining, while Japanese production was 0.7 Mt/a. Another 1.5 Mt were recycled each year, bringing the total world market to 6.5 Mt/a. Since then, the global production has increased from 10.7 Mt/a in 2010 to 17.88 Mt/a in 2023. The two biggest producers of virgin acetic acid are Celanese and BP Chemicals. Other major producers include Millennium Chemicals, Sterling Chemicals, Samsung, Eastman, and . Methanol carbonylation Most acetic acid is produced by methanol carbonylation. In this process, methanol and carbon monoxide react to produce acetic acid according to the equation: The process involves iodomethane as an intermediate, and occurs in three steps. A metal carbonyl catalyst is needed for the carbonylation (step 2). Two related processes exist for the carbonylation of methanol: the rhodium-catalyzed Monsanto process, and the iridium-catalyzed Cativa process. The latter process is greener and more efficient and has largely supplanted the former process. Catalytic amounts of water are used in both processes, but the Cativa process requires less, so the water-gas shift reaction is suppressed, and fewer by-products are formed. By altering the process conditions, acetic anhydride may also be produced in plants using rhodium catalysis. Acetaldehyde oxidation Prior to the commercialization of the Monsanto process, most acetic acid was produced by oxidation of acetaldehyde. This remains the second-most-important manufacturing method, although it is usually not competitive with the carbonylation of methanol. The acetaldehyde can be produced by hydration of acetylene. This was the dominant technology in the early 1900s. Light naphtha components are readily oxidized by oxygen or even air to give peroxides, which decompose to produce acetic acid according to the chemical equation, illustrated with butane: Such oxidations require metal catalyst, such as the naphthenate salts of manganese, cobalt, and chromium. The typical reaction is conducted at temperatures and pressures designed to be as hot as possible while still keeping the butane a liquid. Typical reaction conditions are and 55 atm. Side-products may also form, including butanone, ethyl acetate, formic acid, and propionic acid. These side-products are also commercially valuable, and the reaction conditions may be altered to produce more of them where needed. However, the separation of acetic acid from these by-products adds to the cost of the process. Similar conditions and catalysts are used for butane oxidation, the oxygen in air to produce acetic acid can oxidize acetaldehyde. Using modern catalysts, this reaction can have an acetic acid yield greater than 95%. The major side-products are ethyl acetate, formic acid, and formaldehyde, all of which have lower boiling points than acetic acid and are readily separated by distillation. Ethylene oxidation Acetaldehyde may be prepared from ethylene via the Wacker process, and then oxidised as above. In more recent times, chemical company Showa Denko, which opened an ethylene oxidation plant in Ōita, Japan, in 1997, commercialised a cheaper single-stage conversion of ethylene to acetic acid. The process is catalyzed by a palladium metal catalyst supported on a heteropoly acid such as silicotungstic acid. A similar process uses the same metal catalyst on silicotungstic acid and silica: It is thought to be competitive with methanol carbonylation for smaller plants (100–250 kt/a), depending on the local price of ethylene. Oxidative fermentation For most of human history, acetic acid bacteria of the genus Acetobacter have made acetic acid, in the form of vinegar. Given sufficient oxygen, these bacteria can produce vinegar from a variety of alcoholic foodstuffs. Commonly used feeds include apple cider, wine, and fermented grain, malt, rice, or potato mashes. The overall chemical reaction facilitated by these bacteria is: A dilute alcohol solution inoculated with Acetobacter and kept in a warm, airy place will become vinegar over the course of a few months. Industrial vinegar-making methods accelerate this process by improving the supply of oxygen to the bacteria. The first batches of vinegar produced by fermentation probably followed errors in the winemaking process. If must is fermented at too high a temperature, acetobacter will overwhelm the yeast naturally occurring on the grapes. As the demand for vinegar for culinary, medical, and sanitary purposes increased, vintners quickly learned to use other organic materials to produce vinegar in the hot summer months before the grapes were ripe and ready for processing into wine. This method was slow, however, and not always successful, as the vintners did not understand the process. One of the first modern commercial processes was the "fast method" or "German method", first practised in Germany in 1823. In this process, fermentation takes place in a tower packed with wood shavings or charcoal. The alcohol-containing feed is trickled into the top of the tower, and fresh air supplied from the bottom by either natural or forced convection. The improved air supply in this process cut the time to prepare vinegar from months to weeks. Nowadays, most vinegar is made in submerged tank culture, first described in 1949 by Otto Hromatka and Heinrich Ebner. In this method, alcohol is fermented to vinegar in a continuously stirred tank, and oxygen is supplied by bubbling air through the solution. Using modern applications of this method, vinegar of 15% acetic acid can be prepared in only 24 hours in batch process, even 20% in 60-hour fed-batch process. Anaerobic fermentation Species of anaerobic bacteria, including members of the genus Clostridium or Acetobacterium, can convert sugars to acetic acid directly without creating ethanol as an intermediate. The overall chemical reaction conducted by these bacteria may be represented as: These acetogenic bacteria produce acetic acid from one-carbon compounds, including methanol, carbon monoxide, or a mixture of carbon dioxide and hydrogen: This ability of Clostridium to metabolize sugars directly, or to produce acetic acid from less costly inputs, suggests that these bacteria could produce acetic acid more efficiently than ethanol-oxidizers like Acetobacter. However, Clostridium bacteria are less acid-tolerant than Acetobacter. Even the most acid-tolerant Clostridium strains can produce vinegar in concentrations of only a few per cent, compared to Acetobacter strains that can produce vinegar in concentrations up to 20%. At present, it remains more cost-effective to produce vinegar using Acetobacter, rather than using Clostridium and concentrating it. As a result, although acetogenic bacteria have been known since 1940, their industrial use is confined to a few niche applications. Uses Acetic acid is a chemical reagent for the production of chemical compounds. The largest single use of acetic acid is in the production of vinyl acetate monomer, closely followed by acetic anhydride and ester production. The volume of acetic acid used in vinegar is comparatively small. Vinyl acetate monomer The primary use of acetic acid is the production of vinyl acetate monomer (VAM). In 2008, this application was estimated to consume a third of the world's production of acetic acid. The reaction consists of ethylene and acetic acid with oxygen over a palladium catalyst, conducted in the gas phase. Vinyl acetate can be polymerised to polyvinyl acetate or other polymers, which are components in paints and adhesives. Ester production The major esters of acetic acid are commonly used as solvents for inks, paints and coatings. The esters include ethyl acetate, n-butyl acetate, isobutyl acetate, and propyl acetate. They are typically produced by catalyzed reaction from acetic acid and the corresponding alcohol: , R = general alkyl group For example, acetic acid and ethanol gives ethyl acetate and water. Most acetate esters, however, are produced from acetaldehyde using the Tishchenko reaction. In addition, ether acetates are used as solvents for nitrocellulose, acrylic lacquers, varnish removers, and wood stains. First, glycol monoethers are produced from ethylene oxide or propylene oxide with alcohol, which are then esterified with acetic acid. The three major products are ethylene glycol monoethyl ether acetate (EEA), ethylene glycol monobutyl ether acetate (EBA), and propylene glycol monomethyl ether acetate (PMA, more commonly known as PGMEA in semiconductor manufacturing processes, where it is used as a resist solvent). This application consumes about 15% to 20% of worldwide acetic acid. Ether acetates, for example EEA, have been shown to be harmful to human reproduction. Acetic anhydride The product of the condensation of two molecules of acetic acid is acetic anhydride. The worldwide production of acetic anhydride is a major application, and uses approximately 25% to 30% of the global production of acetic acid. The main process involves dehydration of acetic acid to give ketene at 700–750 °C. Ketene is thereafter reacted with acetic acid to obtain the anhydride: Acetic anhydride is an acetylation agent. As such, its major application is for cellulose acetate, a synthetic textile also used for photographic film. Acetic anhydride is also a reagent for the production of heroin and other compounds. Use as solvent As a polar protic solvent, acetic acid is frequently used for recrystallization to purify organic compounds. Acetic acid is used as a solvent in the production of terephthalic acid (TPA), the raw material for polyethylene terephthalate (PET). In 2006, about 20% of acetic acid was used for TPA production. Acetic acid is often used as a solvent for reactions involving carbocations, such as Friedel-Crafts alkylation. For example, one stage in the commercial manufacture of synthetic camphor involves a Wagner-Meerwein rearrangement of camphene to isobornyl acetate; here acetic acid acts both as a solvent and as a nucleophile to trap the rearranged carbocation. Glacial acetic acid is used in analytical chemistry for the estimation of weakly alkaline substances such as organic amides. Glacial acetic acid is a much weaker base than water, so the amide behaves as a strong base in this medium. It then can be titrated using a solution in glacial acetic acid of a very strong acid, such as perchloric acid. Medical use Acetic acid injection into a tumor has been used to treat cancer since the 1800s. Acetic acid is used as part of cervical cancer screening in many areas in the developing world. The acid is applied to the cervix and if an area of white appears after about a minute the test is positive. Acetic acid is an effective antiseptic when used as a 1% solution, with broad spectrum of activity against streptococci, staphylococci, pseudomonas, enterococci and others. It may be used to treat skin infections caused by pseudomonas strains resistant to typical antibiotics. While diluted acetic acid is used in iontophoresis, no high quality evidence supports this treatment for rotator cuff disease. As a treatment for otitis externa, it is on the World Health Organization's List of Essential Medicines. Foods Acetic acid has per 100 g. Vinegar is typically no less than 4% acetic acid by mass. Legal limits on acetic acid content vary by jurisdiction. Vinegar is used directly as a condiment, and in the pickling of vegetables and other foods. Table vinegar tends to be more diluted (4% to 8% acetic acid), while commercial food pickling employs solutions that are more concentrated. The proportion of acetic acid used worldwide as vinegar is not as large as industrial uses, but it is by far the oldest and best-known application. Reactions Organic chemistry Acetic acid undergoes the typical chemical reactions of a carboxylic acid. Upon treatment with a standard base, it converts to metal acetate and water. With strong bases (e.g., organolithium reagents), it can be doubly deprotonated to give . Reduction of acetic acid gives ethanol. The OH group is the main site of reaction, as illustrated by the conversion of acetic acid to acetyl chloride. Other substitution derivatives include acetic anhydride; this anhydride is produced by loss of water from two molecules of acetic acid. Esters of acetic acid can likewise be formed via Fischer esterification, and amides can be formed. When heated above , acetic acid decomposes to produce carbon dioxide and methane, or to produce ketene and water: Reactions with inorganic compounds Acetic acid is mildly corrosive to metals including iron, magnesium, and zinc, forming hydrogen gas and salts called acetates: Because aluminium forms a passivating acid-resistant film of aluminium oxide, aluminium tanks are used to transport acetic acid. Containers lined with glass, stainless steel or polyethylene are also used for this purpose. Metal acetates can also be prepared from acetic acid and an appropriate base, as in the popular "baking soda + vinegar" reaction giving off sodium acetate: A colour reaction for salts of acetic acid is iron(III) chloride solution, which results in a deeply red colour that disappears after acidification. A more sensitive test uses lanthanum nitrate with iodine and ammonia to give a blue solution. Acetates when heated with arsenic trioxide form cacodyl oxide, which can be detected by its malodorous vapours. Other derivatives Organic or inorganic salts are produced from acetic acid. Some commercially significant derivatives: Sodium acetate, used in the textile industry and as a food preservative (E262). Copper(II) acetate, used as a pigment and a fungicide. Aluminium acetate and iron(II) acetate—used as mordants for dyes. Palladium(II) acetate, used as a catalyst for organic coupling reactions such as the Heck reaction. Halogenated acetic acids are produced from acetic acid. Some commercially significant derivatives: Chloroacetic acid (monochloroacetic acid, MCA), dichloroacetic acid (considered a by-product), and trichloroacetic acid. MCA is used in the manufacture of indigo dye. Bromoacetic acid, which is esterified to produce the reagent ethyl bromoacetate. Trifluoroacetic acid, which is a common reagent in organic synthesis. Amounts of acetic acid used in these other applications together account for another 5–10% of acetic acid use worldwide. Health and safety Vapour Prolonged inhalation exposure (eight hours) to acetic acid vapours at 10 ppm can produce some irritation of eyes, nose, and throat; at 100 ppm marked lung irritation and possible damage to lungs, eyes, and skin may result. Vapour concentrations of 1,000 ppm cause marked irritation of eyes, nose and upper respiratory tract and cannot be tolerated. These predictions were based on animal experiments and industrial exposure. In 12 workers exposed for two or more years to an airborne average concentration of 51 ppm acetic acid (estimated), symptoms of conjunctive irritation, upper respiratory tract irritation, and hyperkeratotic dermatitis were produced. Exposure to 50 ppm or more is intolerable to most persons and results in intensive lacrimation and irritation of the eyes, nose, and throat, with pharyngeal oedema and chronic bronchitis. Unacclimatised humans experience extreme eye and nasal irritation at concentrations in excess of 25 ppm, and conjunctivitis from concentrations below 10 ppm has been reported. In a study of five workers exposed for seven to 12 years to concentrations of 80 to 200 ppm at peaks, the principal findings were blackening and hyperkeratosis of the skin of the hands, conjunctivitis (but no corneal damage), bronchitis and pharyngitis, and erosion of the exposed teeth (incisors and canines). Solution Concentrated acetic acid (≥ 25%) is corrosive to skin. These burns or blisters may not appear until hours after exposure. The hazardous properties of acetic acid are dependent on the concentration of the (typically aqueous) solution, with the most significant increases in hazard levels at thresholds of 25% and 90% acetic acid concentration by weight. The following table summarizes the hazards of acetic acid solutions by concentration: Concentrated acetic acid can be ignited only with difficulty at standard temperature and pressure, but becomes a flammable risk in temperatures greater than , and can form explosive mixtures with air at higher temperatures with explosive limits of 5.4–16% concentration.
Physical sciences
Carbon–oxygen bond
null
19916613
https://en.wikipedia.org/wiki/Radical%20%28chemistry%29
Radical (chemistry)
In chemistry, a radical, also known as a free radical, is an atom, molecule, or ion that has at least one unpaired valence electron. With some exceptions, these unpaired electrons make radicals highly chemically reactive. Many radicals spontaneously dimerize. Most organic radicals have short lifetimes. A notable example of a radical is the hydroxyl radical (HO·), a molecule that has one unpaired electron on the oxygen atom. Two other examples are triplet oxygen and triplet carbene (꞉) which have two unpaired electrons. Radicals may be generated in a number of ways, but typical methods involve redox reactions, ionizing radiation, heat, electrical discharges, and electrolysis are known to produce radicals. Radicals are intermediates in many chemical reactions, more so than is apparent from the balanced equations. Radicals are important in combustion, atmospheric chemistry, polymerization, plasma chemistry, biochemistry, and many other chemical processes. A majority of natural products are generated by radical-generating enzymes. In living organisms, the radicals superoxide and nitric oxide and their reaction products regulate many processes, such as control of vascular tone and thus blood pressure. They also play a key role in the intermediary metabolism of various biological compounds. Such radicals can even be messengers in a process dubbed redox signaling. A radical may be trapped within a solvent cage or be otherwise bound. Formation Radicals are either (1) formed from spin-paired molecules or (2) from other radicals. Radicals are formed from spin-paired molecules through homolysis of weak bonds or electron transfer, also known as reduction. Radicals are formed from other radicals through substitution, addition, and elimination reactions. Radical formation from spin-paired molecules Homolysis Homolysis makes two new radicals from a spin-paired molecule by breaking a covalent bond, leaving each of the fragments with one of the electrons in the bond. Because breaking a chemical bond requires energy, homolysis occurs under the addition of heat or light. The bond dissociation energy associated with homolysis depends on the stability of a given compound, and some weak bonds are able to homolyze at relatively lower temperatures. Some homolysis reactions are particularly important because they serve as an initiator for other radical reactions. One such example is the homolysis of halogens, which occurs under light and serves as the driving force for radical halogenation reactions. Another notable reaction is the homolysis of dibenzoyl peroxide, which results in the formation of two benzoyloxy radicals and acts as an initiator for many radical reactions. Reduction Classically, radicals form by one-electron reductions. Typically one-electron reduced organic compounds are unstable. Stability is conferred to the radical anion when the charge can be delocalized. Examples include alkali metal naphthenides, anthracenides, and ketyls. Radical formation from other radicals Abstraction Hydrogen abstraction generates radicals. To achieve this reaction, the C-H bond of the H-atom donor must be weak, which is rarely the case in organic compounds. Allylic and especiall doubly allylic C-H bonds are prone to abstraction by O2. This reaction is the basis of drying oils, such as linoleic acid derivatives. Addition In free-radical additions, a radical adds to a spin-paired substrate. When applied to organic compounds, the reaction usually entails addition to an alkene. This addition generates a new radical, which can add to yet another alkene, etc. This behavior underpins radical polymerization, technology that produces many plastics. Elimination Radical elimination can be viewed as the reverse of radical addition. In radical elimination, an unstable radical compound breaks down into a spin-paired molecule and a new radical compound. Shown below is an example of a radical elimination reaction, where a benzoyloxy radical breaks down into a phenyl radical and a carbon dioxide molecule. Stability Stability of organic radicals The generation and reactivity of organic radicals are dependent on both their thermodynamic stability and kinetic stability, also known as the persistency. This distinction is necessary because these two types of stability do not always correlate with each other. For example, benzylic radicals, which are known for their weak benzylic C−H bond strength, are thermodynamically stabilized due to resonance delocalization. However, these radicals are kinetically transient because they can undergo rapid, diffusion-limited dimerization, resulting in a lifetime that is less than a few nanoseconds. To avoid confusion, particularly for carbon-centered radicals, Griller and Ingold introduced the following definitions: "Stabilized should be used to describe a carbon-centered radical, R·, when the R−H bond strength is weaker than the appropriate C−H bond of alkane." "Persistent should be used to describe a radical that has a lifetime that is significantly greater than methyl [radical] under the same condition." While relationships between thermodynamic stability and kinetic persistency is highly case-dependent, organic radicals can be generally stabilized by any or all of these factors: the presence of electronegativity, delocalization, and steric hindrance. The compound 2,2,6,6-tetramethylpiperidinyloxyl illustrates the combination of all three factors. It is a commercially available solid that, aside from being magnetic, behaves like a normal organic compound. Electronegativity Organic radicals are inherently electron deficient thus the greater the electronegativity of the atom on which the unpaired electron resides the less stable the radical. Between carbon, nitrogen, and oxygen, for example, carbon is the most stable and oxygen the least stable. Electronegativity also factors into the stability of carbon atoms of different hybridizations. Greater s-character correlates to higher electronegativity of the carbon atom (due to the close proximity of s orbitals to the nucleus), and the greater the electronegativity the less stable a radical. sp-hybridized carbons (50% s-character) form the least stable radicals compared to sp3-hybridized carbons (25% s-character) which form the most stable radicals. Delocalization The delocalization of electrons across the structure of a radical, also known as its ability to form one or more resonance structures, allows for the electron deficiency to be spread over several atoms, minimizing instability. Delocalization usually occurs in the presence of electron-donating groups, such as hydroxyl groups (−OH), ethers (−OR), adjacent alkenes, and amines (−NH2 or −NR), or electron-withdrawing groups, such as C=O or C≡N. Delocalization effects can also be understood using molecular orbital theory as a lens, more specifically, by examining the intramolecular interaction of the unpaired electron with a donating group's pair of electrons or the empty π* orbital of an electron-withdrawing group in the form of a molecular orbital diagram. The HOMO of a radical is singly-occupied hence the orbital is aptly referred to as the SOMO, or the Singly-Occupied Molecular Orbital. For an electron-donating group, the SOMO interacts with the lower energy lone pair to form a new lower-energy filled bonding-orbital and a singly-filled new SOMO, higher in energy than the original. While the energy of the unpaired electron has increased, the decrease in energy of the lone pair forming the new bonding orbital outweighs the increase in energy of the new SOMO, resulting in a net decrease of the energy of the molecule. Therefore, electron-donating groups help stabilize radicals. With a group that is instead electron-withdrawing, the SOMO then interacts with the empty π* orbital. There are no electrons occupying the higher energy orbital formed, while a new SOMO forms that is lower in energy. This results in a lower energy and higher stability of the radical species. Both donating groups and withdrawing groups stabilize radicals. Another well-known albeit weaker form of delocalization is hyperconjugation. In radical chemistry, radicals are stabilized by hyperconjugation with adjacent alkyl groups. The donation of sigma (σ) C−H bonds into the partially empty radical orbitals helps to differentiate the stabilities of radicals on tertiary, secondary, and primary carbons. Tertiary carbon radicals have three σ C-H bonds that donate, secondary radicals only two, and primary radicals only one. Therefore, tertiary radicals are the most stable and primary radicals the least stable. Steric hindrance Most simply, the greater the steric hindrance the more difficult it is for reactions to take place, and the radical form is favored by default. For example, compare the hydrogen-abstracted form of N-hydroxypiperidine to the molecule TEMPO. TEMPO, or (2,2,6,6-Tetramethylpiperidin-1-yl)oxyl, is too sterically hindered by the additional methyl groups to react making it stable enough to be sold commercially in its radical form. N-Hydroxypiperidine, however, does not have the four methyl groups to impede the way of a reacting molecule so the structure is unstable. Facile H-atom donors The stability of many (or most) organic radicals is not indicated by their isolability but is manifested in their ability to function as donors of H•. This property reflects a weakened bond to hydrogen, usually O−H but sometimes N−H or C−H. This behavior is important because these H• donors serve as antioxidants in biology and in commerce. Illustrative is α-tocopherol (vitamin E). The tocopherol radical itself is insufficiently stable for isolation, but the parent molecule is a highly effective hydrogen-atom donor. The C−H bond is weakened in triphenylmethyl (trityl) derivatives. Inorganic radicals A large variety of inorganic radicals are stable and in fact isolable. Examples include most first-row transition metal complexes. With regard to main group radicals, the most abundant radical in the universe is also the most abundant chemical in the universe, H•. Most main group radicals are not however isolable, despite their intrinsic stability. Hydrogen radicals for example combine eagerly to form H2. Nitric oxide (NO) is well known example of an isolable inorganic radical. Fremy's salt (Potassium nitrosodisulfonate, (KSO3)2NO) is a related example. Many thiazyl radicals are known, despite limited extent of π resonance stabilization. Many radicals can be envisioned as the products of breaking of covalent bonds by homolysis. The homolytic bond dissociation energies, usually abbreviated as "ΔH°" are a measure of bond strength. Splitting H2 into 2 H•, for example, requires a ΔH° of +435 kJ/mol, while splitting Cl2 into two Cl• requires a ΔH° of +243 kJ/mol. For weak bonds, homolysis can be induced thermally. Strong bonds require high energy photons or even flames to induce homolysis. Diradicals Diradicals are molecules containing two radical centers. Dioxygen (O2) is an important example of a stable diradical. Singlet oxygen, the lowest-energy non-radical state of dioxygen, is less stable than the diradical due to Hund's rule of maximum multiplicity. The relative stability of the oxygen diradical is primarily due to the spin-forbidden nature of the triplet-singlet transition required for it to grab electrons, i.e., "oxidize". The diradical state of oxygen also results in its paramagnetic character, which is demonstrated by its attraction to an external magnet. Diradicals can also occur in metal-oxo complexes, lending themselves for studies of spin forbidden reactions in transition metal chemistry. Carbenes in their triplet state can be viewed as diradicals centred on the same atom, while these are usually highly reactive persistent carbenes are known, with N-heterocyclic carbenes being the most common example. Triplet carbenes and nitrenes are diradicals. Their chemical properties are distinct from the properties of their singlet analogues. Occurrence of radicals Combustion A familiar radical reaction is combustion. The oxygen molecule is a stable diradical, best represented by •O–O•. Because spins of the electrons are parallel, this molecule is stable. While the ground state of oxygen is this unreactive spin-unpaired (triplet) diradical, an extremely reactive spin-paired (singlet) state is available. For combustion to occur, the energy barrier between these must be overcome. This barrier can be overcome by heat, requiring high temperatures. The triplet-singlet transition is also "forbidden". This presents an additional barrier to the reaction. It also means molecular oxygen is relatively unreactive at room temperature except in the presence of a catalytic heavy atom such as iron or copper. Combustion consists of various radical chain reactions that the singlet radical can initiate. The flammability of a given material strongly depends on the concentration of radicals that must be obtained before initiation and propagation reactions dominate leading to combustion of the material. Once the combustible material has been consumed, termination reactions again dominate and the flame dies out. As indicated, promotion of propagation or termination reactions alters flammability. For example, because lead itself deactivates radicals in the gasoline-air mixture, tetraethyl lead was once commonly added to gasoline. This prevents the combustion from initiating in an uncontrolled manner or in unburnt residues (engine knocking) or premature ignition (preignition). When a hydrocarbon is burned, a large number of different oxygen radicals are involved. Initially, hydroperoxyl radical (HOO•) are formed. These then react further to give organic hydroperoxides that break up into hydroxyl radicals (HO•). Polymerization Many polymerization reactions are initiated by radicals. Polymerization involves an initial radical adding to non-radical (usually an alkene) to give new radicals. This process is the basis of the radical chain reaction. The art of polymerization entails the method by which the initiating radical is introduced. For example, methyl methacrylate (MMA) can be polymerized to produce Poly(methyl methacrylate) (PMMA – Plexiglas or Perspex) via a repeating series of radical addition steps: Newer radical polymerization methods are known as living radical polymerization. Variants include reversible addition-fragmentation chain transfer (RAFT) and atom transfer radical polymerization (ATRP). Being a prevalent radical, O2 reacts with many organic compounds to generate radicals together with the hydroperoxide radical. Drying oils and alkyd paints harden due to radical crosslinking initiated by oxygen from the atmosphere. Atmospheric radicals The most common radical in the lower atmosphere is molecular dioxygen. Photodissociation of source molecules produces other radicals. In the lower atmosphere, important radical are produced by the photodissociation of nitrogen dioxide to an oxygen atom and nitric oxide (see below), which plays a key role in smog formation—and the photodissociation of ozone to give the excited oxygen atom O(1D) (see below). The net and return reactions are also shown ( and , respectively). In the upper atmosphere, the photodissociation of normally unreactive chlorofluorocarbons (CFCs) by solar ultraviolet radiation is an important source of radicals (see eq. 1 below). These reactions give the chlorine radical, Cl•, which catalyzes the conversion of ozone to O2, thus facilitating ozone depletion (– below). Such reactions cause the depletion of the ozone layer, especially since the chlorine radical is free to engage in another reaction chain; consequently, the use of chlorofluorocarbons as refrigerants has been restricted. In biology Radicals play important roles in biology. Many of these are necessary for life, such as the intracellular killing of bacteria by phagocytic cells such as granulocytes and macrophages. Radicals are involved in cell signalling processes, known as redox signaling. For example, radical attack of linoleic acid produces a series of 13-hydroxyoctadecadienoic acids and 9-hydroxyoctadecadienoic acids, which may act to regulate localized tissue inflammatory and/or healing responses, pain perception, and the proliferation of malignant cells. Radical attacks on arachidonic acid and docosahexaenoic acid produce a similar but broader array of signaling products. Radicals may also be involved in Parkinson's disease, senile and drug-induced deafness, schizophrenia, and Alzheimer's. The classic free-radical syndrome, the iron-storage disease hemochromatosis, is typically associated with a constellation of free-radical-related symptoms including movement disorder, psychosis, skin pigmentary melanin abnormalities, deafness, arthritis, and diabetes mellitus. The free-radical theory of aging proposes that radicals underlie the aging process itself. Similarly, the process of mitohormesis suggests that repeated exposure to radicals may extend life span. Because radicals are necessary for life, the body has a number of mechanisms to minimize radical-induced damage and to repair damage that occurs, such as the enzymes superoxide dismutase, catalase, glutathione peroxidase and glutathione reductase. In addition, antioxidants play a key role in these defense mechanisms. These are often the three vitamins, vitamin A, vitamin C and vitamin E and polyphenol antioxidants. Furthermore, there is good evidence indicating that bilirubin and uric acid can act as antioxidants to help neutralize certain radicals. Bilirubin comes from the breakdown of red blood cells' contents, while uric acid is a breakdown product of purines. Too much bilirubin, though, can lead to jaundice, which could eventually damage the central nervous system, while too much uric acid causes gout. Reactive oxygen species Reactive oxygen species or ROS are species such as superoxide, hydrogen peroxide, and hydroxyl radical, commonly associated with cell damage. ROS form as a natural by-product of the normal metabolism of oxygen and have important roles in cell signaling. Two important oxygen-centered radicals are superoxide and hydroxyl radical. They derive from molecular oxygen under reducing conditions. However, because of their reactivity, these same radicals can participate in unwanted side reactions resulting in cell damage. Excessive amounts of these radicals can lead to cell injury and death, which may contribute to many diseases such as cancer, stroke, myocardial infarction, diabetes and major disorders. Many forms of cancer are thought to be the result of reactions between radicals and DNA, potentially resulting in mutations that can adversely affect the cell cycle and potentially lead to malignancy. Some of the symptoms of aging such as atherosclerosis are also attributed to radical induced oxidation of cholesterol to 7-ketocholesterol. In addition radicals contribute to alcohol-induced liver damage, perhaps more than alcohol itself. Radicals produced by cigarette smoke are implicated in inactivation of alpha 1-antitrypsin in the lung. This process promotes the development of emphysema. Oxybenzone has been found to form radicals in sunlight, and therefore may be associated with cell damage as well. This only occurred when it was combined with other ingredients commonly found in sunscreens, like titanium oxide and octyl methoxycinnamate. ROS attack the polyunsaturated fatty acid, linoleic acid, to form a series of 13-hydroxyoctadecadienoic acid and 9-hydroxyoctadecadienoic acid products that serve as signaling molecules that may trigger responses that counter the tissue injury which caused their formation. ROS attacks other polyunsaturated fatty acids, e.g. arachidonic acid and docosahexaenoic acid, to produce a similar series of signaling products. Reactive oxygen species are also used in controlled reactions involving singlet dioxygen known as type II photooxygenation reactions after Dexter energy transfer (triplet-triplet annihilation) from natural triplet dioxygen and triplet excited state of a photosensitizer. Typical chemical transformations with this singlet dioxygen species involve, among others, conversion of cellulosic biowaste into new poylmethine dyes. Depiction in chemical reactions In chemical equations, radicals are frequently denoted by a dot placed immediately to the right of the atomic symbol or molecular formula as follows: Radical reaction mechanisms use single-headed arrows to depict the movement of single electrons: The homolytic cleavage of the breaking bond is drawn with a "fish-hook" arrow to distinguish from the usual movement of two electrons depicted by a standard curly arrow. The second electron of the breaking bond also moves to pair up with the attacking radical electron. Radicals also take part in radical addition and radical substitution as reactive intermediates. Chain reactions involving radicals can usually be divided into three distinct processes. These are initiation, propagation, and termination. Initiation reactions are those that result in a net increase in the number of radicals. They may involve the formation of radicals from stable species as in Reaction 1 above or they may involve reactions of radicals with stable species to form more radicals. Propagation reactions are those reactions involving radicals in which the total number of radicals remains the same. Termination reactions are those reactions resulting in a net decrease in the number of radicals. Typically two radicals combine to form a more stable species, for example: 2 Cl• → Cl2 History and nomenclature Until late in the 20th century the word "radical" was used in chemistry to indicate any connected group of atoms, such as a methyl group or a carboxyl, whether it was part of a larger molecule or a molecule on its own. A radical is often known as an R group. The qualifier "free" was then needed to specify the unbound case. Following recent nomenclature revisions, a part of a larger molecule is now called a functional group or substituent, and "radical" now implies "free". However, the old nomenclature may still appear in some books. The term radical was already in use when the now obsolete radical theory was developed. Louis-Bernard Guyton de Morveau introduced the phrase "radical" in 1785 and the phrase was employed by Antoine Lavoisier in 1789 in his Traité Élémentaire de Chimie. A radical was then identified as the root base of certain acids (the Latin word "radix" meaning "root"). Historically, the term radical in radical theory was also used for bound parts of the molecule, especially when they remain unchanged in reactions. These are now called functional groups. For example, methyl alcohol was described as consisting of a methyl "radical" and a hydroxyl "radical". Neither are radicals in the modern chemical sense, as they are permanently bound to each other, and have no unpaired, reactive electrons; however, they can be observed as radicals in mass spectrometry when broken apart by irradiation with energetic electrons. In a modern context the first organic (carbon–containing) radical identified was the triphenylmethyl radical, (C6H5)3C•. This species was discovered by Moses Gomberg in 1900. In 1933 Morris S. Kharasch and Frank Mayo proposed that free radicals were responsible for anti-Markovnikov addition of hydrogen bromide to allyl bromide. In most fields of chemistry, the historical definition of radicals contends that the molecules have nonzero electron spin. However, in fields including spectroscopy and astrochemistry, the definition is slightly different. Gerhard Herzberg, who won the Nobel prize for his research into the electron structure and geometry of radicals, suggested a looser definition of free radicals: "any transient (chemically unstable) species (atom, molecule, or ion)". The main point of his suggestion is that there are many chemically unstable molecules that have zero spin, such as C2, C3, CH2 and so on. This definition is more convenient for discussions of transient chemical processes and astrochemistry; therefore researchers in these fields prefer to use this loose definition.
Physical sciences
Chemical reactions
null
19916615
https://en.wikipedia.org/wiki/Electron%20shell
Electron shell
In chemistry and atomic physics, an electron shell may be thought of as an orbit that electrons follow around an atom's nucleus. The closest shell to the nucleus is called the "1 shell" (also called the "K shell"), followed by the "2 shell" (or "L shell"), then the "3 shell" (or "M shell"), and so on further and further from the nucleus. The shells correspond to the principal quantum numbers (n = 1, 2, 3, 4 ...) or are labeled alphabetically with the letters used in X-ray notation (K, L, M, ...). A useful guide when understanding electron shells in atoms is to note that each row on the conventional periodic table of elements represents an electron shell. Each shell can contain only a fixed number of electrons: the first shell can hold up to two electrons, the second shell can hold up to eight electrons, the third shell can hold up to 18, continuing as the general formula of the nth shell being able to hold up to 2(n2) electrons. For an explanation of why electrons exist in these shells, see electron configuration. Each shell consists of one or more subshells, and each subshell consists of one or more atomic orbitals. History In 1913, Niels Bohr proposed a model of the atom, giving the arrangement of electrons in their sequential orbits. At that time, Bohr allowed the capacity of the inner orbit of the atom to increase to eight electrons as the atoms got larger, and "in the scheme given below the number of electrons in this [outer] ring is arbitrary put equal to the normal valency of the corresponding element". Using these and other constraints, he proposed configurations that are in accord with those now known only for the first six elements. "From the above we are led to the following possible scheme for the arrangement of the electrons in light atoms:" The shell terminology comes from Arnold Sommerfeld's modification of the 1913 Bohr model. During this period Bohr was working with Walther Kossel, whose papers in 1914 and in 1916 called the orbits "shells". Sommerfeld retained Bohr's planetary model, but added mildly elliptical orbits (characterized by additional quantum numbers and m) to explain the fine spectroscopic structure of some elements. The multiple electrons with the same principal quantum number (n) had close orbits that formed a "shell" of positive thickness instead of the circular orbit of Bohr's model which orbits called "rings" were described by a plane. The existence of electron shells was first observed experimentally in Charles Barkla's and Henry Moseley's X-ray absorption studies. Moseley's work did not directly concern the study of electron shells, because he was trying to prove that the periodic table was not arranged by weight, but by the charge of the protons in the nucleus. However, because the number of electrons in an electrically neutral atom equals the number of protons, this work was extremely important to Niels Bohr who mentioned Moseley's work several times in his 1962 interview. Moseley was part of Rutherford's group, as was Niels Bohr. Moseley measured the frequencies of X-rays emitted by every element between calcium and zinc and found that the frequencies became greater as the elements got heavier. This led to the theory that electrons were emitting X-rays when they were shifted to lower shells. This led to the conclusion that the electrons were in Kossel's shells with a definite limit per shell, labeling them with the letters K, L, M, N, O, P, and Q. The origin of this terminology was alphabetic. Barkla, who worked independently from Moseley as an X-ray spectrometry experimentalist, first noticed two distinct types of scattering from shooting X-rays at elements in 1909 and named them "A" and "B". Barkla described these two types of X-ray diffraction: the first was unconnected with the type of material used in the experiment and could be polarized. The second diffraction beam he called "fluorescent" because it depended on the irradiated material. It was not known what these lines meant at the time, but in 1911 Barkla decided there might be scattering lines previous to "A", so he began at "K". However, later experiments indicated that the K absorption lines are produced by the innermost electrons. These letters were later found to correspond to the n values 1, 2, 3, etc. that were used in the Bohr model. They are used in the spectroscopic Siegbahn notation. The work of assigning electrons to shells was continued from 1913 to 1925 by many chemists and a few physicists. Niels Bohr was one of the few physicists who followed the chemist's work of defining the periodic table, while Arnold Sommerfeld worked more on trying to make a relativistic working model of the atom that would explain the fine structure of the spectra from a classical orbital physics standpoint through the Atombau approach. Einstein and Rutherford, who did not follow chemistry, were unaware of the chemists who were developing electron shell theories of the periodic table from a chemistry point of view, such as Irving Langmuir, Charles Bury, J.J. Thomson, and Gilbert Lewis, who all introduced corrections to Bohr's model such as a maximum of two electrons in the first shell, eight in the next and so on, and were responsible for explaining valency in the outer electron shells, and the building up of atoms by adding electrons to the outer shells. So when Bohr outlined his electron shell atomic theory in 1922, there was no mathematical formula for the theory. So Rutherford said he was hard put "to form an idea of how you arrive at your conclusions". Einstein said of Bohr's 1922 paper that his "electron-shells of the atoms together with their significance for chemistry appeared to me like a miracle – and appears to me as a miracle even today". Arnold Sommerfeld, who had followed the Atombau structure of electrons instead of Bohr who was familiar with the chemists' views of electron structure, spoke of Bohr's 1921 lecture and 1922 article on the shell model as "the greatest advance in atomic structure since 1913". However, the electron shell development of Niels Bohr was basically the same theory as that of the chemist Charles Rugeley Bury in his 1921 paper. As work continued on the electron shell structure of the Sommerfeld-Bohr Model, Sommerfeld had introduced three "quantum numbers n, k, and m, that described the size of the orbit, the shape of the orbit, and the direction in which the orbit was pointing." Because we use k for the Boltzmann constant, the azimuthal quantum number was changed to ℓ. When the modern quantum mechanics theory was put forward based on Heisenberg's matrix mechanics and Schrödinger's wave equation, these quantum numbers were kept in the current quantum theory but were changed to n being the principal quantum number, and m being the magnetic quantum number. However, the final form of the electron shell model still in use today for the number of electrons in shells was discovered in 1923 by Edmund Stoner, who introduced the principle that the nth shell was described by 2(n2). Seeing this in 1925, Wolfgang Pauli added a fourth quantum number, "spin", during the old quantum theory period of the Sommerfeld-Bohr Solar System atom to complete the modern electron shell theory. Subshells Each shell is composed of one or more subshells, which are themselves composed of atomic orbitals. For example, the first (K) shell has one subshell, called 1s; the second (L) shell has two subshells, called 2s and 2p; the third shell has 3s, 3p, and 3d; the fourth shell has 4s, 4p, 4d and 4f; the fifth shell has 5s, 5p, 5d, and 5f and can theoretically hold more in the 5g subshell that is not occupied in the ground-state electron configuration of any known element. The various possible subshells are shown in the following table: The first column is the "subshell label", a lowercase-letter label for the type of subshell. For example, the "4s subshell" is a subshell of the fourth (N) shell, with the type (s) described in the first row. The second column is the azimuthal quantum number (ℓ) of the subshell. The precise definition involves quantum mechanics, but it is a number that characterizes the subshell. The third column is the maximum number of electrons that can be put into a subshell of that type. For example, the top row says that each s-type subshell (1s, 2s, etc.) can have at most two electrons in it. Each of the following subshells (p, d, f, g) can have 4 more electrons than the one preceding it. The fourth column says which shells have a subshell of that type. For example, looking at the top two rows, every shell has an s subshell, while only the second shell and higher have a p subshell (i.e., there is no "1p" subshell). The final column gives the historical origin of the labels s, p, d, and f. They come from early studies of atomic spectral lines. The other labels, namely g, h, and i, are an alphabetic continuation following the last historically originated label of f. Number of electrons in each shell Each subshell is constrained to hold electrons at most, namely: Each s subshell holds at most 2 electrons Each p subshell holds at most 6 electrons Each d subshell holds at most 10 electrons Each f subshell holds at most 14 electrons Each g subshell holds at most 18 electrons Therefore, the K shell, which contains only an s subshell, can hold up to 2 electrons; the L shell, which contains an s and a p, can hold up to 2 + 6 = 8 electrons, and so forth; in general, the nth shell can hold up to 2n2 electrons. Although that formula gives the maximum in principle, that maximum is only achieved (in known elements) for the first four shells (K, L, M, N). No known element has more than 32 electrons in any one shell. This is because the subshells are filled according to the Aufbau principle. The first elements to have more than 32 electrons in one shell would belong to the g-block of period 8 of the periodic table. These elements would have some electrons in their 5g subshell and thus have more than 32 electrons in the O shell (fifth principal shell). Subshell energies and filling order Although it is sometimes stated that all the electrons in a shell have the same energy, this is an approximation. However, the electrons in one subshell do have exactly the same level of energy, with later subshells having more energy per electron than earlier ones. This effect is great enough that the energy ranges associated with shells can overlap. The filling of the shells and subshells with electrons proceeds from subshells of lower energy to subshells of higher energy. This follows the n + ℓ rule which is also commonly known as the Madelung rule. Subshells with a lower n + ℓ value are filled before those with higher n + ℓ values. In the case of equal n + ℓ values, the subshell with a lower n value is filled first. Because of this, the later shells are filled over vast sections of the periodic table. The K shell fills in the first period (hydrogen and helium), while the L shell fills in the second (lithium to neon). However, the M shell starts filling at sodium (element 11) but does not finish filling till copper (element 29), and the N shell is even slower: it starts filling at potassium (element 19) but does not finish filling till ytterbium (element 70). The O, P, and Q shells begin filling in the known elements (respectively at rubidium, caesium, and francium), but they are not complete even at the heaviest known element, oganesson (element 118). List of elements with electrons per shell The list below gives the elements arranged by increasing atomic number and shows the number of electrons per shell. At a glance, the subsets of the list show obvious patterns. In particular, every set of five elements () before each noble gas (group 18, ) heavier than helium have successive numbers of electrons in the outermost shell, namely three to seven. Sorting the table by chemical group shows additional patterns, especially with respect to the last two outermost shells. (Elements 57 to 71 belong to the lanthanides, while 89 to 103 are the actinides.) The list below is primarily consistent with the Aufbau principle. However, there are a number of exceptions to the rule; for example palladium (atomic number 46) has no electrons in the fifth shell, unlike other atoms with lower atomic number. The elements past 108 have such short half-lives that their electron configurations have not yet been measured, and so predictions have been inserted instead.
Physical sciences
Atomic physics
null
19916686
https://en.wikipedia.org/wiki/Hydrochloric%20acid
Hydrochloric acid
Hydrochloric acid, also known as muriatic acid or spirits of salt, is an aqueous solution of hydrogen chloride (HCl). It is a colorless solution with a distinctive pungent smell. It is classified as a strong acid. It is a component of the gastric acid in the digestive systems of most animal species, including humans. Hydrochloric acid is an important laboratory reagent and industrial chemical. Etymology Because it was produced from rock salt according to the methods of Johann Rudolph Glauber, hydrochloric acid was historically called by European alchemists spirits of salt or acidum salis (salt acid). Both names are still used, especially in other languages, such as , , , , , , , , , , (ensan), (yánsuān), and (yeomsan). Gaseous HCl was called marine acid air. The name muriatic acid has the same origin (muriatic means "pertaining to brine or salt", hence muriate means hydrochloride), and this name is still sometimes used. The name hydrochloric acid was coined by the French chemist Joseph Louis Gay-Lussac in 1814. History 9th–10th century In the early tenth century, the Persian physician and alchemist Abu Bakr al-Razi (–925, Latin: Rhazes) conducted experiments with sal ammoniac (ammonium chloride) and vitriol (hydrated sulfates of various metals), which he distilled together, thus producing the gas hydrogen chloride. In doing so, al-Razi may have stumbled upon a primitive method for producing hydrochloric acid, as perhaps manifested in the following recipe from his ("The Book of Secrets"): However, it appears that in most of his experiments al-Razi disregarded the gaseous products, concentrating instead on the color changes that could be effected in the residue. According to Robert P. Multhauf, hydrogen chloride was produced many times without clear recognition that, by dissolving it in water, hydrochloric acid may be produced. 11th–13th century Drawing on al-Razi's experiments, the ("On Alums and Salts"), an eleventh- or twelfth-century Arabic text falsely attributed to al-Razi and translated into Latin by Gerard of Cremona (11441187), described the heating of metals with various salts, which in the case of mercury resulted in the production of mercury(II) chloride (corrosive sublimate). In this process, hydrochloric acid actually started to form, but it immediately reacted with the mercury to produce corrosive sublimate. Thirteenth-century Latin alchemists, for whom the De aluminibus et salibus was one of the main reference works, were fascinated by the chlorinating properties of corrosive sublimate, and they soon discovered that when the metals are eliminated from the process of heating vitriols, alums, and salts, strong mineral acids can directly be distilled. 14th–15th century Aqua regia One important invention that resulted from the discovery of the mineral acids is aqua regia, a mixture of nitric acid and hydrochloric acid in a 1:3 proportion, capable of dissolving gold. This was first described in pseudo-Geber's De inventione veritatis ("On the Discovery of Truth", after ), where aqua regia was prepared by adding ammonium chloride to nitric acid. The fact that aqua regia typically is defined as a mixture of nitric acid and hydrochloric acid does not mean that hydrochloric acid was discovered before or simultaneously with aqua regia. The isolation of hydrochloric acid happened about 300 years later. The production of hydrochloric acid itself (i.e., as an isolated substance rather than as already mixed with nitric acid) depended on the use of more efficient cooling apparatus, which would only develop in subsequent centuries. 16th–17th century Isolation of hydrochloric acid From the point of view of Western history of chemistry, hydrochloric acid was the last of the three well-known mineral acids for which the method of its production appeared in the literature. Recipes for its production started to appear in the late sixteenth century. The earliest recipes for the production of hydrochloric acid are found in Giovanni Battista Della Porta's (1535–1615) Magiae naturalis ("Natural Magic") and in the works of other contemporary chemists like Andreas Libavius (–1616), Jean Beguin (1550–1620), and Oswald Croll (–1609). Among the historians who have written about this are German chemists Hermann Franz Moritz Kopp (1845) and Edmund Oscar von Lippmann (1938), mining engineer (and future U.S. president) Herbert Hoover with his wife geologist Lou Henry Hoover (1912), Dutch chemist Robert Jacobus Forbes (1948), American chemist Mary Elvira Weeks (1956), and British chemists F. Sherwood Taylor (1957) and J. R. Partington (1960). Italian chemist Ladislao Reti have summarized the result of their efforts thus: Dissolving metals The knowledge of mineral acids such as hydrochloric acid would be of key importance to seventeenth-century chemists like Daniel Sennert (1572–1637) and Robert Boyle (1627–1691), who used their capability to rapidly dissolve metals in their demonstrations of the composite nature of bodies. Industrial developments During the Industrial Revolution in Europe, demand for alkaline substances increased. A new industrial process developed by Nicolas Leblanc of Issoudun, France enabled cheap large-scale production of sodium carbonate (soda ash). In this Leblanc process, common salt is converted to soda ash, using sulfuric acid, limestone, and coal, releasing hydrogen chloride as a by-product. Until the British Alkali Act 1863 and similar legislation in other countries, the excess HCl was often vented into the air. An early exception was the Bonnington Chemical Works where, in 1830, the HCl began to be captured and the hydrochloric acid produced was used in making sal ammoniac (ammonium chloride). After the passage of the act, soda ash producers were obliged to absorb the waste gas in water, producing hydrochloric acid on an industrial scale. In the 20th century, the Leblanc process was effectively replaced by the Solvay process without a hydrochloric acid by-product. Since hydrochloric acid was already fully settled as an important chemical in numerous applications, the commercial interest initiated other production methods, some of which are still used today. After 2000, hydrochloric acid is mostly made by absorbing by-product hydrogen chloride from industrial organic compounds production. Chemical properties Gaseous hydrogen chloride is a molecular compound with a covalent bond between the hydrogen and chlorine atoms. In aqueous solutions dissociation is complete, with the formation of chloride ions and hydrated hydrogen ions (hydronium ions). A combined IR, Raman, X-ray, and neutron diffraction study of concentrated hydrochloric acid showed that the hydronium ion forms hydrogen bonded complexes with other water molecules. (See Hydronium for further discussion of this issue.) The pKa value of hydrochloric acid in aqueous solution is estimated theoretically to be −5.9. A solution of hydrogen chloride in water behaves as a strong acid: the concentration of HCl molecules is effectively zero. Physical properties Physical properties of hydrochloric acid, such as boiling and melting points, density, and pH, depend on the concentration or molarity of HCl in the aqueous solution. They range from those of water at very low concentrations approaching 0% HCl to values for fuming hydrochloric acid at over 40% HCl. Hydrochloric acid as the binary (two-component) mixture of HCl and H2O has a constant-boiling azeotrope at 20.2% HCl and . There are four constant-crystallization eutectic points for hydrochloric acid, between the crystal form of [H3O]Cl (68% HCl), [H5O2]Cl (51% HCl), [H7O3]Cl (41% HCl), [H3O]Cl·5H2O (25% HCl), and ice (0% HCl). There is also a metastable eutectic point at 24.8% between ice and the [H7O3]Cl crystallization. They are all hydronium salts. Production Hydrochloric acid is usually prepared industrially by dissolving hydrogen chloride in water. Hydrogen chloride can be generated in many ways, and thus several precursors to hydrochloric acid exist. The large-scale production of hydrochloric acid is almost always integrated with the industrial scale production of other chemicals, such as in the chloralkali process which produces hydroxide, hydrogen, and chlorine, the latter of which can be combined to produce HCl. Hydrogen chloride is produced by combining chlorine and hydrogen: As the reaction is exothermic, the installation is called an HCl oven or HCl burner. The resulting hydrogen chloride gas is absorbed in deionized water, resulting in chemically pure hydrochloric acid. This reaction can give a very pure product, e.g. for use in the food industry. Industrial market Hydrochloric acid is produced in solutions up to 38% HCl (concentrated grade). Higher concentrations up to just over 40% are chemically possible, but the evaporation rate is then so high that storage and handling require extra precautions, such as pressurization and cooling. Bulk industrial-grade is therefore 30% to 35%, optimized to balance transport efficiency and product loss through evaporation. In the United States, solutions of between 20% and 32% are sold as muriatic acid. Solutions for household purposes in the US, mostly cleaning, are typically 10% to 12%, with strong recommendations to dilute before use. In the United Kingdom, where it is sold as "Spirits of Salt" for domestic cleaning, the potency is the same as the US industrial grade. In other countries, such as Italy, hydrochloric acid for domestic or industrial cleaning is sold as "Acido Muriatico", and its concentration ranges from 5% to 32%. Major producers worldwide include Dow Chemical at 2 million tonnes annually (Mt/year), calculated as HCl gas, Georgia Gulf Corporation, Tosoh Corporation, Akzo Nobel, and Tessenderlo at 0.5 to 1.5 Mt/year each. Total world production, for comparison purposes expressed as HCl, is estimated at 20 Mt/year, with 3 Mt/year from direct synthesis, and the rest as secondary product from organic and similar syntheses. By far, most hydrochloric acid is consumed captively by the producer. The open world market size is estimated at 5 Mt/year. Applications Hydrochloric acid is a strong inorganic acid that is used in many industrial processes such as refining metal. The application often determines the required product quality. Hydrogen chloride, not hydrochloric acid, is used more widely in industrial organic chemistry, e.g. for vinyl chloride and dichloroethane. Pickling of steel One of the most important applications of hydrochloric acid is in the pickling of steel, to remove rust or iron oxide scale from iron or steel before subsequent processing, such as extrusion, rolling, galvanizing, and other techniques. Technical quality HCl at typically 18% concentration is the most commonly used pickling agent for the pickling of carbon steel grades. The spent acid has long been reused as iron(II) chloride (also known as ferrous chloride) solutions, but high heavy-metal levels in the pickling liquor have decreased this practice. The steel pickling industry has developed hydrochloric acid regeneration processes, such as the spray roaster or the fluidized bed HCl regeneration process, which allow the recovery of HCl from spent pickling liquor. The most common regeneration process is the pyrohydrolysis process, applying the following formula: By recuperation of the spent acid, a closed acid loop is established. The iron(III) oxide by-product of the regeneration process is valuable, used in a variety of secondary industries. Production of inorganic compounds Akin to its use for pickling, hydrochloric acid is used to dissolve many metals, metal oxides and metal carbonates. The conversions are often depicted in simplified equations: These processes are used to produce metal chlorides for analysis or further production. pH control and neutralization Hydrochloric acid can be used to regulate the acidity (pH) of solutions. In industry demanding purity (food, pharmaceutical, drinking water), high-quality hydrochloric acid is used to control the pH of process water streams. In less-demanding industry, technical quality hydrochloric acid suffices for neutralizing waste streams and swimming pool pH control. Regeneration of ion exchangers High-quality hydrochloric acid is used in the regeneration of ion exchange resins. Cation exchange is widely used to remove ions such as Na+ and Ca2+ from aqueous solutions, producing demineralized water. The acid is used to rinse the cations from the resins. Na+ is replaced with H+ and Ca2+ with 2 H+. Ion exchangers and demineralized water are used in all chemical industries, drinking water production, and many food industries. Laboratory use Of the common strong mineral acids in chemistry, hydrochloric acid is the monoprotic acid least likely to undergo an interfering oxidation-reduction reaction. It is one of the least hazardous strong acids to handle; despite its acidity, it contains the non-reactive and non-toxic chloride ion. Intermediate-strength hydrochloric acid solutions are quite stable upon storage, maintaining their concentrations over time. These attributes, plus the fact that it is available as a pure reagent, make hydrochloric acid an excellent acidifying reagent. It is also inexpensive. Hydrochloric acid is the preferred acid in titration for determining the amount of bases. Strong acid titrants give more precise results due to a more distinct endpoint. Azeotropic, or "constant-boiling", hydrochloric acid (roughly 20.2%) can be used as a primary standard in quantitative analysis, although its exact concentration depends on the atmospheric pressure when it is prepared. Other Hydrochloric acid is used for a large number of small-scale applications, such as leather processing, household cleaning, and building construction. Oil production may be stimulated by injecting hydrochloric acid into the rock formation of an oil well, dissolving a portion of the rock, and creating a large-pore structure. Oil well acidizing is a common process in the North Sea oil production industry. Hydrochloric acid has been used for dissolving calcium carbonate, e.g. such things as de-scaling kettles and for cleaning mortar off brickwork. When used on brickwork the reaction with the mortar only continues until the acid has all been converted, producing calcium chloride, carbon dioxide, and water: Many chemical reactions involving hydrochloric acid are applied in the production of food, food ingredients, and food additives. Typical products include aspartame, fructose, citric acid, lysine, hydrolyzed vegetable protein as food enhancer, and in gelatin production. Food-grade (extra-pure) hydrochloric acid can be applied when needed for the final product. Presence in living organisms Gastric acid is one of the main secretions of the stomach. It consists mainly of hydrochloric acid and acidifies the stomach content to a pH of 1 to 2. Chloride (Cl−) and hydrogen (H+) ions are secreted separately in the stomach fundus region at the top of the stomach by parietal cells of the gastric mucosa into a secretory network called canaliculi before it enters the stomach lumen. Gastric acid acts as a barrier against microorganisms to prevent infections and is important for the digestion of food. Its low pH denatures proteins and thereby makes them susceptible to degradation by digestive enzymes such as pepsin. The low pH also activates the enzyme precursor pepsinogen into the active enzyme pepsin by self-cleavage. After leaving the stomach, the hydrochloric acid of the chyme is neutralized in the duodenum by bicarbonate. The stomach itself is protected from the strong acid by the secretion of a thick mucus layer, and by secretin induced buffering with sodium bicarbonate. Heartburn or peptic ulcers can develop when these mechanisms fail. Drugs of the antihistaminic and proton pump inhibitor classes can inhibit the production of acid in the stomach, and antacids are used to neutralize excessive existing acid. Hydrochloric acid is also used by osteoclasts alongside proteases for bone resorption. Safety Being a strong acid, hydrochloric acid is corrosive to living tissue and to many materials, but not to rubber. Typically, rubber protective gloves and related protective gear are used when handling concentrated solutions. Vapors or mists are a respiratory hazard, which can be partially mitigated by use of a respirator equipped with cartridges specifically designed to capture hydrochloric acid. Airborne acid is an irritant to the eyes, and may require the use of protective goggles or a facemask. Legal status Hydrochloric acid has been listed as a Table II precursor under the 1988 United Nations Convention Against Illicit Traffic in Narcotic Drugs and Psychotropic Substances because of its use in the production of heroin, cocaine, and methamphetamine.
Physical sciences
Inorganic compounds
null
6436318
https://en.wikipedia.org/wiki/Istanbul%20Metro
Istanbul Metro
The Istanbul Metro () is a rapid transit railway network that serves the city of Istanbul, Turkey. Apart from the M11 line, which is operated by TCDD Taşımacılık, the system is operated by Metro Istanbul, a public enterprise controlled by the Istanbul Metropolitan Municipality. The oldest section of the metro is the M1 line, which opened on 3 September, 1989. As of 2024, the system now includes 159 stations in service, with 36 more under construction. With 243.3 kilometers, Istanbul has the 21st longest metro line in the world and the 4th longest in Europe as of 2024. The system consists of eleven lines: the lines designated M1A, M1B, M2, M3, M6, M7, M9 and M11 are on the European side of the Bosporus, while lines M4, M5 and M8 are on the Asian side. Due to Istanbul's unique geography and the depth of the Bosporus strait which divides the city, the European and Asian metro networks do not connect directly. The two parts of the city are linked through the Marmaray commuter rail line, which is connected to the metro in several places. Four metro lines are under construction on the Asian side: M10 (Pendik Merkez–Fevzi Çakmak), M12 (60. Yıl Parkı–Kazım Karabekir), M13 (Emek–Yenidoğan) and M14 (Altunizade–Bosna Bulvarı). Additionally, extension works on the M7 and M11 lines (on the European side) and the M4 and M5 lines (on the Asian side) are underway. In addition to the Marmaray commuter rail, the metro connects to the F1, Tünel (F2), F3 and F4 funicular lines and with the network of the Istanbul Tram, Metrobüs and the cable cars. History The oldest underground urban rail line in Istanbul is the Tünel, which entered service on 17 January 1875. It is the world's second-oldest underground urban rail line after the London Underground which was built in 1863, and the first underground urban rail line in continental Europe, however this is not recognised as the opening of the metro due to the line being Funicular. The first master plan for a full metro network in Istanbul, titled Avant Projet d'un Métropolitain à Constantinople and conceived by the French engineer L. Guerby, dates to 10 January 1912. The plan comprised a total of 24 stations between the Topkapı and Şişli districts and included a connection through the Golden Horn. Each station would have a platform next to the rail line, while the distance between stations varied from . The blueprints of the project, which was never realized, are today displayed at the Istanbul Technical University Museum. In 1936 the French urban planner Henri Prost proposed a metro network between the districts of Taksim and Beyazıt, to the north and south of the Golden Horn, respectively. In October 1951 the Dutch firm Nedeco proposed a similar route between Taksim and Beyazıt, and in September 1952 the Director of the Paris Transportation Department, Marc Langevin, prepared a 14-chapter report together with his associate Louis Meizzonet for the implementation of the project and its integration with the other means of public transportation in the city. However, these plans never came into effect and all proposals were put on hold until 1987, when the planning for the current Istanbul Metro was made. Construction works for the first 'modern' mass transit railway system started in 1989, with the first stations opening in September. İstanbul Ulaşım A.Ş. (now known as Metro Istanbul) was founded the previous year to operate the system. The M1 was initially called "Hafif Metro" (which literally translates as "light metro"). Although it was built as a fully grade separated line, the M1 line operates with shorter trainsets and shorter station platforms than is standard on a traditional metro line, hence its "light metro" designation. The M1 line was later extended from Aksaray towards the western suburbs, reaching Atatürk Airport in the southwest in 2002. Construction of the M2 line began on 11 September 1992, but faced many challenges due to the numerous archaeological sites that were discovered during the drilling process, which slowed down or fully stopped the construction of many stations, especially in the south. Taking into account the seismic activity in Istanbul, the entire network was built with the cut-and-cover method to withstand an earthquake of up to 9.0 on the Richter magnitude scale. The first section between Taksim and 4. Levent entered service, after some delays, on 16 September 2000. This line is long and has 6 stations, which all look similar but are in different colours. In 2000, there were 8 Alstom-built 4-car train sets in service, which ran every 5 minutes on average and transported 130,000 passengers daily. On 30 January 2009, the first train sets built by Eurotem entered service. Eurotem will build a total of 92 new trainsets for the M2 line. As of 30 January 2009, a total of 34 trainsets, each with 4 cars, were being used on the M2 line. A northern extension from 4. Levent to Maslak was opened on 30 January 2009. On 2 September 2010, the northern (temporary) terminus of Darüşşafaka followed up. The southern extension of the M2 line from Taksim to Yenikapı, across the Golden Horn with Haliç station on the bridge and underground through the historic peninsula, entered service on 15 February 2014. The Taksim-Yenikapı extension is long, with four stations. The total cost of the extension was $593 million. At Yenikapı, it will intersect with the extended M1 line and the Marmaray commuter line, which since its opening in 2013 has offered a connection between the Asian and European sides of the city. The trip between the Şişhane station in Beyoğlu and the Haciosman station in Maslak is long and takes 27 minutes; including Şişhane - Taksim (, 2 minutes), Taksim - 4. Levent (, 12 minutes), and 4. Levent - Haciosman (, 12 minutes.) The total length of the European side of the M2 line will reach when all 16 stations from Hacıosman to Yenikapı will be completed; not including the -long Golden Horn Metro Bridge, the long Taksim-Kabataş tunnel connection with the Seabus port, and the long Yenikapı-Aksaray tunnel connecting the M1 line to the Yenikapı Transfer Center. On the Asian side, construction is in progress of the remaining portion of the long M4 line from Kadıköy to Kaynarca, yielding a total of 19 stations. It cost €751 million and was built by the Astaldi / Makyol / Gülermak consortium. The first section opened on 17 August 2012, terminating in Kartal. Construction of the long M5 line from Üsküdar via Ümraniye to Çekmeköy started in March 2012. Operations The Metro operates from 06:00 AM to 12 AM every 6–12 minutes, except for the M8 line, which temporarily operates from 06:00 AM to 23:00 PM. During peak hours the intervals could be reduced to 3 to 4 minutes. The metro has a flat fare of 17.70 TL, a student rate of 8.64 TL, and a fare of 12.67 TL for teachers and senior residents, when used with the Istanbulkart, the city's universal contactless payment card. Night operations On 30 August 2019 as a gesture for the Victory Day of Turkey, Mayor of Istanbul, Ekrem İmamoğlu, announced that many lines of Istanbul were going to provide 24 hours service during the weekends with trains at 20 minutes interval. Since August 2019, the Istanbul Metro started to provide 24 hours service for some lines on weekends and public holidays, but this service was discontinued in March 2020 due to coronavirus pandemic. On March 5, 2022 Metro Istanbul announced that the night metro service would be re-started on the following lines: M1A Yenikapı-Atatürk Airport, M1B Yenikapı-Kirazli, M2 Yenikapı-Hacıosman, M4 Kadıköy-Sabiha Gökçen Airport, M5 Üsküdar-Çekmeköy, and M6 Levent-Boğaziçi University/Hisarüstü, and later added M7 Mecidiyeköy-Mahmutbey on July 14, 2022. Lines Each line is given a different color and name. The letter "M" stands for metro, while "T" stands for tram, "F" stands for funicular, "TF" stands for cable car and "B" for suburban. Frutiger LT Pro is used as the font for the icons. Stations The Istanbul Metro system has a total of 159 stations in operation with 35 more under construction. Since the majority of the system is underground, stations are generally accessed going down from street level. At every station entrance there is a post with the Istanbul metro "M" logo and the station name underneath it except for the M11 line which has the "U" logo next to its own entrances. Entrances are usually built into sidewalks along a street, although many stations of newer lines have their entrances from street level on small plazas. Like Moscow Metro, the majority of the stations of the Istanbul Metro are generally deep level due to the city's hilly geography. Out of the 159 operating stations of the Istanbul Metro: 143 are fully underground 7 are elevated stations 7 are on an embankment or at-grade 2 are partially underground Mezzanines Most stations have a mezzanine directly below street level, which allows passengers to enter the stations from multiple locations and proceed to the correct platform without having to cross any streets. Inside the mezzanines are ticket machines and turnstiles, where passengers must pay to enter fare control zones and proceed to the trains. In some stations, mezzanines connect directly to nearby buildings and structures, such as shopping malls or business centres. Each station concourse or mezzanine are patrolled by Istanbul Metro security guards to prevent fare evasion and crime. Stations with large walkways toward different exits, such as Taksim, also have travelators to cover the long walking distances. Upon entering the station, passengers may use ticket vending machines or staffed ticket booths to purchase their fare, which can be stored on the Istanbulkart contactless smart card. After entering the fare-controlled area, via the turnstiles, passengers may continue further down to the platform level. Platforms Since the different lines of the Istanbul Metro have different specifications, most notably car length, there is no uniform length for platforms. The oldest platforms of the network, built between 1989 and 2002 on the M1 line, are and can accommodate trains up to 4 cars long. Platforms on the M2, M3, M4, M7 and M11 lines span about and can each accommodate trains up to 8 cars long. Platforms on the M5 line can accommodate up to 6-car trains, whiles platforms on the M6, M8, and M9 lines are accommodate 4-car trains. Sections of platforms are subject to close during off-peak hours, especially on the M2 line. While platforms on the M2 line are open fully during peak hours, capacity is reduced to 4-car trains during off-peak hours. Platforms on the M5, M7, M8 and M11 lines are protected by platform edge doors. Except on funicular lines, most platforms of the Istanbul Metro consist of two side platforms or one island platform. Five stations consist of two island platforms, serving three tracks in a configuration known as the Spanish Solution. These stations are Otogar on the M1 line, Yenikapı and Sanayi on the M2 line, Olimpiyat on the M9 line, and Bostancı on the M4 line. Design The stations on the M1 line, which were built in the 1980s and 1990s, are very functional and plain in design. Like many stations built during this period, the stations use fluorescent tubes as light sources, and are decorated with tiled floors and ceiling panels made of white plastic ceiling slats. Platforms of stations on the M1 line are supported by square and tiled central columns (e.g. Aksaray, Emniyet - Fatih and Topkapı - Ulubatlı.) Stations on the M2 line built between 2000 and 2011, care was taken to use light colors. For this reason, the basic wall color of the stations is white. In order to make people not confuse the stations and to make them more memorable, some characteristic patterns were also implemented. (e.g. colored stripes/tiles on walls). Stations on the M7 line, which were built between 2017 and 2020, mostly use a combination of gray and white colors. Since the line has fully automated driverless vehicles, passengers are prevented from falling onto the tracks by using platform screen doors (PSDs). Technical specifications This table lists technical characteristics of the metro lines that are currently in service or under construction. Alignment and interchanges Apart from the Haliç station on the M2 line, about half of the M1 (mostly M1A) line, Alibeyköy and Kağıthane valleys crossing by the M7 line, viaduct section at Menekşe River of the M3 line and viaduct section at Olimpiyat of the M9 line, the lines are fully underground. All station names are on the bus lines as well. The M2 line has an interchange between F1 in Taksim and an interchange tunnel with the Zincirlikuyu Metrobus station at the Gayrettepe station. There is also a transfer station at Yenikapı with Marmaray, M1 line and İDO Ferry Port; from where it is possible to take the high-speed catamaran Seabus departing for Bursa, Bandırma or Yalova; as well as the other Seabus ports of Istanbul such as Bostancı, Kadıköy, Bakırköy and Kabataş. The M3 line has an interchange with the M9 line at the İkitelli Sanayi station, M7 line at the Mahmutbey station, M1 line at the Kirazlı station, M2, M1 and Metrobus at the İncirli station, and lastly Marmaray (B1) and YHT at Özgürlük Meydanı Station. The M4 line has a vapur (traditional ferry), motorboat, İDO and nostalgic tram interchange at Kadıköy which is the heart of Istanbul's Asian side. One can also interchange to Marmaray at the Ayrılık Ceşmesi station. Also at Ünalan / Uzunçayır, the line has a Metrobus interchange just like Gayrettepe in M2. The M5 line has an interchange with the Marmaray and İDO at the Üsküdar station. Also at Altunizade the line has a Metrobus interchange. The M6 line has an interchange with the M2 line at the Levent station. The M7 line interchanges with the M2 line at the Mecidiyeköy station, with the T4 tramway line at the Kiptaş-Venezia/Karadeniz station, and with the M3 at the Mahmutbey station. The M8 line has an interchange with the Marmaray at the Bostancı station, with the M4 line at the Kozyatağı station and with the M5 line at the Dudullu station. The M9 has an interchange with the M3 at Ikitelli station. The M11 line interchanges with the M2 line and Metrobus at Gayrettepe station, and with the M7 line at Kağıthane station. Rolling stock Aside from running on standard gauge tracks and all models having 4 doors per side, the rolling stock is generally incompatible, with few track connections to other lines and rolling stock generally will never run on any line other than whatever they first were delivered to. As each line is generally self-contained, this has rarely become an issue. History The first Istanbul full metro rolling stocks, which entered service on 16 September 2000, on the Taksim - 4. Levent line, were built by Alstom. These trains are air-conditioned and equipped with LCD screens, and share a similar exterior design to the first three generations of rolling stock used on the Caracas Metro in Venezuela. On 30 January 2009, the first 8 trains (each with 4 wagons) built by Eurotem (the Turkish factory of Hyundai Rotem) entered service. Today the system has 268 trains. These trains are also air conditioned and equipped with LCD screens, as well as dynamic route map showing the location and direction of the train. In September 2009, CAF signed a contract to supply 144 units for the M4 metro line, amounting to 1.1 million euros. These metro units are formed with 4 cars for a total length of 90 meters and have a maximum transport capacity of 1300 passengers. In August 2013, tender for the 126 driverless train units for the M5 line was won by CAF and Mitsubishi with 119 million euros. The first units were delivered in November 2016. In March 2016, Eurotem signed a contract to supply 300 driverless units for the M7 line, amounting to 280.200.000 euros. Rolling stocks by line M1A and M1B Although line M1 is a (light) metro line, its rolling stock—in use since 1989—is made up of typical ABB light rail vehicles (LRVs). These are partly the same as those used on the T4 tramway line. The rolling stock of the line is planned to be refurbished and made driverless. M2 and M6 The rolling stock on the M2 and M6 lines, which totals a number of 192 units, is made up of Alstom vehicles from France and Hyundai Rotem vehicles from South Korea which are manufactured in Adapazarı by Eurotem. All wagons have 8 double doors, making them have 4 openings on each side per wagon, the rolling stock also includes a fast passenger change, heating, air conditioning and broadcasting system. These lines are the only lines to use third rail power. M3 and M9 The rolling stock on the M3 line, consists of 68 Alstom Metropolis AM4 (similar-looking vehicles are used at M4 Budapest Metro) units. Some of them are also used on the M9 Ataköy-Olimpiyat metro line. M4 Line M4 uses CAF rolling stock. Trains operate in 4-car sets and can couple to form 8-car trains. M5 Line M5 uses driverless 6-car CAF rolling stock. M7 and M8 Line M7 uses driverless Hyundai Rotem rolling stock. Trains are made of 4-car units and can couple to form 8-car trains, although these are not currently used. Line M8 uses very similar trains, but only as 4-car sets. M11 Line M11 uses CRRC rolling stock. Future extensions As part of the Istanbul Metropolitan Municipality goal of expanding the size of the city rail transportation network to by 2030, the Istanbul Metro has several lines that are under construction or planned. Since the city does not have a widespread railway network, the Metropolitan Municipality aims to connect urban areas of the city that do not have access to the Istanbul two coastal railways with metro lines. Due to the city's unique geography and depth of the Bosporus which divides the city, none of the current Istanbul Metro lines cross the strait; lines are wholly located either on the European side or the Asian side of the city. In 2019, The Ministry of Transport and Infrastructure began the planning of a line that would link İncirli with Söğütlüçeşme via a tunnel under the Bosphorus also designed to accommodate road traffic. The following metro lines are under construction: Construction of the following metro lines are planned or on hold: European side M7 metro line extension The M7 Kabataş-Yıldız metro line extension is under construction. The extension will have 2 new stations and is expected to go into service in 2025. It will interconnect with ferries at Beşiktaş and Kabataş stations, with T1 trams line at Kabataş station and with F1 funicular line at Kabataş station. M11 Metro Line The M11 Gayrettepe-Arnavutköy Hastane metro extension to Halkalı is under construction. The extension from Arnavutköy Hastane to Halkalı will be opened in late 2025. Asian side M4 metro line M4 Kadıköy–İçmeler Metro Line is extending to İçmeler. Planned opening date of Kaynarca Merkez is 2025. Construction between Kaynarca Merkez and İçmeler has not started yet. Kaynarca Merkez (M10 Line Interchange) Çamçeşme Kavakpınar Esenyalı Aydıntepe İçmeler (Marmaray Interchange) M5 metro line M5 Üsküdar–Çekmeköy-Sultanbeyli Metro Line was being extended a tender signed in April 2017 but construction was restarted in November 2019. The section between Çekmeköy and Samandıra Merkez opened on 16 March 2024 and the remaining section to Sultanbeyli is expected to open in 2025. Veysel Karani Hasanpaşa Sultanbeyli M12 metro line The M12 60. Yıl Parkı-Kazım Karabekir Metro Line will connect the underserved Ataşehir district with the regional centres of densely populated Ümraniye and integrate those districts into the Istanbul rapid rail system. Planned opening date is 2025. 60. Yıl Parkı Tütüncü Mehmet Efendi (Marmaray Interchange) Sahrayıcedit Yenisahra (M4 Line Interchange) Ataşehir Finans Merkezi Site Atakent Çarşı (M5 Line Interchange) SBÜ Hastanesi Kazım Karabekir M14 metro line The M14 Altunizade-Bosna Bulvarı Metro Line will serve the Çamlıca Hill, Çamlıca Mosque and Bosna Boulevard and will be connect the M5 metro line. Planned opening date is 2025. Altunizade (M5 Interchange) Ferah Mahallesi Çamlıca Camii Bosna Bulvarı Network overview Metro, suburban rail, tram, funicular, ropeway and Metrobus (as of March 2024): In operation: 446.15 km / 351 stations Under construction: 84.5 km / 53 stations Planned: ? km / ? stations Sum: ? km / ? stations Network map
Technology
Europe_2
null
1929534
https://en.wikipedia.org/wiki/Solar%20neutrino
Solar neutrino
A solar neutrino is a neutrino originating from nuclear fusion in the Sun's core, and is the most common type of neutrino passing through any source observed on Earth at any particular moment. Neutrinos are elementary particles with extremely small rest mass and a neutral electric charge. They only interact with matter via weak interaction and gravity, making their detection very difficult. This has led to the now-resolved solar neutrino problem. Much is now known about solar neutrinos, but research in this field is ongoing. History and background Homestake experiment The timeline of solar neutrinos and their discovery dates back to the 1960s, beginning with the two astrophysicists John N. Bahcall and Raymond Davis Jr. The experiment, known as the Homestake experiment, named after the town in which it was conducted (Homestake, South Dakota), aimed to count the solar neutrinos arriving at Earth. Bahcall, using a solar model he developed, came to the conclusion that the most effective way to study solar neutrinos would be via the chlorine-argon reaction. Using his model, Bahcall was able to calculate the number of neutrinos expected to arrive at Earth from the Sun. Once the theoretical value was determined, the astrophysicists began pursuing experimental confirmation. Davis developed the idea of taking hundreds of thousands of liters of perchloroethylene, a chemical compound made up of carbon and chlorine, and searching for neutrinos using a chlorine-argon detector. The process was conducted very far underground, hence the decision to conduct the experiment in Homestake as the town was home to the Homestake Gold Mine. By conducting the experiment deep underground, Bahcall and Davis were able to avoid cosmic ray interactions which could affect the process and results. The entire experiment lasted several years as it was able to detect only a few chlorine to argon conversions each day, and the first results were not yielded by the team until 1968. To their surprise, the experimental value of the solar neutrinos present was less than 20% of the theoretical value Bahcall calculated. At the time, it was unknown if there was an error with the experiment or with the calculations, or if Bahcall and Davis did not account for all variables, but this discrepancy gave birth to what became known as the solar neutrino problem. Further experimentation Davis and Bahcall continued their work to understand where they may have gone wrong or what they were missing, along with other astrophysicists who also did their own research on the subject. Many reviewed and redid Bahcall's calculations in the 1970s and 1980s, and although there was more data making the results more precise, the difference still remained. Davis even repeated his experiment changing the sensitivity and other factors to make sure nothing was overlooked, but he found nothing and the results still showed "missing" neutrinos. By the end of the 1970s, the widely expected result was the experimental data yielded about 39% of the calculated number of neutrinos. In 1969, Bruno Pontecorvo, an Italo-Russian astrophysicist, suggested a new idea that maybe we do not quite understand neutrinos like we think we do, and that neutrinos could change in some way, meaning the neutrinos that are released by the sun changed form and were no longer neutrinos the way neutrinos were thought of by the time they reached Earth where the experiment was conducted. This theory Pontecorvo had would make sense in accounting for the discrepancy between the experimental and theoretical results that persisted. Solution to solar neutrino problem Pontecorvo was never able to prove his theory, but he was on to something with his thinking. In 2002, results from an experiment conducted 2100 meters underground at the Sudbury Neutrino Observatory proved and supported Pontecorvo's theory and discovered that neutrinos released from the Sun can in fact change form or flavor because they are not completely massless. This discovery of neutrino oscillation solved the solar neutrino problem, nearly 40 years after Davis and Bahcall began studying solar neutrinos. Neutrino observatories Super-Kamiokande The Super-Kamiokande is a 50,000 ton water Cherenkov detector underground. The primary uses for this detector in Japan in addition to neutrino observation is cosmic ray observation as well as searching for proton decay. In 1998, the Super-Kamiokande was the site of the Super-Kamiokande experiment which led to the discovery of neutrino oscillation, the process by neutrinos change their flavor, either to electron, muon or tau. The Super-Kamiokande experiment began in 1996 and is still active. In the experiment, the detector works by being able to spot neutrinos by analyzing water molecules and detecting electrons being removed from them which then produces a blue Cherenkov light, which is produced by neutrinos. Therefore, when this detection of blue light happens it can be inferred that a neutrino is present and counted. The Sudbury Neutrino Observatory The Sudbury Neutrino Observatory (SNO), a underground observatory in Sudbury, Canada, is the other site where neutrino oscillation research was taking place in the late 1990s and early 2000s. The results from experiments at this observatory along with those at Super-Kamiokande are what helped solve the solar neutrino problem. The SNO is also a heavy-water Cherenkov detector and designed to work the same way as the Super-Kamiokande. The Neutrinos when reacted with heavy water produce the blue Cherenkov light, signaling the detection of neutrinos to researchers and observers. Borexino The Borexino detector is located at the Laboratori Nazionali de Gran Sasso, Italy. Borexino is an actively used detector, and experiments are on-going at the site. The goal of the Borexino experiment is measuring low energy, typically below 1 MeV, solar neutrinos in real-time. The detector is a complex structure consisting of photomultipliers, electrons, and calibration systems making it equipped to take proper measurements of the low energy solar neutrinos. Photomultipliers are used as the detection device in this system as they are able to detect light for extremely weak signals. Solar neutrinos are able to provide direct insight into the core of the Sun because that is where the solar neutrinos originate. Solar neutrinos leaving the Sun's core reach Earth before light does due to the fact solar neutrinos do not interact with any other particle or subatomic particle during their path, while light (photons) bounces around from particle to particle. The Borexino experiment used this phenomenon to discover that the Sun releases the same amount of energy currently as it did a 100,000 years ago. Formation process Solar neutrinos are produced in the core of the Sun through various nuclear fusion reactions, each of which occurs at a particular rate and leads to its own spectrum of neutrino energies. Details of the more prominent of these reactions are described below. The main contribution comes from the proton–proton chain. The reaction is: or in words: two protons deuteron + positron + electron neutrino. Of all Solar neutrinos, approximately 91% are produced from this reaction. As shown in the figure titled "Solar neutrinos (proton–proton chain) in the standard solar model", the deuteron will fuse with another proton to create a 3He nucleus and a gamma ray. This reaction can be seen as: The isotope 4He can be produced by using the 3He in the previous reaction which is seen below. With both helium-3 and helium-4 now in the environment, one of each weight of helium nucleus can fuse to produce beryllium: Beryllium-7 can follow two different paths from this stage: It could capture an electron and produce the more stable lithium-7 nucleus and an electron neutrino, or alternatively, it could capture one of the abundant protons, which would create boron-8. The first reaction via lithium-7 is: This lithium-yielding reaction produces approximately 7% of the solar neutrinos. The resulting lithium-7 later combines with a proton to produce two nuclei of helium-4. The alternative reaction is proton capture, that produces boron-8, which then beta+ decays into beryllium-8 as shown below: This alternative boron-yielding reaction produces about 0.02% of the solar neutrinos; although so few that they would conventionally be neglected, these rare solar neutrinos stand out because of their higher average energies. The asterisk (*) on the beryllium-8 nucleus indicates that it is in an excited, unstable state. The excited beryllium-8 nucleus then splits into two helium-4 nuclei: Observed data The highest flux of solar neutrinos come directly from the proton–proton interaction, and have a low energy, up to 400 keV. There are also several other significant production mechanisms, with energies up to 18 MeV. From the Earth, the amount of neutrino flux at Earth is around 7·1010 particles·cm−2·s −1. The number of neutrinos can be predicted with great confidence by the standard solar model, but the number of neutrinos detected on Earth versus the number of neutrinos predicted are different by a factor of a third, which is the solar neutrino problem. Solar models additionally predict the location within the Sun's core where solar neutrinos should originate, depending on the nuclear fusion reaction which leads to their production. Future neutrino detectors will be able to detect the incoming direction of these neutrinos with enough precision to measure this effect. The energy spectrum of solar neutrinos is also predicted by solar models. It is essential to know this energy spectrum because different neutrino detection experiments are sensitive to different neutrino energy ranges. The Homestake experiment used chlorine and was most sensitive to solar neutrinos produced by the decay of the beryllium isotope 7Be. The Sudbury Neutrino Observatory is most sensitive to solar neutrinos produced by 8B. The detectors that use gallium are most sensitive to the solar neutrinos produced by the proton–proton chain reaction process, however they were not able to observe this contribution separately. The observation of the neutrinos from the basic reaction of this chain, proton–proton fusion in deuterium, was achieved for the first time by Borexino in 2014. In 2012 the same collaboration reported detecting low-energy neutrinos for the proton–electron–proton (pep reaction) that produces 1 in 400 deuterium nuclei in the Sun. The detector contained 100 metric tons of liquid and saw on average 3 events each day (due to C production) from this relatively uncommon thermonuclear reaction. In 2014, Borexino reported a successful direct detection of neutrinos from the pp-reaction at a rate of 144±33/day, consistent with the predicted rate of 131±2/day that was expected based on the standard solar model prediction that the pp-reaction generates 99% of the Sun's luminosity and their analysis of the detector's efficiency. And in 2020, Borexino reported the first detection of CNO cycle neutrinos from deep within the solar core. Note that Borexino measured neutrinos of several energies; in this manner they have demonstrated experimentally, for the first time, the pattern of solar neutrino oscillations predicted by the theory. Neutrinos can trigger nuclear reactions. By looking at ancient ores of various ages that have been exposed to solar neutrinos over geologic time, it may be possible to interrogate the luminosity of the Sun over time, which, according to the standard solar model, has changed over the eons as the (presently) inert byproduct helium has accumulated in its core. Key contributing astrophysicists Wolfgang Pauli was the first to suggest the idea of a particle such as the neutrino existing in our universe in 1930. He believed such a particle to be completely massless. This was the belief amongst the astrophysics community until the solar neutrino problem was solved. Frederick Reines, from the University of California at Irvine, and Clyde Cowan were the first astrophysicists to detect neutrinos in 1956. They won a Nobel Prize in Physics for their work in 1995. Raymond Davis and John Bahcall are the pioneers of solar neutrino studies. While Bahcall never won a Nobel Prize, Davis along with Masatoshi Koshiba won the Nobel Prize in Physics in 2002 after the solar neutrino problem was solved for their contributions in helping solve the problem. Pontecorvo, known as the first astrophysicist to suggest the idea neutrinos have some mass and can oscillate, never received a Nobel Prize for his contributions due to his passing in 1993. Arthur B. McDonald, a Canadian physicist, was a key contributor in building the Sudbury Neutrino Observatory (SNO) in the mid 1980s and later became the director of the SNO and leader of the team that solved the solar neutrino problem. McDonald, along with Japanese physicist Kajita Takaaki both received a Nobel Prize for their work discovering the oscillation of neutrinos in 2015. Current research and findings The critical issue of the solar neutrino problem, that many astrophysicists interested in solar neutrinos studied and attempted to solve in late 1900s and early 2000s, is solved. In the 21st century, even without a main problem to solve, there is still unique and novel research ongoing in this field of astrophysics. Solar neutrino flux at keV energies This research, published in 2017, aimed to solve the solar neutrino and antineutrino flux for extremely low energies (keV range). Processes at these low energies consisted vital information that told researchers about the solar metallicity. Solar metallicity is the measure of elements present in the particle that are heavier than hydrogen and helium, typically in this field this element is usually iron. The results from this research yielded significantly different findings compared to past research in terms of the overall flux spectrum. Currently technology does not yet exist to put these findings to the test. Limiting neutrino magnetic moments with Borexino Phase-II solar neutrino data This research, published in 2017, aimed to search for the solar neutrino effective magnetic moment. The search was completed using data from exposure from the Borexino experiment's second phase which consisted of data over 1291.5 days (3.54 years). The results yielded that the electron recoil spectrum shape was as expected with no major changes or deviations from it.
Physical sciences
Solar System
Astronomy
1930277
https://en.wikipedia.org/wiki/Empusidae
Empusidae
Empusidae is a family of plant-mimicking mantises (see Mantodea), consisting of 10 genera, in two subfamilies. Unlike many other mantis families, the Empusidae are a monophyletic lineage. Empusidae mantises are ambush predators, with mouthparts adapted to feeding on other insects and small animals. The majority of Empusidae species are distributed throughout Africa, but they are also found in Southeast Asia and in the southern parts of Europe. Natural history The Empusidae species Gongylus gongylodes (Linné, 1758) was the first mantis species ever to be described. Since Gongylus mantises have been fascinating entomologists and have been bred in captivity for a long time, their behavior and breeding preferences are well known, such as a defensive behavior of displaying a hissing noise by rubbing the anterior edges of its serrated fore wings to the femur of the hind legs. Morphology The about 28 species of empusid mantis are all relatively large and bizarre looking. The prothorax is always surrounded by a crest and the femur of the middle and hind legs often have flap-like appendages. The pronotum is characteristically elongated and the abdomen is often lobed. Members of the Mantoidea superfamily possess a cyclopean ear, an organ situated on the metathorax, which has been proven to be an adaptation to bat predation. The presence of this adaptation has been dated to originate in the early Eocene. Phylogeny The Empusidae belong to the superfamily Hymenopoidea, together with the Hymenopodidae. Phylogenetic studies place the Empusidae as a sister group to the Hymenopodidae. The Empusidae and Hymenopodidae are, in turn, placed as sister groups to all other ambush mantises. The latest phylogeny was revised by Svenson et al. 2015. The Mantodea Species File currently includes two subfamilies: Blepharodinae Blepharodes Bolivar, 1890 Blepharopsis Rehn, 1902 Empusinae This subfamily is divided into two tribes: Empusini subtribe Empusina Dilatempusa Roy, 2004 Empusa Illiger, 1798 Gongylus Thunberg, 1815 Hypsicorypha Krauss, 1892 subtribe Idolomorphina Chopardempusa Paulian, 1958 Hemiempusa Saussure & Zehntner, 1895 Idolomorpha Burmeister, 1838 Idolomorphini Idolomantis Uvarov, 1940 Historical findings In 2017-2018, a rock carving of an Empusidae with raptorial forearms was revealed in the Teimareh rock art site in the Khomeyn County, Iran. An engraved, insect-like image has a 14-cm length and 11-cm width with two circles at its sides which probably dates 40,000–4,000 years back. This motif is analogous to the famous 'squatter man' petroglyph encountered at several locations around the world. Images
Biology and health sciences
Insects: General
Animals
1930814
https://en.wikipedia.org/wiki/Hydrogen%20iodide
Hydrogen iodide
Hydrogen iodide (HI) is a diatomic molecule and hydrogen halide. Aqueous solutions of HI are known as hydroiodic acid or hydriodic acid, a strong acid. Hydrogen iodide and hydroiodic acid are, however, different in that the former is a gas under standard conditions, whereas the other is an aqueous solution of the gas. They are interconvertible. HI is used in organic and inorganic synthesis as one of the primary sources of iodine and as a reducing agent. Properties of hydrogen iodide HI is a colorless gas that reacts with oxygen to give water and iodine. With moist air, HI gives a mist (or fumes) of hydroiodic acid. It is exceptionally soluble in water, giving hydroiodic acid. One liter of water will dissolve 425 liters of HI gas, the most concentrated solution having only four water molecules per molecule of HI. Hydroiodic acid Hydroiodic acid is not pure hydrogen iodide, but a mixture containing it. Commercial "concentrated" hydroiodic acid usually contains 48–57% HI by mass. The solution forms an azeotrope boiling at 127 °C with 57% HI, 43% water. The high acidity is caused by the dispersal of the ionic charge over the anion. The iodide ion radius is much larger than the other common halides, which results in the negative charge being dispersed over a large space. By contrast, a chloride ion is much smaller, meaning its negative charge is more concentrated, leading to a stronger interaction between the proton and the chloride ion. This weaker H+···I− interaction in HI facilitates dissociation of the proton from the anion and is the reason HI is the strongest acid of the hydrohalides.  Ka ≈ 1010  Ka ≈ 109  Ka ≈ 106 Synthesis The industrial preparation of HI involves the reaction of I2 with hydrazine, which also yields nitrogen gas: When performed in water, the HI must be distilled. HI can also be distilled from a solution of NaI or other alkali iodide in concentrated phosphoric acid (note that concentrated sulfuric acid will not work for acidifying iodides, as it will oxidize the iodide to elemental iodine). Another way HI may be prepared is by bubbling hydrogen sulfide steam through an aqueous solution of iodine, forming hydroiodic acid (which is distilled) and elemental sulfur (this is filtered): Additionally, HI can be prepared by simply combining H2 and I2: This is a reversible reaction (using conditions 250°C) This method is usually employed to generate high-purity samples. For many years, this reaction was considered to involve a simple bimolecular reaction between molecules of H2 and I2. However, when a mixture of the gases is irradiated with the wavelength of light equal to the dissociation energy of I2, about 578 nm, the rate increases significantly. This supports a mechanism whereby I2 first dissociates into 2 iodine atoms, which each attach themselves to a side of an H2 molecule and break the : In the laboratory, another method involves hydrolysis of PI3, the iodine analog of PBr3. In this method, I2 reacts with phosphorus to create phosphorus triiodide, which then reacts with water to form HI and phosphorous acid: Key reactions and applications Solutions of hydrogen iodide are easily oxidized by air: is dark brown in color, which makes aged solutions of HI often appear dark brown. Like HBr and HCl, HI adds to alkenes: HI is also used in organic chemistry to convert primary alcohols into alkyl iodides. This reaction is an SN2 substitution, in which the iodide ion replaces the "activated" hydroxyl group (water): HI is preferred over other hydrogen halides because the iodide ion is a much better nucleophile than bromide or chloride, so the reaction can take place at a reasonable rate without much heating. This reaction also occurs for secondary and tertiary alcohols, but substitution occurs via the SN1 pathway. HI (or HBr) can also be used to cleave ethers into alkyl iodides and alcohols, in a reaction similar to the substitution of alcohols. This type of cleavage is significant because it can be used to convert a chemically stable and inert ether into more reactive species. In this example diethyl ether is split into ethanol and iodoethane: The reaction is regioselective, as iodide tends to attack the less sterically hindered ether carbon. If an excess of HI is used, the alcohol formed in this reaction will be converted to a 2nd equivalent of alkyl iodide, as in the conversion of primary alcohols into alkyl iodides. HI is subject to the same Markovnikov and anti-Markovnikov guidelines as HCl and HBr. Although harsh by modern standards, HI was commonly employed as a reducing agent early on in the history of organic chemistry. Chemists in the 19th century attempted to prepare cyclohexane by HI reduction of benzene at high temperatures, but instead isolated the rearranged product, methylcyclopentane (see the article on cyclohexane). As first reported by Kiliani, hydroiodic acid reduction of sugars and other polyols results in the reductive cleavage of several or even all hydroxy groups, although often with poor yield and/or reproducibility. In the case of benzyl alcohols and alcohols with α-carbonyl groups, reduction by HI can provide synthetically useful yields of the corresponding hydrocarbon product (). This process can be made catalytic in HI using red phosphorus to reduce the formed I2.
Physical sciences
Hydrogen compounds
Chemistry
1932067
https://en.wikipedia.org/wiki/Greenland%20shark
Greenland shark
The Greenland shark (Somniosus microcephalus), also known as the gurry shark or grey shark, is a large shark of the family Somniosidae ("sleeper sharks"), closely related to the Pacific and southern sleeper sharks. Inhabiting the North Atlantic and Arctic Oceans, they are notable for their exceptional longevity, although they are poorly studied due to the depth and remoteness of their natural habitat. Greenland sharks have the longest lifespan of any known vertebrate, estimated to be between 250 and 500 years. They are among the largest extant species of shark, reaching a maximum confirmed length of long and weighing over . They reach sexual maturity at about 150 years of age, and their pups are born alive after an estimated gestation period of 8 to 18 years. The shark is a generalist feeder, consuming a variety of available foods, including carrion. Greenland shark meat is toxic to mammals due to its high levels of trimethylamine N-oxide, although a treated form of it is eaten in Iceland as a delicacy known as kæstur hákarl. Because they live deep in remote parts of the northern oceans, Greenland sharks are not considered a threat to humans. A possible attack occurred in August 1936 on two British fisherman, but the species was never identified. Description The Greenland shark is one of the largest known extant species of shark, with adults growing to around 400 to 500 cm. The largest confirmed specimen measured up to long and weighed around . The all-tackle International Game Fish Association (IGFA) record for this species is . It rivals the Pacific sleeper shark (possibly up to long) for the largest species in the family Somniosidae. The Greenland shark is a thickset species, with a short, rounded snout, small eyes, and small dorsal and pectoral fins. The gill openings are very small for the species' great size. Female Greenland sharks are typically larger than males. Coloration can range from pale creamy-gray to blackish-brown and the body is typically uniform in color, though whitish spots or faint dark streaks are occasionally seen on the back. The shark is often infested by the copepod Ommatokoita elongata, a crustacean which attaches itself to the shark's eyes. It was speculated that the copepod may display bioluminescence and thus attract prey for the shark in a mutualistic relationship, but this hypothesis has not been verified. These parasites also damage the eyeball in several ways, leading to almost complete blindness. This does not seem to reduce the life expectancy or predatory ability of Greenland sharks, due to their strong reliance on smell and hearing. The genome of the Greenland shark was published in 2024. It is 6.45 Gb (billion base pairs) in length. Dentition When feeding on large carcasses, the shark employs a rolling motion of its jaw. The 48 to 52 teeth of the upper jaw are very thin and pointed, lacking serrations. These upper jaw teeth act as an anchor while the lower jaw proceeds to cut massive chunks out of the prey. The 48 to 52 lower teeth are interlocking, broad and square in shape, containing short, smooth cusps that point outward. Teeth in the two halves of the lower jaw are strongly pitched in opposite directions. Behavior Diet As both scavengers and active predators, Greenland sharks have established themselves as apex predators in Arctic ecosystems. They primarily eat fish (cod, wolffish, haddock, and skates) and seal. Some Greenland sharks have been found to also eat minke whale. Small Greenland sharks eat predominantly squid, as well as sea birds, crabs, amphipods, marine snails, brittle stars, sea urchins, and jellyfish, while the larger sharks that are greater than were discovered eating prey such as epibenthic and benthic fishes, as well as seals and small cetaceans such as oceanic dolphins and porpoises. The largest of these sharks were found having eaten redfish, as well as other higher trophic level prey. It is proposed that, because of their slow speeds and low twitch speed muscle fiber, Greenland sharks hunt marine mammals such as seals and smaller cetaceans that are asleep, injured, or sick. Regarding most other benthic species, they utilize their cryptic coloration, and approach prey undetected before closing the remaining distance, expanding their buccal cavity to create suction, drawing in prey. This is the likely explanation for why the gut contents of Greenland sharks are often whole prey specimens. Greenland sharks have also been found with remains of moose, polar bear, horse, and reindeer (in one case an entire reindeer body) in their stomachs. The Greenland shark is known to be a scavenger and is attracted by the smell of rotting meat in the water. The sharks have frequently been observed gathering around fishing boats. They also scavenge on seals. Although such a large shark could easily consume a human swimmer, the frigid waters it typically inhabits make the likelihood of attacks on people very low. No cases of predation on humans have been verified. Movement and migration The Greenland shark prefers cold water temperatures (—1.1 to 7.4 °C) and deep water (100 to 1,200m). As an ectotherm living in a just-above-freezing environment, this species is sluggish and slow-moving, with the lowest swim speed and tail-beat frequency for its size across all fish species, which most likely correlates with its very slow metabolism and extreme longevity. It swims at an average of 0.34 m·s−1, with its fastest cruising speed only reaching 0.74 m·s−1. Because this top speed is a fraction of that of a typical seal in their diet, biologists are uncertain how the sharks are able to prey on the seals. It is hypothesized that they may ambush them while they sleep. Greenland sharks migrate annually based on depth and temperature rather than distance, although some do travel. During the winter, the sharks congregate in the shallows (up to 80° north) for warmth but migrate separately in summer to the deeps or even farther south. The species has been observed at a depth of by a submersible investigating the wreck of the SS Central America that lies about east of Cape Hatteras, North Carolina. Daily vertical migration between shallower and deeper waters has also been recorded. In August 2013, researchers from Florida State University caught a Greenland shark in the Gulf of Mexico at a depth of , where the water temperature was . Four previous records of Greenland shark were reported from Cuba and the northern Gulf of Mexico. A more typical depth range is , with the species often occurring in relatively shallow waters in the far north and deeper in the southern part of its range. In April 2022, a large Somniosus shark was caught and subsequently released on Glover's Reef off the coast of Belize. This shark was identified as being either a Greenland shark or a Greenland/Pacific sleeper shark hybrid. This observation is notable for being the first possible record of a Greenland shark from the Western Caribbean, and being caught on a nearshore coral reef (the only other record of this species from the Caribbean was made from a deep-water habitat off the Caribbean coast of Colombia). The discovery indicates that Greenland sharks may have a wider distribution in the tropics, primarily at greater depths, than previously believed. Longevity The Greenland shark has the longest known lifespan of all vertebrate species. It is estimated that the species has a lifespan of at least 272 years, with the oldest individual estimated to be 392 ± 120 years of age. Estimates of age were made using radiocarbon dating of crystals within the lenses of their eyes. Greenland sharks are estimated to reach sexual maturity at around 150 years of age at which point females measure around 4.19 ± 0.04 meters and males measure around 2.84 ± 0.06 meters. One Greenland shark was tagged off the coast of Greenland in 1936 and recaptured in 1952. Its measurements suggest that Greenland sharks grow at a rate of per year. Efforts to conserve Greenland sharks are particularly important due to their extreme longevity, long maturation periods, and the heightened sensitivity of large shark populations. Reproduction Greenland sharks are born alive (a process known as ovoviviparity) after an estimated gestation period of 8–18 years. Estimates of litter size have varied across studies. Some studies suggest that this species produce up to 10 pups per litter, each initially measuring some 38–42 cm in length. Based on these estimates, It is thought that, due to their extreme longevity, Greenland sharks can have between 200 and 700 pups during their lifetime. Within a Greenland shark's uterus, villi serve a key function in supplying oxygen to embryos. It is speculated that oxygen supply is a major limiting factor in the size of litters. Other studies, however, have estimated that Greenland sharks may produce from 200 to 324 pups per litter, measuring between 35–45 cm in length. Physiological adaptations Like other elasmobranchii, Greenland sharks have high concentrations of the two nitrogenous compounds urea and trimethylamine N-oxide (TMAO) in their tissues, which increase their buoyancy and function as osmoprotectants. TMAO also counteracts the protein-destabilizing tendencies of urea and deep-water pressure. Its presence in the tissues of both elasmobranch and teleost fish has been found to increase with depth. The blood of Greenland sharks contains three major types of hemoglobin, made up of two copies of  globin combined with two copies of three very similar  subunits. These three types show very similar oxygenation and carbonylation properties, which are unaffected by urea, an important compound in marine elasmobranchii physiology. They display identical electronic absorption and resonance in Raman spectroscopy, indicating that their heme-pocket structures are identical or highly similar. The hemoglobins also have a lower affinity for oxygen compared to temperate sharks. These characteristics are interpreted as adaptations to living at great water depths. Threats The shark has historically been hunted for its liver oil up until the development of synthetic oils and cessation of export of liver oil and skin from Greenland in the 1960s. In the 1970s, the species was seen as a problem for other fisheries in western Norway and the government subsidized a fishery to reduce the stock of the species. Today, the Greenland shark is primarily caught as bycatch in industrial fisheries. While about 25 Greenland sharks are caught per year by artisanal fisheries targeting the species in Iceland, 3,500 are caught annually as bycatch in the Arctic and Atlantic Oceans. The shark is likely affected by quantity, dynamics, and distribution of Arctic sea ice. The rate of projected loss of sea ice will continue to negatively influence the abundance, distribution and availability of prey, while, at the same time, providing greater access for fishing fleets. There is greater potential for new fisheries to develop as more productive and abundant southerly species invade the warming Arctic waters. Conservation and management Greenland sharks are recognized as the longest-lived vertebrates on earth. They have a slow growth rate, late maturity period, and low fecundity, making the management and conservation of this species very important. As a result of their low productivity and extreme longevity, this species is particularly susceptible to overfishing and bycatch. Therefore, Greenland sharks' longevity and conservative life history traits, in tandem with their vulnerability to accidental catching and commercial fishing, promotes a growing concern for the sustainability of this species. Hákarl The flesh of the Greenland shark is toxic because of the presence of high concentrations of urea and trimethylamine oxide (TMAO). If the meat is eaten without pretreatment, the ingested TMAO is metabolized into trimethylamine, which may be a uremic toxin. Occasionally, sled dogs that eat the flesh are unable to stand up because of this effect. Similar toxic effects occur with the related Pacific sleeper shark, but not in most other shark species. Greenland shark meat is produced and eaten in Iceland where, today, it is known as a delicacy called hákarl. To make the shark safe for human consumption, it is first fermented and then dried in a process that can take multiple months. The shark was traditionally fermented by burying the meat in gravel pits near the ocean for at least several weeks. In the present day, shark cuts are typically fermented in containers that are perforated to allow liquid to drain. The fermentation process converts urea into ammonia and TMAO into TMA, which then drains as liquid from the meat. The meat is then excavated and hung in strips to dry for several more months. Inuit legends The Greenland shark's poisonous flesh has a high urea content, which gave rise to the Inuit legend of Skalugsuak, the first Greenland shark. The legend says that an old woman washed her hair in urine (a common practice to kill head lice) and dried it with a cloth. The cloth blew into the ocean to become Skalugsuak. Another legend tells of Sedna, whose father cut off her fingers while drowning her, with each finger turning into a sea creature, including Skalugsuak. The Greenland shark plays a role in cosmologies of the Inuit from the Canadian Eastern Arctic and Greenland. Igloolik Inuit believe that the shark lives within the urine pot of Sedna, goddess of the sea, and consequently, its flesh has a urine-like smell and acts as a helping spirit to shamans.
Biology and health sciences
Sharks
Animals
1932886
https://en.wikipedia.org/wiki/Salp
Salp
A salp (: salps, also known colloquially as “sea grape”) or salpa (: salpae or salpas) is a barrel-shaped, planktonic tunicate in the family Salpidae. It moves by contracting, thereby pumping water through its gelatinous body; it is one of the most efficient examples of jet propulsion in the animal kingdom. The salp strains the pumped water through its internal feeding filters, feeding on phytoplankton. Distribution Salps are common in equatorial, temperate, and cold seas, where they can be seen at the surface, singly or in long, stringy colonies. The most abundant concentrations of salps are in the Southern Ocean (near Antarctica), where they sometimes form enormous swarms, often in deep water, and are sometimes even more abundant than krill. Since 1910, while krill populations in the Southern Ocean have declined, salp populations appear to be increasing. Salps have been seen in increasing numbers along the coast of Washington, United States. Life cycle Salps have a complex life cycle, with an obligatory alternation of generations. Both portions of the life cycle exist together in the seas—they look quite different, but both are mostly transparent, tubular, gelatinous animals that are typically between long. The solitary life history phase, also known as an oozooid, is a single, barrel-shaped animal that reproduces asexually by producing a chain of tens to hundreds of individuals, which are released from the parent at a small size. The chain of salps is the 'aggregate' portion of the life cycle. The aggregate individuals are also known as blastozooids; they remain attached together while swimming and feeding, and each individual grows in size. Each blastozooid in the chain reproduces sexually (the blastozooids are sequential hermaphrodites, first maturing as females, and are fertilized by male gametes produced by older chains), with a growing embryo oozooid attached to the body wall of the parent. The growing oozooids are eventually released from the parent blastozooids, and then continue to feed and grow as the solitary asexual phase, closing the life cycle of salps. The alternation of generations allows for a fast generation time, with both solitary individuals and aggregate chains living and feeding together in the sea. When phytoplankton is abundant, this rapid reproduction leads to fairly short-lived blooms of salps, which eventually filter out most of the phytoplankton. The bloom ends when enough food is no longer available to sustain the enormous population of salps. Occasionally, mushroom corals and those of the genus Heteropsammia are known to feed on salps during blooms. History The incursion of a large number of salps (Salpa fusiformis) into the North Sea in 1920 led to a failure of the Scottish herring fishery. Oceanographic importance A reason for the success of salps is how they respond to phytoplankton blooms. When food is plentiful, salps can quickly bud off clones, which graze on the phytoplankton and can grow at a rate which is probably faster than that of any other multicellular animal, quickly stripping the phytoplankton from the sea. But if the phytoplankton is too dense, the salps can clog and sink to the bottom. During these blooms, beaches can become slimy with mats of salp bodies, and other planktonic species can experience fluctuations in their numbers due to competition with the salps. Sinking fecal pellets and bodies of salps carry carbon to the seafloor, and salps are abundant enough to have an effect on the ocean's biological pump. Consequently, large changes in their abundance or distribution may alter the ocean's carbon cycle, and potentially play a role in climate change. Nervous systems and relationships to other animals Salps are closely related to the pelagic tunicate groups Doliolida and Pyrosoma, as well as to other bottom-living (benthic) tunicates. Although salps appear similar to jellyfish because of their simple body form and planktonic behavior, they are chordates: animals with dorsal nerve cords, related to vertebrates (animals with backbones). Small fish swim inside salps as protection from predators. Classification The World Register of Marine Species lists the following genera and species in the order Salpida: Order Salpida Family Salpidae Subfamily Cyclosalpinae Genus Cyclosalpa de Blainville, 1827 Cyclosalpa affinis (Chamisso, 1819) Cyclosalpa bakeri Ritter, 1905 Cyclosalpa foxtoni Van Soest, 1974 Cyclosalpa ihlei van Soest, 1974 Cyclosalpa pinnata (Forskål, 1775) Cyclosalpa polae Sigl, 1912 Cyclosalpa quadriluminis Berner, 1955 Cyclosalpa sewelli Metcalf, 1927 Cyclosalpa strongylenteron Berner, 1955 Genus Helicosalpa Todaro, 1902 Helicosalpa komaii (Ihle & Ihle-Landenberg, 1936) Helicosalpa virgula (Vogt, 1854) Helicosalpa younti Kashkina, 1973 Subfamily Salpinae Genus Brooksia Metcalf, 1918 Brooksia berneri van Soest, 1975 Brooksia rostrata (Traustedt, 1893) Genus Ihlea Metcalf, 1919 Ihlea magalhanica (Apstein, 1894) Ihlea punctata (Forskål, 1775) Ihlea racovitzai (van Beneden & Selys Longchamp, 1913) Genus Metcalfina Metcalfina hexagona (Quoy & Gaimard, 1824) Genus Pegea Savigny, 1816 Pegea bicaudata (Quoy & Gaimard, 1826) Pegea confederata (Forsskål, 1775) Genus Ritteriella Metcalf, 1919 Ritteriella amboinensis (Apstein, 1904) Ritteriella picteti (Apstein, 1904) Ritteriella retracta (Ritter, 1906) Genus Salpa Forskål, 1775 Salpa aspera Chamisso, 1819 Salpa fusiformis Cuvier, 1804 Salpa gerlachei Foxton, 1961 Salpa maxima Forskål, 1775 Salpa thompsoni (Foxton, 1961) Salpa tuberculata Metcalf, 1918 Salpa younti van Soest, 1973 Genus Soestia (also accepted as Iasis) Soestia cylindrica (Cuvier, 1804) Soestia zonaria (Pallas, 1774) Genus Thalia Thalia cicar van Soest, 1973 Thalia democratica Forskål, 1775 Thalia longicauda Quoy & Gaimard, 1824 Thalia orientalis Tokioka, 1937 Thalia rhinoceros Van Soest, 1975 Thalia rhomboides Quoy & Gaimard, 1824 Thalia sibogae Van Soest, 1973 Genus Thetys Tilesius, 1802 Thetys vagina Tilesius, 1802 Genus Traustedtia Traustedtia multitentaculata Quoy & Gaimard, 1834 Genus Weelia Yount, 1954 Weelia cylindrica (Cuvier, 1804)
Biology and health sciences
Chordata (except vertebrates)
Animals
24027000
https://en.wikipedia.org/wiki/Properties%20of%20water
Properties of water
Water () is a polar inorganic compound that is at room temperature a tasteless and odorless liquid, which is nearly colorless apart from an inherent hint of blue. It is by far the most studied chemical compound and is described as the "universal solvent" and the "solvent of life". It is the most abundant substance on the surface of Earth and the only common substance to exist as a solid, liquid, and gas on Earth's surface. It is also the third most abundant molecule in the universe (behind molecular hydrogen and carbon monoxide). Water molecules form hydrogen bonds with each other and are strongly polar. This polarity allows it to dissociate ions in salts and bond to other polar substances such as alcohols and acids, thus dissolving them. Its hydrogen bonding causes its many unique properties, such as having a solid form less dense than its liquid form, a relatively high boiling point of 100 °C for its molar mass, and a high heat capacity. Water is amphoteric, meaning that it can exhibit properties of an acid or a base, depending on the pH of the solution that it is in; it readily produces both and ions. Related to its amphoteric character, it undergoes self-ionization. The product of the activities, or approximately, the concentrations of and is a constant, so their respective concentrations are inversely proportional to each other. Physical properties Water is the chemical substance with chemical formula ; one molecule of water has two hydrogen atoms covalently bonded to a single oxygen atom. Water is a tasteless, odorless liquid at ambient temperature and pressure. Liquid water has weak absorption bands at wavelengths of around 750 nm which cause it to appear to have a blue color. This can easily be observed in a water-filled bath or wash-basin whose lining is white. Large ice crystals, as in glaciers, also appear blue. Under standard conditions, water is primarily a liquid, unlike other analogous hydrides of the oxygen family, which are generally gaseous. This unique property of water is due to hydrogen bonding. The molecules of water are constantly moving concerning each other, and the hydrogen bonds are continually breaking and reforming at timescales faster than 200 femtoseconds (2 × 10−13 seconds). However, these bonds are strong enough to create many of the peculiar properties of water, some of which make it integral to life. Water, ice, and vapor Within the Earth's atmosphere and surface, the liquid phase is the most common and is the form that is generally denoted by the word "water". The solid phase of water is known as ice and commonly takes the structure of hard, amalgamated crystals, such as ice cubes, or loosely accumulated granular crystals, like snow. Aside from common hexagonal crystalline ice, other crystalline and amorphous phases of ice are known. The gaseous phase of water is known as water vapor (or steam). Visible steam and clouds are formed from minute droplets of water suspended in the air. Water also forms a supercritical fluid. The critical temperature is 647 K and the critical pressure is 22.064 MPa. In nature, this only rarely occurs in extremely hostile conditions. A likely example of naturally occurring supercritical water is in the hottest parts of deep water hydrothermal vents, in which water is heated to the critical temperature by volcanic plumes and the critical pressure is caused by the weight of the ocean at the extreme depths where the vents are located. This pressure is reached at a depth of about 2200 meters: much less than the mean depth of the ocean (3800 meters). Heat capacity and heats of vaporization and fusion Water has a very high specific heat capacity of 4184 J/(kg·K) at 20 °C (4182 J/(kg·K) at 25 °C) —the second-highest among all the heteroatomic species (after ammonia), as well as a high heat of vaporization (40.65 kJ/mol or 2268 kJ/kg at the normal boiling point), both of which are a result of the extensive hydrogen bonding between its molecules. These unusual properties allow water to moderate Earth's climate by buffering large fluctuations in temperature. Most of the additional energy stored in the climate system since 1970 has accumulated in the oceans. The specific enthalpy of fusion (more commonly known as latent heat) of water is 333.55 kJ/kg at 0 °C: the same amount of energy is required to melt ice as to warm ice from −160 °C up to its melting point or to heat the same amount of water by about 80 °C. Of common substances, only that of ammonia is higher. This property confers resistance to melting on the ice of glaciers and drift ice. Before and since the advent of mechanical refrigeration, ice was and still is in common use for retarding food spoilage. The specific heat capacity of ice at −10 °C is 2030 J/(kg·K) and the heat capacity of steam at 100 °C is 2080 J/(kg·K). Density of water and ice The density of water is about : this relationship was originally used to define the gram. The density varies with temperature, but not linearly: as the temperature increases, the density rises to a peak at and then decreases; the initial increase is unusual because most liquids undergo thermal expansion so that the density only decreases as a function of temperature. The increase observed for water from to and for a few other liquids is described as negative thermal expansion. Regular, hexagonal ice is also less dense than liquid water—upon freezing, the density of water decreases by about 9%. These peculiar effects are due to the highly directional bonding of water molecules via the hydrogen bonds: ice and liquid water at low temperature have comparatively low-density, low-energy open lattice structures. The breaking of hydrogen bonds on melting with increasing temperature in the range 0–4 °C allows for a denser molecular packing in which some of the lattice cavities are filled by water molecules. Above 4 °C, however, thermal expansion becomes the dominant effect, and water near the boiling point (100 °C) is about 4% less dense than water at . Under increasing pressure, ice undergoes a number of transitions to other polymorphs with higher density than liquid water, such as ice II, ice III, high-density amorphous ice (HDA), and very-high-density amorphous ice (VHDA). The unusual density curve and lower density of ice than of water is essential for much of the life on earth—if water were most dense at the freezing point, then in winter the cooling at the surface would lead to convective mixing. Once 0 °C are reached, the water body would freeze from the bottom up, and all life in it would be killed. Furthermore, given that water is a good thermal insulator (due to its heat capacity), some frozen lakes might not completely thaw in summer. As it is, the inversion of the density curve leads to a stable layering for surface temperatures below 4 °C, and with the layer of ice that floats on top insulating the water below, even e.g., Lake Baikal in central Siberia freezes only to about 1 m thickness in winter. In general, for deep enough lakes, the temperature at the bottom stays constant at about 4 °C (39 °F) throughout the year (see diagram). Density of saltwater and ice The density of saltwater depends on the dissolved salt content as well as the temperature. Ice still floats in the oceans, otherwise, they would freeze from the bottom up. However, the salt content of oceans lowers the freezing point by about 1.9 °C (due to freezing-point depression of a solvent containing a solute) and lowers the temperature of the density maximum of water to the former freezing point at 0 °C. This is why, in ocean water, the downward convection of colder water is not blocked by an expansion of water as it becomes colder near the freezing point. The oceans' cold water near the freezing point continues to sink. So creatures that live at the bottom of cold oceans like the Arctic Ocean generally live in water 4 °C colder than at the bottom of frozen-over fresh water lakes and rivers. As the surface of saltwater begins to freeze (at −1.9 °C for normal salinity seawater, 3.5%) the ice that forms is essentially salt-free, with about the same density as freshwater ice. This ice floats on the surface, and the salt that is "frozen out" adds to the salinity and density of the seawater just below it, in a process known as brine rejection. This denser saltwater sinks by convection and the replacing seawater is subject to the same process. This produces essentially freshwater ice at −1.9 °C on the surface. The increased density of the seawater beneath the forming ice causes it to sink towards the bottom. On a large scale, the process of brine rejection and sinking cold salty water results in ocean currents forming to transport such water away from the Poles, leading to a global system of currents called the thermohaline circulation. Miscibility and condensation Water is miscible with many liquids, including ethanol in all proportions. Water and most oils are immiscible, usually forming layers according to increasing density from the top. This can be predicted by comparing the polarity. Water being a relatively polar compound will tend to be miscible with liquids of high polarity such as ethanol and acetone, whereas compounds with low polarity will tend to be immiscible and poorly soluble such as with hydrocarbons. As a gas, water vapor is completely miscible with air. On the other hand, the maximum water vapor pressure that is thermodynamically stable with the liquid (or solid) at a given temperature is relatively low compared with total atmospheric pressure. For example, if the vapor's partial pressure is 2% of atmospheric pressure and the air is cooled from 25 °C, starting at about 22 °C, water will start to condense, defining the dew point, and creating fog or dew. The reverse process accounts for the fog burning off in the morning. If the humidity is increased at room temperature, for example, by running a hot shower or a bath, and the temperature stays about the same, the vapor soon reaches the pressure for phase change and then condenses out as minute water droplets, commonly referred to as steam. A saturated gas or one with 100% relative humidity is when the vapor pressure of water in the air is at equilibrium with vapor pressure due to (liquid) water; water (or ice, if cool enough) will fail to lose mass through evaporation when exposed to saturated air. Because the amount of water vapor in the air is small, relative humidity, the ratio of the partial pressure due to the water vapor to the saturated partial vapor pressure, is much more useful. Vapor pressure above 100% relative humidity is called supersaturated and can occur if the air is rapidly cooled, for example, by rising suddenly in an updraft. Vapour pressure Compressibility The compressibility of water is a function of pressure and temperature. At 0 °C, at the limit of zero pressure, the compressibility is . At the zero-pressure limit, the compressibility reaches a minimum of around 45 °C before increasing again with increasing temperature. As the pressure is increased, the compressibility decreases, being at 0 °C and . The bulk modulus of water is about 2.2 GPa. The low compressibility of non-gasses, and of water in particular, leads to their often being assumed as incompressible. The low compressibility of water means that even in the deep oceans at depth, where pressures are 40 MPa, there is only a 1.8% decrease in volume. The bulk modulus of water ice ranges from 11.3 GPa at 0 K up to 8.6 GPa at 273 K. The large change in the compressibility of ice as a function of temperature is the result of its relatively large thermal expansion coefficient compared to other common solids. Triple point The temperature and pressure at which ordinary solid, liquid, and gaseous water coexist in equilibrium is a triple point of water. Since 1954, this point had been used to define the base unit of temperature, the kelvin, but, starting in 2019, the kelvin is now defined using the Boltzmann constant, rather than the triple point of water. Due to the existence of many polymorphs (forms) of ice, water has other triple points, which have either three polymorphs of ice or two polymorphs of ice and liquid in equilibrium. Gustav Heinrich Johann Apollon Tammann in Göttingen produced data on several other triple points in the early 20th century. Kamb and others documented further triple points in the 1960s. Melting point The melting point of ice is at standard pressure; however, pure liquid water can be supercooled well below that temperature without freezing if the liquid is not mechanically disturbed. It can remain in a fluid state down to its homogeneous nucleation point of about . The melting point of ordinary hexagonal ice falls slightly under moderately high pressures, by /atm or about /70 atm as the stabilization energy of hydrogen bonding is exceeded by intermolecular repulsion, but as ice transforms into its polymorphs (see crystalline states of ice) above , the melting point increases markedly with pressure, i.e., reaching at (triple point of Ice VII). Electrical properties Electrical conductivity Pure water containing no exogenous ions is an excellent electronic insulator, but not even "deionized" water is completely free of ions. Water undergoes autoionization in the liquid state when two water molecules form one hydroxide anion () and one hydronium cation (). Because of autoionization, at ambient temperatures pure liquid water has a similar intrinsic charge carrier concentration to the semiconductor germanium and an intrinsic charge carrier concentration three orders of magnitude greater than the semiconductor silicon, hence, based on charge carrier concentration, water can not be considered to be a completely dielectric material or electrical insulator but to be a limited conductor of ionic charge. Because water is such a good solvent, it almost always has some solute dissolved in it, often a salt. If water has even a tiny amount of such an impurity, then the ions can carry charges back and forth, allowing the water to conduct electricity far more readily. It is known that the theoretical maximum electrical resistivity for water is approximately 18.2 MΩ·cm (182 kΩ·m) at 25 °C. This figure agrees well with what is typically seen on reverse osmosis, ultra-filtered and deionized ultra-pure water systems used, for instance, in semiconductor manufacturing plants. A salt or acid contaminant level exceeding even 100 parts per trillion (ppt) in otherwise ultra-pure water begins to noticeably lower its resistivity by up to several kΩ·m. In pure water, sensitive equipment can detect a very slight electrical conductivity of 0.05501 ± 0.0001 μS/cm at 25.00 °C. Water can also be electrolyzed into oxygen and hydrogen gases but in the absence of dissolved ions this is a very slow process, as very little current is conducted. In ice, the primary charge carriers are protons (see proton conductor). Ice was previously thought to have a small but measurable conductivity of 1 S/cm, but this conductivity is now thought to be almost entirely from surface defects, and without those, ice is an insulator with an immeasurably small conductivity. Polarity and hydrogen bonding An important feature of water is its polar nature. The structure has a bent molecular geometry for the two hydrogens from the oxygen vertex. The oxygen atom also has two lone pairs of electrons. One effect usually ascribed to the lone pairs is that the H–O–H gas-phase bend angle is 104.48°, which is smaller than the typical tetrahedral angle of 109.47°. The lone pairs are closer to the oxygen atom than the electrons sigma bonded to the hydrogens, so they require more space. The increased repulsion of the lone pairs forces the O–H bonds closer to each other. Another consequence of its structure is that water is a polar molecule. Due to the difference in electronegativity, a bond dipole moment points from each H to the O, making the oxygen partially negative and each hydrogen partially positive. A large molecular dipole, points from a region between the two hydrogen atoms to the oxygen atom. The charge differences cause water molecules to aggregate (the relatively positive areas being attracted to the relatively negative areas). This attraction, hydrogen bonding, explains many of the properties of water, such as its solvent properties. Although hydrogen bonding is a relatively weak attraction compared to the covalent bonds within the water molecule itself, it is responsible for several of the water's physical properties. These properties include its relatively high melting and boiling point temperatures: more energy is required to break the hydrogen bonds between water molecules. In contrast, hydrogen sulfide (), has much weaker hydrogen bonding due to sulfur's lower electronegativity. is a gas at room temperature, despite hydrogen sulfide having nearly twice the molar mass of water. The extra bonding between water molecules also gives liquid water a large specific heat capacity. This high heat capacity makes water a good heat storage medium (coolant) and heat shield. Cohesion and adhesion Water molecules stay close to each other (cohesion), due to the collective action of hydrogen bonds between water molecules. These hydrogen bonds are constantly breaking, with new bonds being formed with different water molecules; but at any given time in a sample of liquid water, a large portion of the molecules are held together by such bonds. Water also has high adhesion properties because of its polar nature. On clean, smooth glass the water may form a thin film because the molecular forces between glass and water molecules (adhesive forces) are stronger than the cohesive forces. In biological cells and organelles, water is in contact with membrane and protein surfaces that are hydrophilic; that is, surfaces that have a strong attraction to water. Irving Langmuir observed a strong repulsive force between hydrophilic surfaces. To dehydrate hydrophilic surfaces—to remove the strongly held layers of water of hydration—requires doing substantial work against these forces, called hydration forces. These forces are very large but decrease rapidly over a nanometer or less. They are important in biology, particularly when cells are dehydrated by exposure to dry atmospheres or to extracellular freezing. Surface tension Water has an unusually high surface tension of 71.99 mN/m at 25 °C which is caused by the strength of the hydrogen bonding between water molecules. This allows insects to walk on water. Capillary action Because water has strong cohesive and adhesive forces, it exhibits capillary action. Strong cohesion from hydrogen bonding and adhesion allows trees to transport water more than 100 m upward. Water as a solvent Water is an excellent solvent due to its high dielectric constant. Substances that mix well and dissolve in water are known as hydrophilic ("water-loving") substances, while those that do not mix well with water are known as hydrophobic ("water-fearing") substances. The ability of a substance to dissolve in water is determined by whether or not the substance can match or better the strong attractive forces that water molecules generate between other water molecules. If a substance has properties that do not allow it to overcome these strong intermolecular forces, the molecules are precipitated out from the water. Contrary to the common misconception, water and hydrophobic substances do not "repel", and the hydration of a hydrophobic surface is energetically, but not entropically, favorable. When an ionic or polar compound enters water, it is surrounded by water molecules (hydration). The relatively small size of water molecules (~ 3 angstroms) allows many water molecules to surround one molecule of solute. The partially negative dipole ends of the water are attracted to positively charged components of the solute, and vice versa for the positive dipole ends. In general, ionic and polar substances such as acids, alcohols, and salts are relatively soluble in water, and nonpolar substances such as fats and oils are not. Nonpolar molecules stay together in water because it is energetically more favorable for the water molecules to hydrogen bond to each other than to engage in van der Waals interactions with non-polar molecules. An example of an ionic solute is table salt; the sodium chloride, NaCl, separates into cations and anions, each being surrounded by water molecules. The ions are then easily transported away from their crystalline lattice into solution. An example of a nonionic solute is table sugar. The water dipoles make hydrogen bonds with the polar regions of the sugar molecule (OH groups) and allow it to be carried away into solution. Quantum tunneling The quantum tunneling dynamics in water was reported as early as 1992. At that time it was known that there are motions which destroy and regenerate the weak hydrogen bond by internal rotations of the substituent water monomers. On 18 March 2016, it was reported that the hydrogen bond can be broken by quantum tunneling in the water hexamer. Unlike previously reported tunneling motions in water, this involved the concerted breaking of two hydrogen bonds. Later in the same year, the discovery of the quantum tunneling of water molecules was reported. Electromagnetic absorption Water is relatively transparent to visible light, near ultraviolet light, and far-red light, but it absorbs most ultraviolet light, infrared light, and microwaves. Most photoreceptors and photosynthetic pigments utilize the portion of the light spectrum that is transmitted well through water. Microwave ovens take advantage of water's opacity to microwave radiation to heat the water inside of foods. Water's light blue color is caused by weak absorption in the red part of the visible spectrum. Structure A single water molecule can participate in a maximum of four hydrogen bonds because it can accept two bonds using the lone pairs on oxygen and donate two hydrogen atoms. Other molecules like hydrogen fluoride, ammonia, and methanol can also form hydrogen bonds. However, they do not show anomalous thermodynamic, kinetic, or structural properties like those observed in water because none of them can form four hydrogen bonds: either they cannot donate or accept hydrogen atoms, or there are steric effects in bulky residues. In water, intermolecular tetrahedral structures form due to the four hydrogen bonds, thereby forming an open structure and a three-dimensional bonding network, resulting in the anomalous decrease in density when cooled below 4 °C. This repeated, constantly reorganizing unit defines a three-dimensional network extending throughout the liquid. This view is based upon neutron scattering studies and computer simulations, and it makes sense in the light of the unambiguously tetrahedral arrangement of water molecules in ice structures. However, there is an alternative theory for the structure of water. In 2004, a controversial paper from Stockholm University suggested that water molecules in the liquid state typically bind not to four but only two others; thus forming chains and rings. The term "string theory of water" (which is not to be confused with the string theory of physics) was coined. These observations were based upon X-ray absorption spectroscopy that probed the local environment of individual oxygen atoms. Molecular structure The repulsive effects of the two lone pairs on the oxygen atom cause water to have a bent, not linear, molecular structure, allowing it to be polar. The hydrogen–oxygen–hydrogen angle is 104.45°, which is less than the 109.47° for ideal sp3 hybridization. The valence bond theory explanation is that the oxygen atom's lone pairs are physically larger and therefore take up more space than the oxygen atom's bonds to the hydrogen atoms. The molecular orbital theory explanation (Bent's rule) is that lowering the energy of the oxygen atom's nonbonding hybrid orbitals (by assigning them more s character and less p character) and correspondingly raising the energy of the oxygen atom's hybrid orbitals bonded to the hydrogen atoms (by assigning them more p character and less s character) has the net effect of lowering the energy of the occupied molecular orbitals because the energy of the oxygen atom's nonbonding hybrid orbitals contributes completely to the energy of the oxygen atom's lone pairs while the energy of the oxygen atom's other two hybrid orbitals contributes only partially to the energy of the bonding orbitals (the remainder of the contribution coming from the hydrogen atoms' 1s orbitals). Chemical properties Self-ionization In liquid water there is some self-ionization giving hydronium ions and hydroxide ions. 2 + The equilibrium constant for this reaction, known as the ionic product of water, , has a value of about at 25 °C. At neutral pH, the concentration of the hydroxide ion () equals that of the (solvated) hydrogen ion (), with a value close to 10−7 mol L−1 at 25 °C. See data page for values at other temperatures. The thermodynamic equilibrium constant is a quotient of thermodynamic activities of all products and reactants including water: However, for dilute solutions, the activity of a solute such as H3O+ or OH− is approximated by its concentration, and the activity of the solvent H2O is approximated by 1, so that we obtain the simple ionic product Geochemistry The action of water on rock over long periods of time typically leads to weathering and water erosion, physical processes that convert solid rocks and minerals into soil and sediment, but under some conditions chemical reactions with water occur as well, resulting in metasomatism or mineral hydration, a type of chemical alteration of a rock which produces clay minerals. It also occurs when Portland cement hardens. Water ice can form clathrate compounds, known as clathrate hydrates, with a variety of small molecules that can be embedded in its spacious crystal lattice. The most notable of these is methane clathrate, 4 , naturally found in large quantities on the ocean floor. Acidity in nature Rain is generally mildly acidic, with a pH between 5.2 and 5.8 if not having any acid stronger than carbon dioxide. If high amounts of nitrogen and sulfur oxides are present in the air, they too will dissolve into the cloud and raindrops, producing acid rain. Isotopologues Several isotopes of both hydrogen and oxygen exist, giving rise to several known isotopologues of water. Vienna Standard Mean Ocean Water is the current international standard for water isotopes. Naturally occurring water is almost completely composed of the neutron-less hydrogen isotope protium. Only 155 ppm include deuterium ( or D), a hydrogen isotope with one neutron, and fewer than 20 parts per quintillion include tritium ( or T), which has two neutrons. Oxygen also has three stable isotopes, with present in 99.76%, in 0.04%, and in 0.2% of water molecules. Deuterium oxide, , is also known as heavy water because of its higher density. It is used in nuclear reactors as a neutron moderator. Tritium is radioactive, decaying with a half-life of 4500 days; exists in nature only in minute quantities, being produced primarily via cosmic ray-induced nuclear reactions in the atmosphere. Water with one protium and one deuterium atom occur naturally in ordinary water in low concentrations (~0.03%) and in far lower amounts (0.000003%) and any such molecules are temporary as the atoms recombine. The most notable physical differences between and , other than the simple difference in specific mass, involve properties that are affected by hydrogen bonding, such as freezing and boiling, and other kinetic effects. This is because the nucleus of deuterium is twice as heavy as protium, and this causes noticeable differences in bonding energies. The difference in boiling points allows the isotopologues to be separated. The self-diffusion coefficient of at 25 °C is 23% higher than the value of . Because water molecules exchange hydrogen atoms with one another, hydrogen deuterium oxide (DOH) is much more common in low-purity heavy water than pure dideuterium monoxide . Consumption of pure isolated may affect biochemical processes—ingestion of large amounts impairs kidney and central nervous system function. Small quantities can be consumed without any ill-effects; humans are generally unaware of taste differences, but sometimes report a burning sensation or sweet flavor. Very large amounts of heavy water must be consumed for any toxicity to become apparent. Rats, however, are able to avoid heavy water by smell, and it is toxic to many animals. Light water refers to deuterium-depleted water (DDW), water in which the deuterium content has been reduced below the standard level. Occurrence Water is the most abundant substance on Earth's surface and also the third most abundant molecule in the universe, after and . 0.23 ppm of the earth's mass is water and 97.39% of the global water volume of 1.38 km3 is found in the oceans. Water is far more prevalent in the outer Solar System, beyond a point called the frost line, where the Sun's radiation is too weak to vaporize solid and liquid water (as well as other elements and chemical compounds with relatively low melting points, such as methane and ammonia). In the inner Solar System, planets, asteroids, and moons formed almost entirely of metals and silicates. Water has since been delivered to the inner Solar System via an as-yet unknown mechanism, theorized to be the impacts of asteroids or comets carrying water from the outer Solar System, where bodies contain much more water ice. The difference between planetary bodies located inside and outside the frost line can be stark. Earth's mass is 0.000023% water, while Tethys, a moon of Saturn, is almost entirely made of water. Reactions Acid–base reactions Water is amphoteric: it has the ability to act as either an acid or a base in chemical reactions. According to the Brønsted-Lowry definition, an acid is a proton () donor and a base is a proton acceptor. When reacting with a stronger acid, water acts as a base; when reacting with a stronger base, it acts as an acid. For instance, water receives an ion from HCl when hydrochloric acid is formed: + + In the reaction with ammonia, , water donates a ion, and is thus acting as an acid: + + Because the oxygen atom in water has two lone pairs, water often acts as a Lewis base, or electron-pair donor, in reactions with Lewis acids, although it can also react with Lewis bases, forming hydrogen bonds between the electron pair donors and the hydrogen atoms of water. HSAB theory describes water as both a weak hard acid and a weak hard base, meaning that it reacts preferentially with other hard species: + → + → + → When a salt of a weak acid or of a weak base is dissolved in water, water can partially hydrolyze the salt, producing the corresponding base or acid, which gives aqueous solutions of soap and baking soda their basic pH: + NaOH + Ligand chemistry Water's Lewis base character makes it a common ligand in transition metal complexes, examples of which include metal aquo complexes such as to perrhenic acid, which contains two water molecules coordinated to a rhenium center. In solid hydrates, water can be either a ligand or simply lodged in the framework, or both. Thus, consists of [Fe2(H2O)6]2+ centers and one "lattice water". Water is typically a monodentate ligand, i.e., it forms only one bond with the central atom. Organic chemistry As a hard base, water reacts readily with organic carbocations; for example in a hydration reaction, a hydroxyl group () and an acidic proton are added to the two carbon atoms bonded together in the carbon-carbon double bond, resulting in an alcohol. When the addition of water to an organic molecule cleaves the molecule in two, hydrolysis is said to occur. Notable examples of hydrolysis are the saponification of fats and the digestion of proteins and polysaccharides. Water can also be a leaving group in SN2 substitution and E2 elimination reactions; the latter is then known as a dehydration reaction. Water in redox reactions Water contains hydrogen in the oxidation state +1 and oxygen in the oxidation state −2. It oxidizes chemicals such as hydrides, alkali metals, and some alkaline earth metals. One example of an alkali metal reacting with water is: 2 Na + 2 → + 2 + 2 Some other reactive metals, such as aluminium and beryllium, are oxidized by water as well, but their oxides adhere to the metal and form a passive protective layer. Note that the rusting of iron is a reaction between iron and oxygen that is dissolved in water, not between iron and water. Water can be oxidized to emit oxygen gas, but very few oxidants react with water even if their reduction potential is greater than the potential of . Almost all such reactions require a catalyst. An example of the oxidation of water is: 4 + 2 → 4 AgF + 4 HF + Electrolysis Water can be split into its constituent elements, hydrogen, and oxygen, by passing an electric current through it. This process is called electrolysis. The cathode half reaction is: 2 + 2 → The anode half reaction is: 2 → + 4 + 4 The gases produced bubble to the surface, where they can be collected or ignited with a flame above the water if this was the intention. The required potential for the electrolysis of pure water is 1.23 V at 25 °C. The operating potential is actually 1.48 V or higher in practical electrolysis. History Henry Cavendish showed that water was composed of oxygen and hydrogen in 1781. The first decomposition of water into hydrogen and oxygen, by electrolysis, was done in 1800 by English chemist William Nicholson and Anthony Carlisle. In 1805, Joseph Louis Gay-Lussac and Alexander von Humboldt showed that water is composed of two parts hydrogen and one part oxygen. Gilbert Newton Lewis isolated the first sample of pure heavy water in 1933. The properties of water have historically been used to define various temperature scales. Notably, the Kelvin, Celsius, Rankine, and Fahrenheit scales were, or currently are, defined by the freezing and boiling points of water. The less common scales of Delisle, Newton, Réaumur, and Rømer were defined similarly. The triple point of water is a more commonly used standard point today. Nomenclature The accepted IUPAC name of water is oxidane or simply water, or its equivalent in different languages, although there are other systematic names which can be used to describe the molecule. Oxidane is only intended to be used as the name of the mononuclear parent hydride used for naming derivatives of water by substituent nomenclature. These derivatives commonly have other recommended names. For example, the name hydroxyl is recommended over oxidanyl for the –OH group. The name oxane is explicitly mentioned by the IUPAC as being unsuitable for this purpose, since it is already the name of a cyclic ether also known as tetrahydropyran. The simplest systematic name of water is hydrogen oxide. This is analogous to related compounds such as hydrogen peroxide, hydrogen sulfide, and deuterium oxide (heavy water). Using chemical nomenclature for type I ionic binary compounds, water would take the name hydrogen monoxide, but this is not among the names published by the International Union of Pure and Applied Chemistry (IUPAC). Another name is dihydrogen monoxide, which is a rarely used name of water, and mostly used in the dihydrogen monoxide parody. Other systematic names for water include hydroxic acid, hydroxylic acid, and hydrogen hydroxide, using acid and base names. None of these exotic names are used widely. The polarized form of the water molecule, , is also called hydron hydroxide by IUPAC nomenclature. Water substance is a rare term used for H2O when one does not wish to specify the phase of matter (liquid water, water vapor, some form of ice, or a component in a mixture) though the term "water" is also used with this general meaning. Oxygen dihydride is another way of referring to water, but modern usage often restricts the term "hydride" to ionic compounds (which water is not).
Physical sciences
Inorganic compounds
null
2676138
https://en.wikipedia.org/wiki/Stabilizer%20%28aeronautics%29
Stabilizer (aeronautics)
An aircraft stabilizer is an aerodynamic surface, typically including one or more movable control surfaces, that provides longitudinal (pitch) and/or directional (yaw) stability and control. A stabilizer can feature a fixed or adjustable structure on which any movable control surfaces are hinged, or it can itself be a fully movable surface such as a stabilator. Depending on the context, "stabilizer" may sometimes describe only the front part of the overall surface. In the conventional aircraft configuration, separate vertical (fin) and horizontal (tailplane) stabilizers form an empennage positioned at the tail of the aircraft. Other arrangements of the empennage, such as the V-tail configuration, feature stabilizers which contribute to a combination of longitudinal and directional stabilization and control. Longitudinal stability and control may be obtained with other wing configurations, including canard, tandem wing and tailless aircraft. Some types of aircraft are stabilized with electronic flight control; in this case, fixed and movable surfaces located anywhere along the aircraft may serve as active motion dampers or stabilizers. Horizontal stabilizers A horizontal stabilizer is used to maintain the aircraft in longitudinal balance, or trim: it exerts a vertical force at a distance so the summation of pitch moments about the center of gravity is zero. The vertical force exerted by the stabilizer varies with flight conditions, in particular according to the aircraft lift coefficient and wing flaps deflection which both affect the position of the center of pressure, and with the position of the aircraft center of gravity (which changes with aircraft loading and fuel consumption). Transonic flight makes special demands on horizontal stabilizers; when the local speed of the air over the wing reaches the speed of sound there is a sudden move aft of the center of pressure. Another role of a horizontal stabilizer is to provide longitudinal static stability. Stability can be defined only when the vehicle is in trim; it refers to the tendency of the aircraft to return to the trimmed condition if it is disturbed. This maintains a constant aircraft attitude, with unchanging pitch angle relative to the airstream, without active input from the pilot. Ensuring static stability of an aircraft with a conventional wing requires that the aircraft center of gravity be ahead of the center of pressure, so a stabilizer positioned at the rear of the aircraft will produce lift in the downwards direction. The elevator serves to control the pitch axis; in case of a fully movable tail, the entire assembly acts as a control surface. Wing-stabilizer interaction The upwash and downwash associated with the generation of lift is the source of aerodynamic interaction between the wing and stabilizer, which translates into a change in the effective angle of attack for each surface. The influence of the wing on a tail is much more significant than the opposite effect and can be modeled using the Prandtl lifting-line theory; however, an accurate estimation of the interaction between multiple surfaces requires computer simulations or wind tunnel tests. Horizontal stabilizer configurations Conventional tailplane In the conventional configuration the horizontal stabilizer is a small horizontal tail or tailplane located to the rear of the aircraft. This is the most common configuration. On many aircraft, the tailplane assembly consists of a fixed surface fitted with a hinged aft elevator surface. Trim tabs may be used to relieve pilot input forces. Alternatively, some light aircraft such as the Piper PA-24 Comanche and the Piper PA-28 Cherokee have an all-moving stabilizer known as a stabilator, with no separate elevator. Stabilators are also found in many supersonic aircraft, where a separate elevator control would cause unacceptable drag. Most airliners and transport aircraft feature a large, slow-moving trimmable tail plane which is combined with independently-moving elevators. The elevators are controlled by the pilot or autopilot and primarily serve to change the aircraft's attitude, while the whole assembly is used to trim (maintaining horizontal static equilibrium) and stabilize the aircraft in the pitch axis. In the Boeing 737, the adjustable stabilizer trim system is powered by an electrically operated jackscrew. Variants on the conventional configuration include the T-tail, Cruciform tail, Twin tail and Twin-boom mounted tail. Three-surface aircraft Three-surface aircraft such as the Piaggio P.180 Avanti or the Scaled Composites Triumph and Catbird, the tailplane is a stabilizer as in conventional aircraft; the frontplane, called foreplane or canard, provides lift and serves as a balancing surface. Some earlier three-surface aircraft, such as the Curtiss AEA June Bug or the Voisin 1907 biplane, were of conventional layout with an additional front pitch control surface which was called "elevator" or sometimes "stabilisateur". Lacking elevators, the tailplanes of these aircraft were not what is now called conventional stabilizers. For example, the Voisin was a tandem-lifting layout (main wing and rear wing) with a foreplane that was neither stabilizing nor mainly lifting; it was called an "équilibreur" ("balancer"), and used as a pitch control and trim surface. Canard aircraft In the canard configuration, a small wing, or foreplane, is located in front of the main wing. Some authors call it a stabilizer or give to the foreplane alone a stabilizing role, although as far as pitch stability is concerned, a foreplane is generally described as a destabilizing surface, the main wing providing the stabilizing moment in pitch. In naturally unstable aircraft, the canard surfaces may be used as an active part of the artificial stability system, and are sometimes named horizontal stabilizers. Tailless aircraft Tailless aircraft lack a separate horizontal stabilizer. In a tailless aircraft, the horizontal stabilizing surface is part of the main wing. Longitudinal stability in tailless aircraft is achieved by designing the aircraft so that its aerodynamic center is behind the center of gravity. This is generally done by modifying the wing design, for example by varying the angle of incidence in the span-wise direction (wing washout or twist), or by using reflexed camber airfoils. Vertical stabilizers A vertical stabilizer provides directional (or yaw) stability and usually comprises a fixed fin and movable control rudder hinged to its rear edge.<ref>Daroll Stinton, The design of the aeroplane, lateral and directional stability and spinning</ref> Less commonly, there is no hinge and the whole fin surface is pivoted for both stability and control. When an aircraft encounters a horizontal gust of wind, yaw stability causes the aircraft to turn into the wind, rather than turn in the same direction. Fuselage geometry, engine nacelles and rotating propellers all influence lateral static stability and affect the required size of the stabilizer. Not all aircraft have a vertical stabilizer. Instead wing sweep and dihedral can provide a similar degree of directional stability, while directional control is often effected by adding drag on the side of the aircraft the aircraft is to be turned towards, either in the form of spoilers or split ailerons. Tailless directional stabilization and control Although the use of a vertical stabilizer is most common, it is possible to obtain directional stability with no discrete vertical stabilizer. This occurs when the wing is swept back and in some cases, as for example on the Rogallo wing often used for hang gliders, means that no fin is needed. Stabilization. When a swept wing is rotated in yaw the outer wing sweep is reduced, so increasing drag, while the inner wing sweep increases, reducing drag. This change in the drag distribution creates a restoring moment. Control. A way to get yaw control is to use differential air braking to affect the drag directly. This technique is suited to Electronic flight controls, as on the Northrop Grumman B-2 flying wing. Combined longitudinal–directional stabilizers On some aircraft, horizontal and vertical stabilizers are combined in a pair of surfaces named V-tail. In this arrangement, two stabilizers (fins and rudders) are mounted at 90–120° to each other, giving a larger horizontal projected area than vertical one as in the majority of conventional tails. The moving control surfaces are then named ruddervators.A portmanteau of rudder & elevator The V-tail thus acts as both a yaw and a pitch stabilizer. Although it may seem that the V-tail configuration can result in a significant reduction of the tail wetted area, it suffers from an increase in control-actuation complexity, as well as complex and detrimental aerodynamic interaction between the two surfaces. This often results in an upsizing in the total area that reduces or negates the original benefit. The Beechcraft Bonanza light aircraft was originally designed with a V-tail. Others combined layouts exist. The General Atomics MQ-1 Predator unmanned aircraft has an inverted V-tail. The tail surfaces of the Lockheed XFV could be described as a V-tail with surfaces that extended through the fuselage to the opposite side. The LearAvia Lear Fan had a Y-tail''. All twin tail arrangements with a tail dihedral angle will provide a combination of longitudinal and directional stabilization.
Technology
Aircraft components
null
2676271
https://en.wikipedia.org/wiki/Canid%20hybrid
Canid hybrid
Canid hybrids are the result of interbreeding between the species of the subfamily Caninae. Genetic considerations The wolf-like canids are a group of large carnivores that are genetically closely related because they all possess 78 chromosomes, arranged in 39 pairs and are karyologically indistinguishable from each other. The group includes the genera Canis, Cuon, Lupulella and Lycaon. The members are the domestic dog (C. lupus familiaris), gray wolf (C. lupus), dingo (C. lupus dingo), coyote (C. latrans), golden jackal (C. aureus), African wolf (C. lupaster), Ethiopian wolf (C. simensis), dhole (Cuon alpinus), black-backed jackal (Lupulella mesomelas), side-striped jackal (L. adusta) and African wild dog (Lycaon pictus). Newly proposed members include the red wolf (Canis rufus), and the eastern wolf (Canis lycaon), subject to a resolution of the dispute as to whether these constitute separate species in their own right or whether they are sub-species of the gray wolf. The members of Canis can potentially interbreed, however, it is believed that Cuon, Lupulella and Lycaon cannot breed with each other or with Canis. The Lupulella genus (the side-striped jackal and black-backed jackal), could theoretically interbreed with each other to produce fertile offspring, but a study of the maternal mitochondrial DNA of the black-backed jackal could find no evidence of genotypes from its most likely mate, the side-striped jackal, indicating that male black-backed jackals had not bred with their sister species. When the differences in number and arrangement of chromosomes is too great, hybridization becomes less and less likely. Other members of the wider dog family, Canidae, such as South American canids, true foxes, bat-eared foxes, or raccoon dogs which diverged 7 to 10 million years ago, are less closely related to the wolf-like canids, have fewer chromosomes and cannot hybridize with them. (recently proven, partly incorrect, see pampas fox with dog below) For instance, the red fox has 34 metacentric chromosomes and from 0 to 8 small B chromosomes, the raccoon dog has 42 chromosomes, and the fennec fox has 64 chromosomes. Wolf hybrids Wolfdog hybrid The domestic dog (Canis familiaris) is a domesticated species of the gray wolf (Canis lupus), along with the dingo (Canis lupus dingo). Therefore, crosses between these species are biologically unremarkable and not a hybridization in the same sense as an interbreeding between different species of Canidae. Wolves are different from domestic dogs in that wolves usually have slimmer chests, longer legs, and they also have stronger jaws than those of the domestic dog subspecies. The difference in appearance from the wolf to the domestic dog becomes even larger when a mix of the two animals is created. Wolfdogs do not have one common description of their appearance because it varies from one breeding cycle to the next. It differs from cycle to cycle because the number of wolf genes inherited in the animal differs greatly and is recorded in a percentage form. The general layout for describing the percentage of wolfdogs is as follows: 1-49% is considered low content (LC), 50-74% is considered to be mid-content (MC), and 75% and higher is considered to be high content (HC). The percentage of the amount of wolf in a wolfdog decides what the animal will look like. For example, if a wolfdog is 25% husky and 75% wolf, it will appear more like a wolf than a husky because it contains more genes from the wolf. This means that the appearance of the wolfdog will most likely contain a narrower chest, longer legs, and sharper teeth because it inherited more traits from the wolf parent. People wanting to improve domestic dogs or create an exotic pet may breed domestic dogs to wolves. Gray wolves have been crossed with dogs that have a wolf-like appearance, such as German Shepherds to form the Czechoslovakian Wolfdog. The breeding of wolf–dog crosses is controversial, with opponents purporting that it produces an animal unfit as a domestic pet. A number of wolfdog breeds are in development. The first generation crosses (one wolf parent, one dog parent) generally are backcrossed to domestic dogs to maintain a domestic temperament and consistent conformation. Dingo hybrids The dingo (Canis lupus dingo) breeds freely with other domestic dogs. This is now so widespread that in some areas, dingoes are now mostly mixed-breed dogs, crossed in recent times with dogs from other parts of the world. The dingo is closely related to the New Guinea singing dog though recent DNA sequencing of a 'pure' wild dingo from South Australia suggests that the dingo is 'intermediate' between wolves and domestic dogs. This would make dingos a subspecies of wolf and so interbreeding between dingos and domestic dogs is also not a hybridization in the same sense as an interbreeding between different species of Canidae. Some dingo hybrids are accepted back into the wild dingo population, where they breed with pure dingoes. The Australian Cattle Dog and Australian Stumpy Tail Cattle Dog breeds are known to have been created by crossing domesticated herding dogs, like the Collie, with the dingo. Coyote hybrids Coydogs Coydogs (the offspring of a male coyote and a female domestic dog) are naturally occurring red or blond color variations of the coyote and feral dogs. The breeding cycles of domestic dogs and coyotes are not synchronized and this makes interbreeding uncommon. If interbreeding had been common, each successive generation of the coyote population would have acquired more and more dog-like traits. Coywolves Hybridization between gray wolves and coyotes has long been recognized both in the wild and in captivity. In an evolutionary biology research conducted by a team of researchers in the Uppsala University, analysis of control region haplotypes of the mitochondrial DNA and sex chromosomes from Mexican wolves, a critically endangered subspecies of the gray wolf once nearly driven to extinction in the wild, confirmed the presence of coyote markers in some of the wolves. The study suggests that at some point in time, female coyotes managed to mate with some of the male wolves of the remnant wild Mexican wolf populations. Analysis on the haplotype of some coyotes from Texas also detected the presence of male wolf introgression, such as Y chromosomes from the gray wolves in the southern coyotes. In one cryptozoological investigation on a corpse of what was initially labelled as a chupacabra, examinations conducted by the UC Davis team and the Texas State University concluded based on the sex chromosomes that the male animal was in fact another coyote and wolf hybrid sired by a male Mexican wolf. DNA analysis consistently shows that all existing red wolves carry coyote genes. This has caused a problem for canid taxonomy, as hybrids are not normally thought of as species, though the convention is to continue to refer to red wolves as a subspecies of the gray wolf, Canis lupus rufus, with no mention of the coyote taxon latrans. In recent history, the taxonomic status of the red wolf has been widely debated. Mech (1970) suggested that red wolves may be fertile hybrid offspring from gray wolf (Canis lupus) and coyote (C. latrans) interbreeding. Wayne and Jenks (1991) and Roy et al. (1994b, 1996) supported this suggestion with genetic analysis. Phillips and Henry (1992) present logic supporting the contention that the red wolf is a subspecies of the gray wolf. However, recent genetic and morphological evidence suggests that the red wolf is a unique taxon. Wilson et al. (2000) report that gray wolves (Canis lupus lycaon) in southern Ontario appear genetically very similar to the red wolf and that these two canids may be subspecies of one another and not a subspecies of gray wolf. Wilson et al. (2000) propose that red wolves and C. lupus lycaon should be a separate species, C. lycaon, with their minor differences acknowledged via subspecies designation. North American wolf biologists and geneticists also concluded that C. rufus and C. lupus lycaon were genetically more similar to each other than either was to C. lupus or C. latrans (B. T. Kelly, unpubl.). In 2002, morphometric analyses of skulls also indicate that the red wolf is likely not to be a gray wolf–coyote hybrid (Nowak 2002). Therefore, while the red wolf's taxonomic status remains unclear, there is mounting evidence to support C. rufus as a unique canid taxon. Classifying animals commonly referred to as "eastern coyotes" or "northeastern coyotes" has become a problem for taxonomists, as it is unclear what new taxon will be used to refer to this new population of animals. African Canid hybrids The Ethiopian wolf's conservation is threatened by dog hybridisation. Animals resulting from Ethiopian wolf-dog hybridisation tend to be more heavily built than pure wolves, and have shorter muzzles and different coat patterns. Management plans for hybridization with dogs involve sterilization of known hybrids. Incidences of Ethiopian wolf-dog hybridization have been recorded in Bale's Web Valley. At least four hybrids were identified and sterilized in the area. Although hybridization has not been detected elsewhere, scientists are concerned that it could pose a threat to the wolf population's genetic integrity, resulting in outbreeding depression or a reduction in fitness, though this does not appear to have taken place. The African gold wolf is known to hybridize with both domestic dogs and Ethiopian wolves, as well as Golden jackals. Jackal hybrids Although hybridization between wolves and golden jackals has never been observed, evidence of such occurrences was discovered through mtDNA analysis on jackals in Bulgaria. Although there is no genetic evidence of gray wolf-jackal hybridization in the Caucasus Mountains, there have been cases where otherwise genetically pure golden jackals have displayed remarkably gray wolf-like phenotypes, to the point of being mistaken for wolves by trained biologists. In The Variation of Animals and Plants Under Domestication, Charles Darwin wrote:Several years ago, I saw confined in the Zoological Gardens of London a female hybrid from an English dog and jackal, which even in this the first generation was so sterile that, as I was assured by her keeper, she did not fully exhibit her proper periods; but this case, from numerous instances have occurred of fertile hybrids from these two animals, was certainly exceptional. Robert Armitage Sterndale mentioned experimental golden jackal/dog hybrids from British India in his Natural History of Mammals in India and Ceylon, noting that glaring jackal traits could be exhibited in hybrids even after three generations of crossing them with dogs. In Russia, golden jackal/Lapponian Herder hybrids were bred as sniffer dogs because jackals have a superior sense of smell and Lapponian Herders are good cold climate dogs. Also, Fox Terrier, Norwegian Lundehund, and Spitz blood were combined to create the Sulimov dog. As well as a superior sense of smell, important at low temperatures where substances are less volatile and therefore less pungent, Sulimov dogs are small-sized and can work in confined spaces. When tired, their normally curled tails droop, making it clear to the handler that the dog needs to be rested. The jackal hybrids were bred by Klim Sulimov, senior research assistant at the D.S. Likhachev Scientific Research Institute for Cultural Heritage and Environmental Protection in Russia. Male jackal pups had to be fostered on a Siberian Husky bitch in order to imprint the jackals on dogs. Female jackals accepted male Huskies more readily. The half-breed jackal-dogs were difficult to train and were bred back to Huskies to produce quarter-breed hybrids (quadroons). These hybrids were small, agile, trainable and had an excellent sense of smell. Twenty-five jackal-dog hybrids are used by Aeroflot at Sheremetyevo International Airport in Moscow for functions including bomb-sniffing. Their breeding program dates back to 1975, but it was not applied to bomb detection until 2002. Pampas fox hybrid Crossings between canids of a different genus is extremely rare. In 2021, a female canid with unusual phenotypic characteristics was found in Vacaria City, Rio Grande do Sul, Brazil. DNA analysis indicates that the canid was a hybrid between a pampas fox and a domestic dog. Dubbed a 'Dogxim' or 'graxorra', this finding is the first documented case of hybridisation detected between these two species. Legality Dog hybrids kept as pets are prohibited in certain jurisdictions, or are classed as wild animals and must be housed in the same way as purebred wolves. In the United States, legislation differs greatly from state to state. In New York, the law does not allow an individual to house or own a dog hybrid of any kind, even if there is a low percentage of wolf genes in the hybrid. States such as Indiana and Arkansas allow the ownership of hybrid animals, but they regulate it strictly with health records, immunization records, and registration of the animal, while other states, such as Arizona, do not have any laws about owning a wolfdog hybrid. States may or may not create their own laws regarding the issue of wolfdog hybrids.
Biology and health sciences
Hybrids
Animals
2679362
https://en.wikipedia.org/wiki/Negative%20impedance%20converter
Negative impedance converter
The negative impedance converter (NIC) is an active circuit which injects energy into circuits in contrast to an ordinary load that consumes energy from them. This is achieved by adding or subtracting excessive varying voltage in series to the voltage drop across an equivalent positive impedance. This reverses the voltage polarity or the current direction of the port and introduces a phase shift of 180° (inversion) between the voltage and the current for any signal generator. The two versions obtained are accordingly a negative impedance converter with voltage inversion (VNIC) and a negative impedance converter with current inversion (INIC). The basic circuit of an INIC and its analysis is shown below. Basic circuit and analysis INIC is a non-inverting amplifier (the op-amp and the voltage divider , on the figure) with a resistor () connected between its output and input. The op-amp output voltage is The current going from the operational amplifier output through resistor toward the source is , and So the input experiences an opposing current that is proportional to , and the circuit acts like a resistor with negative resistance In general, elements , , and need not be pure resistances (i.e., they may be capacitors, inductors, or impedance networks). Application By using an NIC as a negative resistor, it is possible to let a real generator behave (almost) like an ideal generator, (i.e., the magnitude of the current or of the voltage generated does not depend on the load). An example for a current source is shown in the figure on the right. The current generator and the resistor within the dotted line is the Norton representation of a circuit comprising a real generator and is its internal resistance. If an INIC is placed in parallel to that internal resistance, and the INIC has the same magnitude but inverted resistance value, there will be and in parallel. Hence, the equivalent resistance is That is, the combination of the real generator and the INIC will now behave like a composed ideal current source; its output current will be the same for any load . In particular, any current that is shunted away from the load into the Norton equivalent resistance will be supplied by the INIC instead. The ideal behavior in this application depends upon the Norton resistance and the INIC resistance being matched perfectly. As long as , the equivalent resistance of the combination will be greater than ; however, if , then the effect of the INIC will be negligible. However, when the circuit is unstable (e.g., when in an unloaded system). In particular, the surplus current from the INIC generates positive feedback that causes the voltage driving the load to reach its power supply limits. By reducing the impedance of the load (i.e., by causing the load to draw more current), the generator–NIC system can be rendered stable again. In principle, if the Norton equivalent current source was replaced with a Thévenin equivalent voltage source, a VNIC of equivalent magnitude could be placed in series with the voltage source's series resistance. Any voltage drop across the series resistance would then be added back to the circuit by the VNIC. However, a VNIC implemented as above with an operational amplifier must terminate on an electrical ground, and so this use is not practical. Because any voltage source with nonzero series resistance can be represented as an equivalent current source with finite parallel resistance, an INIC will typically be placed in parallel with a source when used to improve the impedance of the source. Negative impedance circuits The negative of any impedance can be produced by a negative impedance converter (INIC in the examples below), including negative capacitance and negative inductance. NIC can further be used to design floating impedances - like a floating negative inductor.
Technology
Functional circuits
null
4920126
https://en.wikipedia.org/wiki/Python%20%28genus%29
Python (genus)
Python is a genus of constricting snakes in the Pythonidae family native to the tropics and subtropics of the Eastern Hemisphere. The name python was proposed by François Marie Daudin in 1803 for non-venomous flecked snakes. Currently, 10 python species are recognized as valid taxa. Three formerly considered python subspecies have been promoted, and a new species recognized. Taxonomy The generic name Python was proposed by François Marie Daudin in 1803 for non-venomous snakes with a flecked skin and a long split tongue. In 1993, seven python species were recognized as valid taxa. On the basis of phylogenetic analyses, between seven and 13 python species are recognized. Distribution and habitat In Africa, pythons are native to the tropics south of the Sahara, but not in the extreme south-western tip of southern Africa (Western Cape) or in Madagascar. In Asia, they occur from Bangladesh, Nepal, India, Pakistan, and Sri Lanka, including the Nicobar Islands, through Myanmar, east to Indochina, southern China, Hong Kong and Hainan, as well as in the Malayan region of Indonesia and the Philippines. Invasive Some suggest that P. bivittatus and P. sebae have the potential to be problematic invasive species in South Florida. In early 2016, after a culling operation yielded 106 pythons, Everglades National Park officials suggested that "thousands" may live within the park, and that the species has been breeding there for some years. More recent data suggest that these pythons would not withstand winter climates north of Florida, contradicting previous research suggesting a more significant geographic potential range. Uses Python skin is used to make clothing, such as vests, belts, boots and shoes, or fashion accessories such as handbags. It may also be stretched and formed as the sound board of some string musical instruments, such as the erhu spike-fiddle, sanxian and the sanshin lutes. With a high demand of snake skin in the current fashion industry, countries in Africa and Southern Asia partake in the legal and illegal selling of python skin. Providing an extremely low pay for the hunters with an extremely high selling product for the consumers, there is an enormous gap between the beginning and end of the snake skin trade. As pets Many Python species, such as P. regius, P. brongersmai, P. bivittatus and M. reticulatus, are popular to keep as pets due to their ease of care, docile temperament, and vibrant colors, with some rare mutations having been sold for several thousands of dollars. Pythons in the pet trade are sourced from the wild, or from captive females that were taken from the wild ('captive-born'), or from parents that themselves were born in a captive setting ('captive-bred') Despite controversy that has arisen from media reports, with proper safety procedures pet pythons are relatively safe to own. Etymology The word 'Python' is derived from the Latin word 'pȳthon' and the Greek word 'πύθων', both referring to the "serpent slain, who was fabled to have been called Pythius in commemoration of his victory near Delphi by Apollo according to the myth".
Biology and health sciences
Snakes
Animals
4920788
https://en.wikipedia.org/wiki/Borrelia
Borrelia
Borrelia is a genus of bacteria of the spirochete phylum. Several species cause Lyme disease, also called Lyme borreliosis, a zoonotic, vector-borne disease transmitted by ticks. Other species of Borrelia cause relapsing fever, and are transmitted by ticks or lice, depending on the species of bacteria. A few Borrelia species as Candidatus Borrelia mahuryensis harbor intermediate genetic features between Lyme disease and relapsing fever Borrelia. The genus is named after French biologist Amédée Borrel (1867–1936), who first documented the distinction between a species of Borrelia, B. anserina, and the other known type of spirochete at the time, Treponema pallidum. This bacterium must be viewed using dark-field microscopy, which make the cells appear white against a dark background. Borrelia species are grown in Barbour-Stoenner-Kelly medium. Of 52 known species of Borrelia, 20 are members of the Lyme disease group (with an additional 3 proposed), 29 belong to the relapsing fever group, and two are members of a genetically distinct third group typically found in reptiles. A proposal has been made to split the Lyme disease group based on genetic diversity and move them to their own genus, Borelliella, but this change is not widely accepted. This bacterium uses hard and soft ticks and lice as vectors. Testing for the presence of the bacteria in a human includes two-tiered serological testing, including immunoassays and immunoblotting. Biology Borrelia species are members of the family Spirochaetaceae, so present the characteristic spirochete (spiral) shape. Most species are obligate anaerobes, although some are aerotolerant. Borrelia species have an outer membrane that contains a substance similar to lipopolysaccharides, an inner membrane, and a layer of peptidoglycan in a periplasmic space, which classifies them as Gram-negative. However, this result is not easily visualized using Gram staining. They are typically 20–30 μm long and 0.2–0.3 μm wide. Spirochetes move using axial filaments called endoflagella in their periplasmic space. The filaments rotate in this space, between the outer membrane and the peptidoglycan layer, propelling the bacterium forward in a corkscrew-like motion. The outer membrane of Borrelia species contains outer surface proteins (Osp) that play a role in their virulence. Phylogeny The currently accepted taxonomy is based on the List of Prokaryotic names with Standing in Nomenclature (LPSN) and National Center for Biotechnology Information (NCBI). Species incertae sedis:: "Ca. Borrelia africana" Ehounoud et al. 2016 "Ca. Borrelia algerica" Fotso et al. 2015 "Ca. Borrelia aligera" Norte et al. 2020 "Ca. Borrelia amblyommatis" corrig. Jiang et al. 2021 ["Ca. Borrelia javanense" Jiang et al. 2021] Borrelia baltazardii corrig. Karimi et al. 1979 ex Karimi et al. 1983 B. bissettii (Margos et al. 2016) Gupta 2020 Borrelia brasiliensis Davis 1952 "Ca. B. caatinga" de Oliveira et al. 2023 Borrelia caucasica (Kandelaki 1945) Davis 1957 Borrelia dugesii (Mazzotti 1949) Davis 1957 "Ca. Borrelia fainii" Qiu et al. 2019 Borrelia graingeri (Heisch 1953) Davis 1957 Borrelia harveyi (Garnham 1947) Davis 1948 "Ca. Borrelia ibitipocensis" corrig. Muñoz-Leal et al. 2020 "Ca. Borrelia ivorensis" Ehounoud et al. 2016 "Ca. Borrelia johnsonii" Schwan et al. 2009 "Ca. Borrelia kalaharica" Fingerle et al. 2016 Borrelia latyschewii (Sofiev 1941) Davis 1948 "Borrelia lonestari" Barbour et al. 1996 "Ca. Borrelia mahuryensis" Binetruy et al. 2020 Borrelia mazzottii Davis 1956 "Borrelia merionesi" Hougen 1974 non (Blanc & Maurice 1948) Davis 1948 "Borrelia microti" (Rafyi 1946) Davis 1948 "Ca. Borrelia mvumii" Mitani et al. 2004 "Borrelia myelophthora" (Steiner 1931) Ahrens & Muschner 1958 "Ca. Borrelia paulista" Weck et al. 2022 "B. rubricentralis" Gofton et al. 2023 "Ca. B. sibirica" Sabitova et al. 2022 Borrelia theileri (Laveran 1903) Bergey et al. 1925 Borrelia tillae Zumpt & Organ 1961 "B. undatumii" Gofton et al. 2023 "Borreliella andersonii" (Marconi, Liveris & Schwartz 1995) Adeolu & Gupta 2014 Borreliella carolinensis (Rudenko et al. 2011) Adeolu & Gupta 2015 "Borreliella lusitaniae" (Le Fleche et al. 1997) Adeolu & Gupta 2014 "Borreliella tanukii" (Fukunaga et al. 1997) Adeolu & Gupta 2014 Borreliella sinica (Masuzawa et al. 2001) Adeolu & Gupta 2015 "Ca. Borreliella texasensis" (Lin et al. 2005) Adeolu & Gupta 2014 Vectors Ticks Hard ticks of the family Ixodidae are common vectors of Borellia bacteria and are the only type of ticks shown to transmit Lyme disease bacteria to humans. Some tick species of the Ambylomma genus are vectors of Candidatus Borrelia mahuryensis in South America. Other species are carried by soft ticks. The soft tick Ornithodoros carries the species of Borellia that cause relapsing fever. Another species, B. anserina, is carried by the soft tick Argas. Inside the ticks, the bacteria grow in the midgut and then travel to the salivary glands to be transmitted to a new host. Ticks can spread the bacteria to each other when co-feeding. If an animal has been infected by a tick and then is bitten by a second tick, the second tick can become infected. The bacteria are most commonly transmitted to humans through ticks in the nymph stage of development, because they are smaller and less likely to be noticed and removed. The ticks must have around 36 to 48 hours of contact with a host to successfully transmit the bacteria. Lice Lice that feed on infected humans acquire the Borrelia organisms that then multiply in the hemolymph and gut of the lice. When an infected louse feeds on an uninfected human, the organism gains access when the victim crushes the louse or scratches the area where the louse is feeding. The U. S. Centers for Disease Control and Prevention reported that no credible evidence shows that lice can carry Borrelia. Pathology Lyme disease Of the 52 known species of Borrelia, 20 belong to the Lyme disease group and are transmitted by ticks. Eight are known to cause Lyme disease or Borreliosis. The major Borrelia species causing Lyme disease are Borrelia burgdorferi, Borrelia afzelii, and Borrelia garinii. All species that cause Lyme disease are referred to collectively as B. burgdorferi sensu lato, while B. burgdorferi itself is specified as B. burgdorferi sensu stricto. B. burgdorferi was previously believed to be the only species to cause Lyme disease in the US, but B. bissettiae and a new species called B. mayonii cause Lyme disease in the US, as well. The remaining five human pathogenic species occur only in Europe and Asia. Relapsing fever Relapsing fever (RF) borreliosis often occurs with severe bacteremia. Twenty-five species of Borrelia are known to cause relapsing fever. While most species use the soft tick family Argasidae as their vector, some outliers live in hard ticks or lice. Relapsing fever can be spread epidemically through lice or endemically through ticks. B. recurrentis, a common species underlying relapsing fever, is transmitted by the human body louse; no other animal reservoir of B. recurrentis is known. B. recurrentis infects the person via mucous membranes and then invades the bloodstream. Other tick-borne relapsing infections are acquired from other species, such as B. hermsii, B. parkeri, or B. miyamotoi, which can be spread from rodents, and serve as a reservoir for the infection, via a tick vector. B. hermsii and B. recurrentis cause very similar diseases, although the disease associated with B. hermsii has more relapses and is responsible for more fatalities, while the disease caused by B. recurrentis has longer febrile and afebrile intervals and a longer incubation period. Borellia miyamotoi disease Diagnosis Direct tests include culture of Borrelia from skin, blood, or cerebrospinal fluid (CSF), and detection of genetic material by polymerase chain reaction in skin, blood, or synovial fluid. Two-tiered serological testing is performed for differential diagnosis of Borrelia infection. The first-tier tests detect specific antibodies (IgM and IgG together or separately) and include enzyme-linked immunoassays (e.g. ELISAs) and immunofluorescent assays. Positive results for first-tier tests are confirmed using second-tier testing. The second tier consists of standardized immunoblotting, either by using Western blots or blots striped with diagnostically important purified antigens. Positive results for second-tier tests are confirmatory for the presence of Borrelia infection. Spirochetes can also be seen using Wright-stained blood smears.
Biology and health sciences
Spirochaetes
Plants
4923690
https://en.wikipedia.org/wiki/Body%20hair
Body hair
Body hair or androgenic hair is terminal hair that develops on the human body during and after puberty. It is different from head hair and also from less visible vellus hair, which is much finer and lighter in colour. Growth of androgenic hair is related to the level of androgens (male hormones) and the density of androgen receptors in the dermal papillae. Both must reach a threshold for the proliferation of hair follicle cells. From childhood onward, regardless of sex, vellus hair covers almost the entire area of the human body. Exceptions include the lips, the backs of the ears, palms of hands, soles of the feet, certain external genital areas, the navel, and scar tissue. Density of hair – i.e. the number of hair follicles per unit area of skin – varies from person to person. In many cases, areas on the human body that contain vellus hair will begin to produce darker and thicker body hair during puberty, such as the first growth of beard hair on a male and female adolescent's previously smooth chin; although it may appear thinner on the female. Androgenic hair follows the same growth pattern as the hair that grows on the scalp, but with a shorter anagen phase and longer telogen phase. While the anagen phase for the hair on one's head lasts for years, the androgenic hair growth phase for body hair lasts a few months. The telogen phase for hair lasts for varying lengths of time, depending on where the hair is, from a few weeks up to nearly a year. This shortened growing period and extended dormant period explains why the hair on the head tends to be much longer than other hair found on the body. Differences in length seen in comparing the hair on the back of the hand and pubic hair, for example, can be explained by varied growth cycles in those regions. The same goes for differences in body hair length seen in different people, especially when comparing men and women. Distribution Like much of the hair on the human body, leg, arm, chest, and back hair begin as vellus hair. As people age, the hair in these regions begins to grow darker and more abundantly. This growth occurs during or after puberty. Men will often have more abundant, coarser hair on the arms and back, while women tend to have a less drastic change in the hair growth in these areas but do experience a significant change in thickness of hairs. However, some women will grow darker, longer hair in one or more of these regions. Chest and abdomen Vellus hair grows on the chest and abdomen of both sexes at all stages of development. During the final stages of puberty and extending into adulthood, men grow increasing amounts of terminal hair over the chest and abdomen areas. Adult women can also grow terminal hairs around the areola, though in many cultures these hairs are removed. Arms Arm hair grows on a human's forearms, sometimes even on the elbow area, and rarely on a human's bicep, triceps, and/or shoulders. Terminal arm hair is concentrated on the wrist end of the forearm, extending over the hand. Terminal hair growth in adolescent males is often much more intense than that in females, particularly for individuals with dark hair. In some cultures, it is common for women to remove arm hair, though this practice is less frequent than that of leg hair removal. Terminal hair growth on arms is a secondary sexual characteristic in boys and appears in the last stages of puberty. Vellus arm hair is usually concentrated on the elbow end of the forearm and often ends on the lower part of the upper arm. This type of intense arm vellus hair growth sometimes occurs in girls and children of both sexes until puberty. Even though this causes the arms to appear hairy, it is not caused solely by testosterone. The hair is softer and different from terminal arm hair, in texture. The longest arm hair ever recorded was done so in California by David Reed in 2017. In 2024, Macie Davis-Southerland measured one hair at 7.24 inches long. Feet Visible hair appearing on the top surfaces of the feet and toes generally begins with the onset of puberty. Terminal hair growth on the feet is typically more intense in adult and adolescent males than in females. Legs Leg hair sometimes appears at the onset of adulthood, with the legs of men more often hairier than those of women. For a variety of reasons, people may shave their leg hair, including cultural practice or individual needs. Around the world, women generally shave their leg hair more regularly than men, to conform with the social norms of many cultures, many of which perceive smooth skin as a sign of youth, beauty, and in some cultures, hygiene. However, athletes of both sexes – swimmers, runners, cyclists and bodybuilders in particular – may shave their androgenic hair to reduce friction, highlight muscular development or to make it easier to get into and out of skin-tight clothing. Pubic Pubic hair is a collection of coarse hair found in the pubic region. It will often also grow on the thighs and abdomen. Zoologist Desmond Morris disputes theories that it developed to signal sexual maturity or protect the skin from chafing during copulation, and prefers the explanation that pubic hair acts as a scent trap. Also, both sexes having thick pubic hairs act as a partial cushion during intercourse. The genital area of males and females are first inhabited by shorter, lighter vellus hairs that are next to invisible and only begin to develop into darker, thicker pubic hair at puberty. At this time, the pituitary gland secretes gonadotropin hormones which trigger the production of testosterone in the testicles and ovaries, promoting pubic hair growth. The average ages pubic hair begins to grow in males and females are 12 and 11, respectively. However, in some females, pubic hair has been known to start growing as early as age 8. Just as individual people differ in scalp hair color, they can also differ in pubic hair color. Differences in thickness, growth rate, and length are also evident. Axillary Underarm hair normally starts to appear at the beginning of puberty, with growth usually completed by the end of the teenage years. Today in much of the world, it is common for women to regularly shave their underarm hair. The prevalence of this practice varies widely, though. The practice became popular for cosmetic reasons around 1915 in the United States and United Kingdom, when one or more magazines showed a woman in a dress with shaved underarms. Regular shaving became feasible with the introduction of the safety razor at the beginning of the 20th century. While underarm shaving was quickly adopted in some English speaking countries, especially in the US and Canada, it did not become widespread in Europe until well after World War II. Since then the practice has spread worldwide; some men also choose to shave their armpits. Facial Facial hair grows primarily on or around one's face. Both men and women experience facial hair growth. Like pubic hair, non-vellus facial hair will begin to grow in around puberty. Moustaches in young men usually begin to grow in at around the age of puberty, although some men may not grow a moustache until they reach late teens or at all. In some cases facial hair development may take longer to mature than the late teens, and some men experience no facial hair development even at an older age. It is common for many women to develop a few facial hairs under or around the chin, along the sides of the face (in the area of sideburns), or on the upper lip. These may appear at any age after puberty but are often seen in women after menopause due to decreased levels of estrogen. A darkening of the vellus hair of the upper lip in women is not considered true facial hair, though it is often referred to as a "moustache"; the appearance of these dark vellus hairs may be lessened by bleaching. A relatively small number of women are able to grow enough facial hair to have a distinct beard. In some cases, female beard growth is the result of a hormonal imbalance (usually androgen excess), or a rare genetic disorder known as hypertrichosis. Sometimes it is caused by use of anabolic steroids. Cultural pressure leads most women to remove facial hair, as it may be viewed as a social stigma. Development Hair follicles are to varying degrees sensitive to androgens, primarily testosterone and its derivatives, particularly dihydrotestosterone, with different areas on the body having different sensitivity. As androgen levels increase, the rate of hair growth and the weight of the hairs increase. Genetic factors determine both individual levels of androgen and the hair follicle's sensitivity to androgen, as well as other characteristics such as hair colour, type of hair and hair retention. Rising levels of androgens during puberty cause vellus hair to transform into terminal hair over many areas of the body. The sequence of appearance of terminal hair reflects the level of androgen sensitivity, with pubic hair being the first to appear due to the area's special sensitivity to androgen. The appearance of pubic hair in both sexes is usually seen as an indication of the start of a person's puberty. There is a sexual differentiation in the amount and distribution of androgenic hair, with men tending to have more terminal hair in more areas. This includes facial hair, chest hair, abdominal hair, leg hair, arm hair, and foot hair. (See Table 1 for development of male body hair during puberty.) Women retain more of the less visible vellus hair, although leg, arm, and foot hair can be noticeable on women. It is not unusual for women to have a few terminal hairs around their nipples as well. In the later decades of life, especially after the fifth decade, there begins a noticeable reduction in body hair especially in the legs. The reason for this is not known but it could be due to poorer circulation, lower free circulating hormone amounts or other reasons. Source: Function Androgenic hair provides tactile sensory input by transferring hair movement and vibration via the shaft to sensory nerves within the skin. Follicular nerves detect displacement of hair shafts and other nerve endings in the surrounding skin detect vibration and distortions of the skin around the follicles. Androgenic hair extends the sense of touch beyond the surface of the skin into the air and space surrounding it, detecting air movements as well as hair displacement from contact by insects or objects. Evolution Determining the evolutionary function of androgenic hair must take into account both human evolution and the thermal properties of hair itself. The thermodynamic properties of hair are based on the properties of the keratin strands and amino acids that combine into a 'coiled' structure. This structure lends to many of the properties of hair, such as its ability to stretch and return to its original length. This coiled structure does not predispose curly or frizzy hair, both of which are defined by oval or triangular hair follicle cross-sections. Evolution of less body hair Hair is a very good thermal conductor and aids heat transfer both into and out of the body. When goose bumps are observed, small muscles (arrector pili muscle) contract to raise the hairs either to provide insulation, by reducing cooling of the skin by air convection, or in response to central nervous stimulus, similar to the feeling of "hairs standing up on the back of your neck". This phenomenon also occurs when static charge is built up and stored in the hair. Keratin however can easily be damaged by excessive heat and dryness, suggesting that extreme sun exposure, perhaps due to a lack of clothing, would result in perpetual hair destruction, eventually resulting in the genes being bred out in favor of high skin pigmentation. It is also true that parasites can live on and in hair thus peoples who preserved their body hair would have required greater general hygiene to prevent diseases. Markus J. Rantala of the Department of Biological and Environmental Science, University of Jyväskylä, Finland, said humans evolved by "natural selection" to be hairless when the trade off of "having fewer parasites" became more important than having a "warming, furry coat". P. E. Wheeler of the Department of Biology at Liverpool Polytechnic said quadrupedal savanna mammals of similar volume to humans have body hair to keep warm while only larger quadrupedal savanna mammals lack body hair, because their body volume itself is enough to keep them warm. Therefore Wheeler said that humans, who should have body hair based on predictions of body volume alone for savanna mammals, evolved no body hair after evolving bipedalism, which he said reduced the amount of body area exposed to the sun by 40%, reducing the solar warming effect on the human body. Loss of fur occurred at least two million years ago, but possibly as early as 3.3 million years ago judging from the divergence of head and pubic lice, and aided persistence hunting (the ability to catch prey in very long distance chases) in the warm savannas where hominins first evolved. The two main advantages are felt to be bipedal locomotion and greater thermal load dissipation capacity due to better sweating and less hair. Sexual selection Markus J. Rantala of the Department of Biological and Environmental Science, University of Jyväskylä, Finland, said the existence of androgen dependent hair on men could be explained by sexual attraction whereby hair on the genitals would trap pheromones and hair on the chin would make the chin appear more massive. Across populations In 1876, Oscar Peschel wrote that North Asiatic Mongols, Native Americans, Malays, Hottentots and Bushmen have little to no body hair, while Semitic peoples, Indo-Europeans, and Southern Europeans (especially the Portuguese and Spanish) have extensive body hair. Anthropologist Joseph Deniker said in 1901 that the very hirsute peoples are the Ainus, Uyghurs, Iranians, Australian aborigines (Arnhem Land being less hairy), Toda, Dravidians and Melanesians, while the most glabrous peoples are the Indigenous Americans, San, and East Asians, who include Chinese, Koreans, Mongols, and Malays. Deniker said that hirsute peoples tend to have thicker beards, eyelashes, and eyebrows but fewer hairs on their scalp. C.H. Danforth and Mildred Trotter of the Department of Anatomy at Washington University in St. Louis conducted a study using American army soldiers of European origin in 1922 where they concluded that dark-haired white men are generally more hairy than fair-haired white men. H. Harris, publishing in the British Journal of Dermatology in 1947, wrote Native Americans have the least body hair, Han Chinese people and black people have little body hair, white people have more body hair than black people and Ainu have the most body hair. Anthropologist Arnold Henry Savage Landor described the Ainu as having hairy bodies. Stewart W. Hindley and Albert Damon of the Department of Anthropology at Stanford University studied, in 1973, the frequency of hair on the middle finger joint (mid-phalangeal hair) of Solomon Islanders, as a part of a series of anthropometric studies of these populations. They summarize other studies on prevalence of this trait as reporting, in general, that Caucasoids are more likely to have hair on the middle finger joint than Negroids, Australoids and Mongoloids, and collect the following frequencies from previously published literature: Andamanese 0%, Inuit 1%, African American 16% or 28%, Ethiopians 25.6%, Mexicans of the Yucatan 20.9%, Penobscot and Shinnecock 22.7%, Gurkha 33.6%, Japanese 44.6%, various Hindus 40–50%, Egyptians 52.3%, Near Eastern peoples 62–71%, various Europeans 60–80%. However, they never completed an Androgenic hair map. According to anthropologist and professor Ashley Montagu in 1989, many East Asian people and African populations such as the San people are less hairy than Europeans and West Asian peoples. Montagu said that the hairless feature is a neotenous trait. Eike-Meinrad Winkler and Kerrin Christiansen of the Institut für Humanbiologie studied, in 1993, the Kavango people and !Kung people from Namibia of body hair and hormone levels to investigate the reason some Africans did not have bodies as hairy as Europeans. Winkler and Christiansen concluded the difference in hairiness between some Africans and Europeans had to do with differences in androgen or estradiol production, in androgen metabolism, and in sex hormone action in the target cells. Valerie Anne Randall of the Department of Biomedical Sciences, University of Bradford, said in 1994 beard growth in Caucasian men increases until the mid-thirties due to a delay caused by growth cycles changing from vellus hair to terminal hair. Randall said white men and women are hairier than Japanese men and women even with the same total plasma androgen levels. Randall says that the reason for some people being hairy and some people not being hairy is unclear, but that it probably is related to differing sensitivity of hair follicles to 5α-reductase. Rodney P. R. Dawber of the Oxford Hair Foundation said in 1997 that East Asian males have little or no facial or body hair and Dawber also said that Mediterranean males are covered with an exuberant pelage. Milkica Nešić and her colleagues from the Department of Physiology at the University of Niš, Serbia, cited prior studies in a 2010 publication indicating that the frequency of hair on the middle finger joint (mid-phalangeal hair) in Europeans is significantly higher than in African populations, where the lowest values were found, and "completely absent" in Northern Native American (Inuit) populations. Their own study found that the latter was part of a wider trend of "Mongoloid" peoples having less hair overall. Androgenic hair as biometric It has been shown that individuals can be uniquely identified by their androgenic hair patterns. For example, even when one's particular distinguishing features such as face and tattoos are obscured, persons can still be identified by their hair on other parts of their body.
Biology and health sciences
Integumentary system
Biology
4924787
https://en.wikipedia.org/wiki/Adansonia%20digitata
Adansonia digitata
Adansonia digitata, the African baobab, is the most widespread tree species of the genus Adansonia, the baobabs, and is native to the African continent and the southern Arabian Peninsula (Yemen, Oman). These are long-lived pachycauls; radiocarbon dating has shown some individuals to be over 2,000 years old. They are typically found in dry, hot savannas of sub-Saharan Africa, where they dominate the landscape and reveal the presence of a watercourse from afar. They have traditionally been valued as sources of food, water, health remedies or places of shelter and are a key food source for many animals. They are steeped in legend and superstition. In recent years, many of the largest, oldest trees have died, for unknown reasons. Common names for the baobab include monkey-bread tree, upside-down tree, and cream of tartar tree. Description African baobabs are trees that often grow as solitary individuals, and are large and distinctive elements of savanna or scrubland vegetation. They grow to a height of . The trunk is typically very broad and fluted or cylindrical, often with a buttressed, spreading base. Trunks may reach a diameter of , and may be made up of multiple stems fused around a hollow core. The hollow core found in many tree species is the result of wood removal, such as decay of the oldest, internal part of the trunk. In baobabs, however, many of the largest and oldest of the trees have a hollow core that is the result of a fused circle of three to eight stems sprouting from roots. The bark is gray and usually smooth. The main branches can be massive. All baobabs are deciduous, losing their leaves in the dry season, and remaining leafless for about eight months of the year. Flowers are large, white and hanging. Fruits are rounded with a thick shell. The leaves are palmately compound with five to seven (sometimes up to nine) leaflets in mature trees, but seedlings and regenerating shoots may have simple leaves. The transition to compound leaves comes with age and may be gradual. African baobabs produce simple leaves much longer than most other Adansonia species. Leaflets are stalkless (sessile) to short-stalked and size is variable. Flowering occurs in both the dry and the wet season. Buds are rounded with a cone-shaped tip. Flowers are showy and sometimes paired, but usually produced singly at the end of a hanging stalk about in length. The calyx is typically made up of five (sometimes three) green triangular bent-back lobes (sepals) with a cream-coloured, hairy interior. The petals are white, roughly the same width and length – up to , and are crumpled in bud. Flowers open during the late afternoon, staying open and fertile for only one night. The fresh flowers have a sweet scent, but after about 24 hours, they start to turn brown and emit a carrion smell. The androecium is white and made up of a long tube of fused stamens (a staminal tube) surrounded by unfused (free) filaments 3–5 cm long. There are a large number of stamens, 720–1,600 per flower, with reports of up to 2,000. Styles are white, growing through the staminal tube and projecting beyond it. They are usually bent at right-angles and topped with an irregular stigma. Pollen grains are spherical with spikes over the surface, typical of the Malvaceae family. Pollen grain diameter is around 50 microns. All Adansonia develop large rounded indehiscent fruits which can be up to long with a woody outer shell. African baobab fruits are quite variable in shape, from nearly round to cylindrical. The shell is thick. Inside is a fleshy, light beige coloured pulp. As it dries, the pulp hardens into a crumbly powder. The seeds are hard and kidney-shaped with a .06-mm-thick coat. They show long-term dormancy, only germinating after fire or passing through an animal's digestive tract. It is thought that this is because the seed coat needs to be cracked or thinned to allow to water to penetrate before the seed can germinate. Water storage Baobab trees store water in their trunks and branches on a seasonal basis as they live in areas of sustained drought and water inaccessibility. The spongy material of the bark allows water to be absorbed deeper into the tissue, as there is rarely enough rain during the wet season to penetrate the litter layer of soil. The U-shaped branches allow for water to trickle down, allowing for maximum absorption over an extended period of time even after the rain stops. The water is absorbed into the vascular tissue of the tree, where it can be moved into the tree's parenchyma cells for long-term storage, or used. A large Baobab can store as much as 136,400 liters of water. During the dry season, the trees will flush out all of their leaves. During this period, the circumference of the trunk will shrink about 2–3 cm and the water content of the stem will drop by about 10%. Dropping leaves during the dry season is done to prevent water loss through transpiration out of the stomata, which would cause the water potentials in the vascular tissue to drop too low and pull water out of the vacuoles in the parenchyma cells. This would lead to the parenchyma cells, which make up the majority of the trunk and branches, to plasmolyze destroying the tree. The water in storage cells is structurally important, which limits their ability to use mass quantities of stored water in times of drought. Baobab trees have much higher water and parenchyma content than most trees, this allows them to grow very large with less energy expenditure. Parenchyma are soft plant tissue cells that are commonly used for water storage in other drought tolerant species like cactus and succulents. The water fluxes from the vascular tissue into the parenchyma cells at the center of the tree with the help of actively transported ions. The ion flux into the cell will shift the concentration gradients, causing water to rush into the cells for long-term storage. Another reason why the water in the trunk can only be used as a buffer for long-term deficits is the distance between the vascular tissue and the parenchyma. The transportation of water from the vascular tissue into storage cells is a very slow process as it is a high-resistance path. The water in the cells at the core of the trunk and the branches would take too much energy from the tree to move back into the vascular tissue for daily use. Longevity The growth rate of baobab trees is determined by ground water or rainfall. The trees produce faint growth rings, but counting growth rings is not a reliable way to age baobabs because some years a tree will form multiple rings and some years none. Radiocarbon dating has provided data on a few individual A. digitata specimens. The Panke baobab in Zimbabwe was some 2,450 years old when it died in 2011, making it the oldest angiosperm ever documented, and two other trees—Dorslandboom in Namibia and Glencoe in South Africa—were estimated to be approximately 2,000 years old. Another specimen known as Grootboom was dated after it died and found to be at least 1,275 years old. Baobabs may be so long-lived in part due to their ability to periodically sprout new stems. Taxonomy The scientific name Adansonia refers to the French explorer and botanist, Michel Adanson (1727–1806), who wrote the first botanical description for the full species. "Digitata" refers to the digits of the hand, as the baobab has compound leaves with normally five (but up to seven) leaflets, akin to a hand. A. digitata is the type species for the genus Adansonia and is the only species in the section Adansonia. All species of Adansonia except A. digitata are diploid; A. digitata is tetraploid. Some populations of African baobab have significant genetic differences and it has been suggested that the taxon contains more than one species. For example, the shape of the fruit varies considerably from region to region. In Angola, the fruits are elongated, rather than round. A proposed new species (Adansonia kilima Pettigrew, et al.), was described in 2012, found in high-elevation sites in eastern and southern Africa. This is now however no longer recognized as a distinct species but considered a synonym of A. digitata. Some high-elevation trees in Tanzania show different genetics and morphology but further study is needed to determine if they should be considered a separate species. History The earliest written reports of African baobab are from a 14th-century travelogue by the Arab traveler Ibn Batuta. The first botanical description was by Alpino (1592) looking at fruits that he observed in Egypt from an unknown source. They were called Bahobab, possibly from the Arabic "bu hibab", meaning "many-seeded fruit". The French explorer and botanist, Michel Adanson observed a baobab tree in 1749 on the island of Sor, Senegal and wrote the first detailed botanical description of the full tree, accompanied with illustrations. Recognizing the connection to the fruit described by Alpino he called the genus Baobab. Linnaeus later renamed the genus Adansonia, to honour Adason, but use of baobab as one of the common names has persisted. Additional common names include monkey-bread tree (the soft, dry fruit is edible), upside-down tree (the sparse branches resemble roots), and cream of tartar tree (cream of tartar) because of the powdery fruit pulp. Distribution and habitat The African Baobab is associated with tropical savannas. It is found in drier climates, is sensitive to water logging and frost and is not found in areas where sand is deep. It is native to mainland Africa, between the latitudes 16° N and 26° S. Some references consider it as introduced to Yemen and Oman while others consider it native there. The tree has also been introduced to many other regions including Australia and Asia. The northern limit of its distribution in Africa is associated with rainfall patterns; only on the Atlantic coast and in the Sudanian savanna does its occurrence venture naturally into the Sahel. On the Atlantic coast, this may be due to spreading after cultivation. Its occurrence is very limited in Central Africa, and it is found only in the very north of South Africa. In East Africa, the trees grow also in shrublands and on the coast. In Angola and Namibia, the baobabs grow in woodlands, and in coastal regions, in addition to savannas. The African Baobab is native to Mauritania, Senegal, Guinea, Sierra Leone, Mali, Burkina Faso, Ghana, Togo, Benin, Niger, Nigeria, northern Cameroon, Chad, Sudan, Congo Republic, DR Congo (formerly Zaire), Eritrea, Ethiopia, southern Somalia, Kenya, Tanzania, Zambia, Zimbabwe, Malawi, Mozambique, Angola, São Tomé, Príncipe, Annobon, South Africa (in Limpopo province, north of the Soutpansberg mountain range), Namibia, Botswana. It is an introduced species in Java, Nepal, Sri Lanka, Philippines, Jamaica, Puerto Rico, Haiti, Dominican Republic, Venezuela, Seychelles, Comoros, India, Guangdong, Fujian, Yunnan and has been planted in Penang, Malaysia, along certain streets. Arab traders introduced it to northwestern Madagascar where baobab trees were often planted at the center of villages. Ecology All baobabs are deciduous, losing their leaves in the dry season, and remaining leafless for about eight months of the year. The African baobab is largely found in savannah habitats, which tend to be fire-prone. Adaptations to survive frequent fires include a thick and fire-resistant bark and thick-shelled fruit. Trees older than about 15 years have thick enough bark to withstand the heat of most savannah fires, while younger trees can resprout after fire. The thick outer shell of the fruit may serve to protect the seeds. Pollination in the African baobab is achieved primarily by fruit bats, in West Africa mainly the straw-coloured fruit bat, Gambian epauletted fruit bat, and the Egyptian fruit bat. The flowers are also visited by galagos, and several kinds of insect. With their hard coat, baobab seeds can withstand drying and remain viable over long periods. The fruits are eaten by many species and the germination potential is improved when seeds have passed through the digestive tract of an animal or have been subjected to fire. Elephants and baboons are main dispersal agents and so the seeds can potentially be dispersed over long distances. The fruits float and the seeds are waterproof, so African baobabs may also be spread by water. Some aspects of the baobab's reproductive biology are not yet understood but it is thought that pollen from another tree may be required to develop fertile seed. Isolated trees without a pollen source from another tree do form fruit, only to abort them at a later stage. The existence of some very isolated trees may then be due to their ability to disperse long distances but self-incompatibility. The fruit, bark, roots and leaves are a key food source for many animals and the trees themselves are an important source of shade and shelter. Conservation The baobab is a protected tree in South Africa, and yet is threatened by various mining and development activities. In the Sahel, the effects of drought, desertification and over-use of the fruit have been cited as causes for concern. As of March 2022 African baobab is not yet classified by the IUCN Red List, although there is evidence that populations may be declining. Many of the largest and oldest African baobabs have died in recent years. Greenhouse gases, climate change, and global warming appear to be factors reducing baobab longevity. Uses People have traditionally valued the trees as sources of food, water, health remedies or places of shelter. The baobab is a traditional food plant in Africa, but is little-known elsewhere. Adanson concluded that the baobab, of all the trees he studied, "is probably the most useful tree in all." He consumed baobab juice twice a day while in Africa, and was convinced that it maintained his health. According to a modern field guide, the juice can help cure diarrhoea. The roots and fruits are edible. The fruit has been suggested to have the potential to improve nutrition, boost food security, foster rural development and support sustainable land care. In Sudanwhere the tree is called tebeldi تبلديpeople make tabaldi juice by soaking and dissolving the dry pulp of the fruit in water, locally known as gunguleiz. Water can also be extracted from some of the trunks. Baobab leaves can be eaten as a relish. Young fresh leaves are cooked in a sauce and sometimes are dried and powdered. The powder is called lalo in Mali and sold in many village markets in Western Africa. The leaves are used in the preparation of a soup termed miyan kuka in Northern Nigeria and are rich in phytochemicals and minerals. The seeds can be pounded into a flour or to extract oil for cooking. Baobab leaves are sometimes used as forage for ruminants in dry season. The oilmeal, which is a byproduct of oil extraction, can also be used as animal feed. Whole fruits or just the fruit pulp can be stored for months under dry conditions. The fiber of the bark can be used to make cloth. In times of drought, elephants consume the juicy wood beneath the bark of the baobab. For export In 2008, the European Union approved the use and consumption of baobab fruit. It is commonly used as an ingredient in smoothies and cereal bars. In 2009, the United States Food and Drug Administration granted generally recognized as safe status to baobab dried fruit pulp as a food ingredient. In culture Along the Zambezi, the tribes believed that baobabs were upright and too proud. The gods became angry and uprooted them and threw them back into the ground upside-down. Evil spirits now cause bad luck to anyone that picks up the sweet white flowers. More specifically, a lion will kill them. In Kafue National Park, one of the largest baobabs is known as "Kondanamwali" or the "tree that eats maidens". The tree fell in love with four beautiful maidens. When they reached puberty, they made the tree jealous by finding husbands. So, one night, during a thunderstorm, the tree opened its trunk and took the maidens inside. A rest house has been built in the branches of the tree. On stormy nights, the crying of the imprisoned maidens can still be heard. Some people believe that women living in kraals where baobabs are plenty will have more children. This is scientifically plausible as those women will have better access to the tree's vitamin-rich leaves and fruits to complement a vitamin-deficient diet. The tree also plays a role in Antoine De Saint-Exupéry's fictional children's book, The Little Prince. In the story, baobabs are described as dangerous plants which must be weeded out from the good plants, lest they overcome a small planet and even break it to pieces. Prominent specimens A number of individual baobab trees attract sightseers due to their age, size, history, location or isolated occurrence. Botswana Around Gweta, Botswana, some have been declared national monuments. Green's Baobab, 27 km south of Gweta was inscribed by the 19th-century hunters and traders Frederick Thomas Green and Hendrik Matthys van Zyl besides other ruthless characters. Fred and Charles Green passed the baobab during an expedition to Lake Ngami and left the inscription "Green's Expedition 1858–1859". An earlier inscription by an unknown traveller reads "1771". About 11 km south of Green's Baobab is the turn-off to Chapman's Baobab, also known as Seven Sisters or Xaugam, i.e. "lion's tail" in Tsoa. It was once an enormous multi-stemmed tree, used by passing explorers, traders and travellers as a navigation beacon. It guided them as they navigated the extensive salt pan northwards, while a hollow in the trunk served as a letterbox. The explorer and hunter James Chapman left an engraving on a large root when he passed the tree with artist Thomas Baines in 1861, but Livingstone, Oswell, Moffat, and Selous also camped here. Livingstone supposedly carved a cross and his initials, and conveyed his 1853 sojourn in Missionary Travels, noting: "about two miles beyond [the immense saltpan Ntwetwe] we unyoked under a fine specimen of baobab, ... It consisted of 6 branches united into one trunk." It had a circumference of 25 m before its constituent trunks collapsed outward on 7 January 2016. Not all its trunks are confirmed dead however, one showing signs of life in 2019. Seven trees known as the Sleeping Sisters or Baines' Baobabs grow on a tiny islet in Kudiakam Pan, Botswana. They are named for Thomas Baines who painted them in May 1862, while en route to Victoria Falls. The fallen giant of Baines' day is still sprouting leaves (as of 2004), and a younger generation of trees are in evidence. The islet is accessible in winter when the pan is dry. Some large specimens have been transplanted to new sites, as was the one at Cresta Mowana lodge in Kasane. Ghana At Saakpuli (also Sakpele) in northern Ghana the site of a 19th-century slave transit camp is marked by a stand of large baobabs, to which slaves were chained. The chains were wrapped around their trunks or around the roots. Similarly, two trees at Salaga in central Ghana are reminders of the slave trade. One, located at the former slave market at the center of town, was replanted at the site of the original to which slaves were shackled. A second larger tree marks the slave cemetery, where bodies of dead slaves were dumped. India Inside the Golconda Fort in Hyderabad, India, is a baobab tree estimated to be 430 years old. It is the largest baobab outside of Africa. Sri Lanka It grows in Mannar peninsula and opposite mainland, Delft island, Wilpaththu and Puththam. Baobab has Tamil vernacular names – Perukku-Maran and Papparappuli. English Name 'Monkey bread. Sinhala name - Aliyagaha (Sri lanka wild life interlude vol l ) It is said that the tree in Pallimunai of Mannar island is the oldest and largest one of 800 years old. Local tradition is that this tree brought to SL by Arabs to feed their camels by its leaves. Madagascar The African baobab in Mahajanga, Madagascar, had a circumference of 21 metres by 2013. It became the symbol of the city and was formerly a place for executions and important meetings. Mozambique The Lebombo Eco Trail tree is about 18.5 m tall with a diameter of almost 22 m. It was found to be about 1400 years old and made up of five stems with ages between 900 and 1400 years, fused in a ring leaving a large central cavity. Namibia The Ombalantu baobab in Namibia has a hollow trunk that can accommodate some 35 people. At times it has served as a chapel, post office, house, and hiding site. The Holboom baobab (Holboom, Nyae Nyae Conservancy, Namibia) is one of the trees with a hollow core. It measures 35.10 m around and radiocarbon dating shows it to be about 1750 years old. Republic of the Congo The Arbre de Brazza is a baobab in the Republic of the Congo under which de Brazza and his companions Dolisie, Chavannes and Ballay made a stop in 1877, as their engraving "EB 1887" still attests. Another engraving, "Mâ Prince", was left by president Nguesso in his youth. Senegal The first botanical description of A. digitata was done by Adanson based on a tree on the island of Sor, Senegal. On the nearby Îles des Madeleines Adanson found a baobab that was in diameter, which bore the carvings of passing mariners on its trunk, including those of Prince Henry the Navigator in 1444 and André Thevet in 1555. When Théodore Monod searched the island in the 20th century, this tree was not to be found. The Gouye Ndiouly or Guy Njulli ("baobab of circumcision") may be the oldest baobab in Senegal and the northern hemisphere. The partially collapsed tree from which new stems have emerged is situated near the bank of the Saloum River at Kahone. It was formerly the venue for the gàmmu, an annual festival during which the kingdom's provincial rulers pledged their loyalty to the king. From 1593 to 1939, 49 kings of the Guélewars dynasty were inducted at this tree. It was beside the place where the Buur Saloum organized circumcision ceremonies, and in 1862, it became the scene of a battle. US Virgin Islands The Grove Place Baobab, listed as a Champion Tree, is believed to be the oldest (250–300 years) of some 100 baobabs on Saint Croix in the US Virgin Islands. It is seen as a living testament to centuries of African presence, as the seeds were likely introduced by an African slave who arrived at the former estate during the 18th century. According to the bronze memorial plaque, twelve women were rounded up during the 1878 Fireburn labor riot, and burned alive beneath the tree. It has since been a rallying place for plantation laborers and unions. Zimbabwe Zimbabwe's Big Tree, near Victoria Falls, stands 25 meters tall and is visited by hundreds of thousands of tourists yearly. Radiocarbon dating has shown this one to be made up of several stems of various ages, with the oldest about 1150 years old. Additional images
Biology and health sciences
Malvales
Plants
4926056
https://en.wikipedia.org/wiki/Insect%20wing
Insect wing
Insect wings are adult outgrowths of the insect exoskeleton that enable insects to fly. They are found on the second and third thoracic segments (the mesothorax and metathorax), and the two pairs are often referred to as the forewings and hindwings, respectively, though a few insects lack hindwings, even rudiments. The wings are strengthened by a number of longitudinal veins, which often have cross-connections that form closed "cells" in the membrane (extreme examples include the dragonflies and lacewings). The patterns resulting from the fusion and cross-connection of the wing veins are often diagnostic for different evolutionary lineages and can be used for identification to the family or even genus level in many orders of insects. Physically, some insects move their flight muscles directly, others indirectly. In insects with direct flight, the wing muscles directly attach to the wing base, so that a small downward movement of the wing base lifts the wing itself upward. Those insects with indirect flight have muscles that attach to and deform the thorax, causing the wings to move as well. The wings are present in only one sex (often the male) in some groups such as velvet ants and Strepsiptera, or are selectively lost in "workers" of social insects such as ants and termites. Rarely, the female is winged but the male not, as in fig wasps. In some cases, wings are produced only at particular times in the life cycle, such as in the dispersal phase of aphids. Wing structure and colouration often vary with morphs, such as in the aphids, migratory phases of locusts and polymorphic butterflies. At rest, the wings may be held flat, or folded a number of times along specific patterns; most typically, it is the hindwings which are folded, but in a few groups such as the vespid wasps, it is the forewings. The evolutionary origin of the insect wing is debated. During the 19th century, the question of insect wing evolution originally rested on two main positions. One position postulated insect wings evolved from pre-existing structures, while the second proposed insect wings were entirely novel formations. The “novel” hypothesis suggested that insect wings did not form from pre-existing ancestral appendages but rather as outgrowths from the insect body wall. Long since, research on insect wing origins has built on the “pre-existing structures” position that was originally proposed in the 19th century. Recent literature has pointed to several ancestral structures as being important to the origin of insect wings. Among these include: gills, respiratory appendages of legs, and lateral (paranotal) and posterolateral projections of the thorax to name a few. According to more current literature, possible candidates include gill-like structures, the paranotal lobe, and the crustacean tergal plate. The latter is based on recent insect genetic research which indicates that insects are pan-crustacean arthropods with a direct crustacean ancestor and shared genetic mechanisms of limb development. Other theories of the origin of insect wings are the paranotal lobe theory, the gill theory and the dual theory of insect wing evolution. These theories postulate that wings either developed from paranotal lobes, extensions of the thoracic terga; that they are modifications of movable abdominal gills as found on aquatic naiads of mayflies; or that insect wings arose from the fusion of pre-existing endite and exite structures each with pre-existing articulation and tracheation. Morphology Internal Each of the wings consists of a thin membrane supported by a system of veins. The membrane is formed by two layers of integument closely apposed, while the veins are formed where the two layers remain separate; sometimes the lower cuticle is thicker and more heavily sclerotized under a vein. Within each of the major veins there is a nerve and a trachea, and, since the cavities of the veins are connected with the hemocoel, hemolymph can flow into the wings. As the wing develops, the dorsal and ventral integumental layers become closely apposed over most of their area forming the wing membrane. The remaining areas form channels, the future veins, in which the nerves and tracheae may occur. The cuticle surrounding the veins becomes thickened and more heavily sclerotized to provide strength and rigidity to the wing. Two types of hair may occur on the wings: microtrichia, which are small and irregularly scattered, and macrotrichia, which are larger, socketed, and may be restricted to veins. The scales of Lepidoptera and Trichoptera are highly modified macrotrichia. Venation In some very small insects, the venation may be greatly reduced. In chalcidoid wasps, for instance, only the subcosta and part of the radius are present. Conversely, an increase in venation may occur by the branching of existing veins to produce accessory veins or by the development of additional, intercalary veins between the original ones, as in the wings of Orthoptera (grasshoppers and crickets). Large numbers of cross-veins are present in some insects, and they may form a reticulum as in the wings of Odonata (dragonflies and damselflies) and at the base of the forewings of Tettigonioidea and Acridoidea (katydids and grasshoppers respectively). The archedictyon is the name given to a hypothetical scheme of wing venation proposed for the very first winged insect. It is based on a combination of speculation and fossil data. Since all winged insects are believed to have evolved from a common ancestor, the archedictyon represents the "template" that has been modified (and streamlined) by natural selection for 200 million years. According to current dogma, the archedictyon contained 6–8 longitudinal veins. These veins (and their branches) are named according to a system devised by John Comstock and George Needham—the Comstock–Needham system: Costa (C) – the leading edge of the wing Subcosta (Sc) – second longitudinal vein (behind the costa), typically unbranched Radius (R) – third longitudinal vein, one to five branches reach the wing margin Media (M) – fourth longitudinal vein, one to four branches reach the wing margin Cubitus (Cu) – fifth longitudinal vein, one to three branches reach the wing margin Anal veins (A1, A2, A3) – unbranched veins behind the cubitus The costa (C) is the leading marginal vein on most insects. Sometimes, there is a small vein above the costa called the precosta, although in almost all extant insects, the precosta is fused with the costa. The costa rarely ever branches because it is at the leading edge, which is associated at its base with the humeral plate. The trachea of the costal vein is perhaps a branch of the subcostal trachea. Located after the costa is the third vein, the subcosta, which branches into two separate veins: the anterior and posterior. The base of the subcosta is associated with the distal end of the neck of the first axillary (see section below). The fourth vein is the radius (R), which is branched into five separate veins. The radius is generally the strongest vein of the wing. Toward the middle of the wing, it forks into a first undivided branch (R1) and a second branch, called the radial sector (Ra), which subdivides dichotomously into four distal branches (R2, R3, R4, R5). Basally, the radius is flexibly united with the anterior end of the second axillary (2Ax). The fifth vein of the wing is the media. In the archetype pattern (A), the media forks into two main branches: a media anterior (MA), which divides into two distal branches (MA1, MA2), and a median sector, or media posterior (MP), which has four terminal branches (M1, M2, M3, M4). In most modern insects the media anterior has been lost, and the usual "media" is the four-branched media posterior with the common basal stem. In the Ephemerida, according to present interpretations of the wing venation, both branches of the media are retained, while in Odonata the persisting media is the primitive anterior branch. The stem of the media is often united with the radius, but when it occurs as a distinct vein its base is associated with the distal median plate (m') or is continuously sclerotized with the latter. The cubitus, the sixth vein of the wing, is primarily two-branched. The primary forking of the takes place near the base of the wing, forming the two principal branches (Cu1, Cu2). The anterior branch may break up into a number of secondary branches, but commonly it forks into two distal branches. The second branch of the cubitus (Cu2) in Hymenoptera, Trichoptera, and Lepidoptera was mistaken by Comstock and Needham for the first anal. Proximally the main stem of the cubitus is associated with the distal median plate (m') of the wing base. Postcubitus (Pcu) is the first anal of the Comstock–Needham system. The postcubitus, however, has the status of an independent wing vein and should be recognized as such. In nymphal wings, its trachea arises between the cubital trachea and the group of vannal tracheae. In the mature wings of more generalized insect the Postcubitus is always associated proximally with the cubitus and is never intimately connected with the flexor sclerite (3Ax) of the wing base. In Neuroptera, Mecoptera, and Trichoptera the postcubitus may be more closely associated with the vannal veins, but its base is always free from the latter. The postcubitus is usually unbranched; it is primitively two branched. The vannal veins (lV to nV) are the anal veins that are immediately associated with the third axillary, and which are directly affected by the movement of this sclerite that brings about the flexion of the wings. In number the vannal veins vary. from 1 to 12, according to the expansion of the vannal area of the wing. The vannal tracheae usually arise from a common tracheal stem in nymphal insects, and the veins are regarded as branches of a single anal vein. Distally the vannal veins are either simple or branched. Jugal Veins (J) of the jugal lobe of the wing is often occupied by a network of irregular veins, or it may be entirely membranous; but sometimes it contains one or two distinct small veins, the first jugal vein, or vena arcuata, and the second jugal vein, or vena cardinalis (2J). C-Sc cross-veins – run between the costa and subcosta R cross-veins – run between adjacent branches of the radius R-M cross-veins – run between the radius and media M-Cu cross-veins – run between the media and cubitus All the veins of the wing are subject to secondary forking and to union by cross-veins. In some orders of insects the cross-veins are so numerous that the whole venational pattern becomes a close network of branching veins and cross-veins. Ordinarily, however, there is a definite number of cross-veins having specific locations. The more constant cross-veins are the humeral cross-vein (h) between costa and subcosta, the radial cross-vein (r) between R and the first fork of Rs, the sectorial cross-vein (s) between the two forks of R8, the median cross-vein (m–m) between M2 and M3, and the mediocubital cross-vein (m-cu) between media and cubitus. The veins of insect wings are characterized by a convex-concave placement, such as those seen in mayflies (i.e., concave is "down" and convex is "up") which alternate regularly and by its triadic type of branching; whenever a vein forks there is always an interpolated vein of the opposite position between the two branches. A concave vein will fork into two concave veins (with the interpolated vein being convex) and the regular alteration of the veins is preserved. The veins of the wing appear to fall into an undulating pattern according to whether they have a tendency to fold up or down when the wing is relaxed. The basal shafts of the veins are convex, but each vein forks distally into an anterior convex branch and a posterior concave branch. Thus the costa and subcosta are regarded as convex and concave branches of a primary first vein, Rs is the concave branch of the radius, posterior media the concave branch of the media, Cu1 and Cu2 are respectively convex and concave, while the primitive Postcubitus and the first vannal have each an anterior convex branch and a posterior concave branch. The convex or concave nature of the veins has been used as evidence in determining the identities of the persisting distal branches of the veins of modern insects, but it has not been demonstrated to be consistent for all wings. Fields Wing areas are delimited and subdivided by fold-lines along which the wing can fold, and flexion-lines along which the wing can flex during flight. The fundamental distinction between the flexion-lines and the fold-lines is often blurred, as fold-lines may permit some flexibility or vice versa. Two constants that are found in nearly all insect wings are the claval (a flexion-line) and jugal folds (or fold line); forming variable and unsatisfactory boundaries. Wing foldings can be very complicated, with transverse folding occurring in the hindwings of Dermaptera and Coleoptera, and in some insects the anal area can be folded like a fan. There are about four different fields found on the insect wings: Remigium Anal area (vannus) Jugal area Axillary area Alula Most veins and crossveins occur in the anterior area of the remigium, which is responsible for most of the flight, powered by the thoracic muscles. The posterior portion of the remigium is sometimes called the clavus; the two other posterior fields are the anal and jugal ares. When the vannal fold has the usual position anterior to the group of anal veins, the remigium contains the costal, subcostal, radial, medial, cubital, and postcubital veins. In the flexed wing the remigium turns posteriorly on the flexible basal connection of the radius with the second axillary, and the base of the mediocubital field is folded medially on the axillary region along the plica basalis (bf) between the median plates (m, m') of the wing base. The vannus is bordered by the vannal fold, which typically occurs between the postcubitus and the first vannal vein. In Orthoptera it usually has this position. In the forewing of Blattidae, however, the only fold in this part of the wing lies immediately before the postcubitus. In Plecoptera the vannal fold is posterior to the postcubitus, but proximally it crosses the base of the first vannal vein. In the cicada the vannal fold lies immediately behind the first vannal vein (lV). These small variations in the actual position of the vannal fold, however, do not affect the unity of action of the vannal veins, controlled by the flexor sclerite (3Ax), in the flexion of the wing. In the hindwings of most Orthoptera a secondary vena dividens forms a rib in the vannal fold. The vannus is usually triangular in shape, and its veins typically spread out from the third axillary like the ribs of a fan. Some of the vannal veins may be branched, and secondary veins may alternate with the primary veins. The vannal region is usually best developed in the hindwing, in which it may be enlarged to form a sustaining surface, as in Plecoptera and Orthoptera. The great fanlike expansions of the hindwings of Acrididae are clearly the vannal regions, since their veins are all supported on the third axillary sclerites on the wing bases, though Martynov (1925) ascribes most of the fan areas in Acrididae to the jugal regions of the wings. The true jugum of the acridid wing is represented only by the small membrane (Ju) mesad of the last vannal vein. The jugum is more highly developed in some other Polyneoptera, as in the Mantidae. In most of the higher insects with narrow wings the vannus becomes reduced, and the vannal fold is lost, but even in such cases the flexed wing may bend along a line between the postcubitus and the first vannal vein. The jugal region, or neala, is a region of the wing that is usually a small membranous area proximal to the base of the vannus strengthened by a few small, irregular veinlike thickenings; but when well developed it is a distinct section of the wing and may contain one or two jugal veins. When the jugal area of the forewing is developed as a free lobe, it projects beneath the humeral angle of the hindwing and thus serves to yoke the two wings together. In the Jugatae group of Lepidoptera it bears a long finger-like lobe. The jugal region was termed the neala ("new wing") because it is evidently a secondary and recently developed part of the wing. The axillary region is the region containing the axillary sclerites and has in general the form of a scalene triangle. The base of the triangle (a-b) is the hinge of the wing with the body; the apex (c) is the distal end of the third axillary sclerite; the longer side is anterior to the apex. The point d on the anterior side of the triangle marks the articulation of the radial vein with the second axillary sclerite. The line between d and c is the plica basalis (bf), or fold of the wing at the base of the mediocubital field. The termen is the outer margin of the wing, between apex and hind or anal angle. At the posterior angle of the wing base in some Diptera there is a pair of membranous lobes (squamae, or calypteres) known as the alula. The alula is well developed in the house fly. The outer squama (c) arises from the wing base behind the third axillary sclerite (3Ax) and evidently represents the jugal lobe of other insects (A, D); the larger inner squama (d) arises from the posterior scutellar margin of the tergum of the wing-bearing segment and forms a protective, hoodlike canopy over the haltere. In the flexed wing the outer squama of the alula is turned upside down above the inner squama, the latter not being affected by the movement of the wing. In many Diptera a deep incision of the anal area of the wing membrane behind the single vannal vein sets off a proximal alar lobe distal to the outer squama of the alula. Joints The various movements of the wings, especially in insects that flex the wings horizontally over the back when at rest, demand a more complicated articular structure at the wing base than a mere hinge of the wing with the body. Each wing is attached to the body by a membranous basal area, but the articular membrane contains a number of small articular sclerites, collectively known as the pteralia. The pteralia include an anterior humeral plate at the base of the costal vein, a group of axillaries (Ax) associated with the subcostal, radial, and vannal veins, and two less definite median plates (m, m') at the base of the mediocubital area. The axillaries are specifically developed only in the wing-flexing insects, where they constitute the flexor mechanism of the wing operated by the flexor muscle arising on the pleuron. Characteristic of the wing base is also a small lobe on the anterior margin of the articular area proximal to the humeral plate, which, in the forewing of some insects, is developed into a large, flat, scale-like flap, the tegula, overlapping the base of the wing. Posteriorly the articular membrane often forms an ample lobe between the wing and the body, and its margin is generally thickened and corrugated, giving the appearance of a ligament, the so-called axillary cord, continuous mesally with the posterior marginal scutellar fold of the tergal plate bearing the wing. The articular sclerites, or pteralia, of the wing base of the wing-flexing insects and their relations to the body and the wing veins, shown diagrammatically, are as follows: Humeral plates First Axillary Second Axillary Third Axillary Fourth Axillary Median plates (m, m') The humeral plate is usually a small sclerite on the anterior margin of the wing base, movable and articulated with the base of the costal vein. Odonata have their humeral plate greatly enlarged, with two muscles arising from the episternum inserted into the Humeral plates and two from the edge of the epimeron inserted into the axillary plate. The first axillary sclerite (lAx) is the anterior hinge plate of the wing base. Its anterior part is supported on the anterior notal wing process of the tergum (ANP); its posterior part articulates with the tergal margin. The anterior end of the sclerite is generally produced as a slender arm, the apex of which (e) is always associated with the base of the subcostal vein (Sc), though it is not united with the latter. The body of the sclerite articulates laterally with the second axillary. The second axillary sclerite (2Ax) is more variable in form than the first axillary, but its mechanical relations are no less definite. It is obliquely hinged to the outer margin of the body of the first axillary, and the radial vein (R) is always flexibly attached to its anterior end (d). The second axillary presents both a dorsal and a ventral sclerotization in the wing base; its ventral surface rests upon the fulcral wing process of the pleuron. The second axillary, therefore, is the pivotal sclerite of the wing base, and it specifically manipulates the radial vein. The third axillary sclerite (3Ax) lies in the posterior part of the articular region of the wing. Its form is highly variable and often irregular, but the third axillary is the sclerite on which is inserted the flexor muscle of the wing (D). Mesally it articulates anteriorly (f) with the posterior end of the second axillary, and posteriorly (b) with the posterior wing process of the tergum (PNP), or with a small fourth axillary when the latter is present. Distally the third axillary is prolonged in a process which is always associated with the bases of the group of veins in the anal region of the wing here termed the vannal veins (V). The third axillary, therefore, is usually the posterior hinge plate of the wing base and is the active sclerite of the flexor mechanism, which directly manipulates the vannal veins. The contraction of the flexor muscle (D) revolves the third axillary on its mesal articulations (b, f) and thereby lifts its distal arm; this movement produces the flexion of the wing. The Fourth Axillary sclerite is not a constant element of the wing base. When present it is usually a small plate intervening between the third axillary and the posterior notal wing process and is probably a detached piece of the latter. The median plates (m, m') are also sclerites that are not so definitely differentiated as specific plates as are the three principal axillaries, but nevertheless they are important elements of the flexor apparatus. They lie in the median area of the wing base distal to the second and third axillaries and are separated from each other by an oblique line (bf) which forms a prominent convex fold during flexion of the wing. The proximal plate (m) is usually attached to the distal arm of the third axillary and perhaps should be regarded as a part of the latter. The distal plate (m') is less constantly present as a distinct sclerite and may be represented by a general sclerotization of the base of the mediocubital field of the wing. When the veins of this region are distinct at their bases, they are associated with the outer median plate. Muscles The muscles that control flight in insects can take up to 10% to 30% of the total body mass. The muscles that control flight vary with the two types of flight found in insects: indirect and direct. Insects that use first, indirect, have the muscles attach to the tergum instead of the wings, as the name suggests. As the muscles contract, the thoracic box becomes distorted, transferring the energy to the wing. There are two "bundles" of muscles, those that span parallel to the tergum, the dorsolongitudinals, and those that are attached to the tegum and extend to the sternum, the dorsoventrals. In direct muscle, the connection is directly from the pleuron (thoracic wall) to individual sclerites located at the base of the wing. The subalar and basilar muscles have ligament attachments to the subalar and basilar sclerites. Here resilin, a highly elastic material, forms the ligaments connecting flight muscles to the wing apparatus. In more derived orders of insects, such as Diptera (flies) and Hymenoptera (wasp), the indirect muscles occupy the greatest volume of the pterothorax and function as the primary source of power for the wingstroke. Contraction of the dorsolongitudinal muscles causes the severe arching of the notum which depresses the wing while contraction of the dorsoventral muscles causes opposite motion of notum. The most primitive extant flying insects, Ephemeroptera (mayflies) and Odonata (dragonflies), use direct muscles that are responsible for developing the needed power for the up and down strokes. Insect wing muscle is a strictly aerobic tissue. Per unit protein it consumes fuel and oxygen at rates taking place in a very concentrated and highly organized tissue so that the steady-state rates per unit volume represent an absolute record in biology. The fuel and oxygen rich blood is carried to the muscles through diffusion occurring in large amounts, in order to maintain the high level of energy used during flight. Many wing muscles are large and may be as large as 10 mm in length and 2 mm in width. Moreover, in some Diptera the fibres are of giant dimensions. For instance, in the very active Rutilia, the cross-section is 1800 μm long and more than 500 μm wide. The transport of fuel and oxygen from the surroundings to the sites of consumption and the reverse transport of carbon dioxide therefore represent a challenge to the biologist both in relation to transport in the liquid phase and in the intricate system of air tubes, i.e. in the tracheal system. Sensors Several types of sensory neurons are found on insect wings: gustatory bristles, mechanosensory bristles, campaniform sensilla, and chordotonal organs. These sensors provide the nervous system with both external and internal proprioceptive feedback necessary for effective flight and grooming. Coupling, folding, and other features In many insect species, the forewing and hindwing can be coupled together, which improves the aerodynamic efficiency of flight by joining the forewing and hindwing into one bigger wing. The most common coupling mechanism (e.g., Hymenoptera and Trichoptera) is a row of small hooks on the forward margin of the hindwing, or "hamuli", which lock onto the forewing, keeping them held together (hamulate coupling). In some other insect species (e.g., Mecoptera, Lepidoptera, and some Trichoptera) the jugal lobe of the forewing covers a portion of the hindwing (jugal coupling), or the margins of the forewing and hindwing overlap broadly (amplexiform coupling), or the hindwing bristles, or frenulum, hook under the retaining structure or retinaculum on the forewing. When at rest, the wings are held over the back in most insects, which may involve longitudinal folding of the wing membrane and sometimes also transverse folding. Folding may sometimes occur along the flexion lines. Though fold lines may be transverse, as in the hindwings of beetles and earwigs, they are normally radial to the base of the wing, allowing adjacent sections of a wing to be folded over or under each other. The commonest fold line is the jugal fold, situated just behind the third anal vein, although, most Neoptera have a jugal fold just behind vein 3A on the forewings. It is sometimes also present on the hindwings. Where the anal area of the hindwing is large, as in Orthoptera and Blattodea, the whole of this part may be folded under the anterior part of the wing along a vannal fold a little posterior to the claval furrow. In addition, in Orthoptera and Blattodea, the anal area is folded like a fan along the veins, the anal veins being convex, at the crests of the folds, and the accessory veins concave. Whereas the claval furrow and jugal fold are probably homologous in different species, the vannal fold varies in position in different taxa. Folding is produced by a muscle arising on the pleuron and inserted into the third axillary sclerite in such a way that, when it contracts, the sclerite pivots about its points of articulation with the posterior notal process and the second axillary sclerite. As a result, the distal arm of the third axillary sclerite rotates upwards and inwards, so that finally its position is completely reversed. The anal veins are articulated with this sclerite in such a way that when it moves they are carried with it and become flexed over the back of the insect. Activity of the same muscle in flight affects the power output of the wing and so it is also important in flight control. In orthopteroid insects, the elasticity of the cuticle causes the vannal area of the wing to fold along the veins. Consequently, energy is expended in unfolding this region when the wings are moved to the flight position. In general, wing extension probably results from the contraction of muscles attached to the basilar sclerite or, in some insects, to the subalar sclerite. Flight Flight mechanisms Two groups of relatively large insects, the Ephemeroptera (mayflies) and the Odonata (dragonflies and damselflies) have the flight muscles attached directly to their wings; the wings can beat no faster than the rate at which nerves can send impulses to command the muscles to beat. All other living winged insects fly using a different mechanism, involving indirect flight muscles which cause the thorax to vibrate; the wings can beat faster than the rate at which the muscles receive nerve impulses. This mechanism evolved once, and is the defining feature (synapomorphy) for the infraclass Neoptera. Aerodynamics There are two basic aerodynamic models of insect flight. Most insects use a method that creates a spiralling leading edge vortex. Some very small insects use the fling and clap or Weis-Fogh mechanism in which the wings clap together above the insect's body and then fling apart. As they fling open, the air gets sucked in and creates a vortex over each wing. This bound vortex then moves across the wing and, in the clap, acts as the starting vortex for the other wing. Circulation and lift are increased, at the price of wear and tear on the wings. Many insects can hover by beating their wings rapidly, requiring sideways stabilization as well as lift. A few insects use gliding flight, without the use of thrust. Evolution Sometime in the Carboniferous Period, some 350 million years ago, when there were only two major land masses, insects began flying. How and why insect wings developed, however, is not well understood, largely due to the scarcity of appropriate fossils from the period of their development in the Lower Carboniferous. Three main theories on the origins of insect flight are that wings developed from paranotal lobes, extensions of the thoracic terga; that they are modifications of movable abdominal gills as found on aquatic naiads of mayflies; or that they developed from thoracic protrusions used as radiators. Fossils Fossils from the Devonian (400 million years ago) are all wingless, but by the Carboniferous (320 million years ago), more than 10 different genera of insects had fully functional wings. There is little preservation of transitional forms between the two periods. The earliest winged insects are from this time period (Pterygota), including the Blattoptera, Caloneurodea, primitive stem-group Ephemeropterans, Orthoptera and Palaeodictyopteroidea. Very early Blattopterans (during the Carboniferous) had a very large discoid pronotum and coriaceous forewings with a distinct CuP vein (an unbranched wing vein, lying near the claval fold and reaching the wing posterior margin). Even though the oldest possible insect fossil is the Devonian Rhyniognatha hirsti, estimated at 396–407 million years old, it possessed dicondylic mandibles, a feature associated with winged insects, although it is later considered as possible myriapod. During the Permian, the dragonflies (Odonata) were the dominant aerial predator and probably dominated terrestrial insect predation as well. True Odonata appeared in the Permian and all are amphibious. Their prototypes are the oldest winged fossils, go back to the Devonian, and are different from other wings in every way. Their prototypes may have had the beginnings of many modern attributes even by late Carboniferous and it is possible that they even captured small vertebrates, for some species had a wing span of 71 cm. The earliest beetle-like species during the Permian had pointed, leather like forewings with cells and pits. Hemiptera, or true bugs had appeared in the form of Arctiniscytina and Paraknightia having forewings with unusual venation, possibly diverging from Blattoptera. A single large wing from a species of Diptera in the Triassic (10 mm instead of usual 2–6 mm) was found in Australia (Mt. Crosby).This family Tilliardipteridae, despite the numerous 'tipuloid' features, should be included in Psychodomorpha sensu Hennig on account of loss of the convex distal 1A reaching wing margin and formation of the anal loop. Hypotheses Paranotal hypothesis: This hypothesis suggests that the insect's wings developed from paranotal lobes, a preadaptation found in insect fossils that is believed to have assisted stabilization while hopping or falling. In favor of this hypothesis is the tendency of most insects, when startled while climbing on branches, to escape by dropping to the ground. Such lobes would have served as parachutes and enable the insect to land more softly. The theory suggests that these lobes gradually grew larger and in a later stage developed a joint with the thorax. Even later would appear the muscles to move these crude wings. This model implies a progressive increase in the effectiveness of the wings, starting with parachuting, then gliding and finally active flight. Still, lack of substantial fossil evidence of the development of the wing joints and muscles poses a major difficulty to the theory, as does the seemingly spontaneous development of articulation and venation. Epicoxal hypothesis: This theory, first proposed in 1870 by Carl Gegenbaur, suggested that a possible origin for insect wings might have been the movable abdominal gills found in many aquatic insects, such as on naiads of mayflies. According to this theory these tracheal gills, which started their way as exits of the respiratory system and over time were modified into locomotive purposes, eventually developed into wings. The tracheal gills are equipped with little winglets that perpetually vibrate and have their own tiny straight muscles. Endite-exite hypothesis: This hypothesis stems from the adaptation of endites and exites, appendages on the respective inner and outer aspects of the primitive arthropod limb. It was advanced by Trueman based on a study by Goldschmidt in 1945 on Drosophila melanogaster, in which a pod variation displayed a mutation transforming normal wings to what was interpreted as a triple-jointed leg arrangement with some additional appendages but lacking the tarsus, where the wing's costal surface normally would be. This mutation was reinterpreted as strong evidence for a dorsal exite and endite fusion, rather than a leg, with the appendages fitting in much better with this hypothesis. The innervation, articulation and musculature required for the evolution of wings are already present in podomeres. Paranota plus leg gene recruitment hypothesis (also known as the dual origin hypothesis): The fossil larvae of Coxoplectoptera provided important new clues to the disputed question of the evolutionary origin of insect wings. Before the larvae fossil discovery the paranotal-hypothesis and the leg-exite-hypothesis have been considered as incompatible alternative explanations, which have both been supported by a set of evidences from the fossil record, comparative morphology, developmental biology and genetics. The expression of leg genes in the ontogeny of the insect wing has been universally considered as conclusive evidence in favour of the leg-exite-hypothesis, which proposes that insect wings are derived from mobile leg appendages (exites). However, the larvae of Coxoplectoptera show that the abdominal gills of mayflies and their ancestors, which are generally considered as corresponding structures to insect wings, articulated within the dorsal tergite plates. This cannot be seen in modern mayfly larvae, because their abdominal tergites and sternites are fused to rings, without any traces left even in embryonic development. If larval gills and wings are corresponding ("serial homologous") structures and thus share the same evolutionary origin, the new results from Coxoplectoptera demonstrate that also wings are of tergal origin, as proposed by the classical paranotal-hypothesis. Staniczek, Bechly & Godunko (2011) therefore suggested a new hypothesis that could reconcile the apparently conflicting evidence from paleontology and developmental genetics: wings first originated as stiff outgrowths of tergal plates (paranota), and only later in evolution became mobile, articulated appendages through secondary recruiting of leg genes. More recent fossil analysis of Paleozoic nymph wing pads provides additional support for the fusion of the paranota elements and arthopodan leg genes. Suggestions have been made that wings may have evolved initially for sailing on the surface of water as seen in some stoneflies. An alternative idea is that it derives from directed aerial gliding descent—a preflight phenomena found in some apterygote, a wingless sister taxa to the winged insects. The earliest fliers were similar to dragonflies with two sets of wings, direct flight muscles, and no ability to fold their wings over their abdomens. Most insects today, which evolved from those first fliers, have simplified to either one pair of wings or two pairs functioning as a single pair and using a system of indirect flight muscles. Natural selection has played an enormous role in refining the wings, control and sensory systems, and anything else that affects aerodynamics or kinematics. One noteworthy trait is wing twist. Most insect wings are twisted, as are helicopter blades, with a higher angle of attack at the base. The twist generally is between 10 and 20 degrees. In addition to this twist, the wing surfaces are not necessarily flat or featureless; most larger insects have wing membranes distorted and angled between the veins in such a way that the cross-section of the wings approximates an airfoil. Thus, the wing's basic shape already is capable of generating a small amount of lift at zero angle of attack. Most insects control their wings by adjusting tilt, stiffness, and flapping frequency of the wings with tiny muscles in the thorax (below). Some insects evolved other wing features that are not advantageous for flight, but play a role in something else, such as mating or protection. Some insects, occupying the biological niches that they do, need to be incredibly maneuverable. They must find their food in tight spaces and be capable of escaping larger predators – or they may themselves be predators, and need to capture prey. Their maneuverability, from an aerodynamic viewpoint, is provided by high lift and thrust forces. Typical insect fliers can attain lift forces up to three times their weight and horizontal thrust forces up to five times their weight. There are two substantially different insect flight mechanisms, and each has its own advantages and disadvantages – just because odonates have a more primitive flight mechanism does not mean they are less able fliers; they are, in certain ways, more agile than anything that has evolved afterward. Morphogenesis While the development of wings in insects is clearly defined in those who are members of Endopterygota, which undergo complete metamorphosis; in these species, the wing develops while in the pupal stage of the insects life cycle. However, insects that undergo incomplete metamorphosis do not have a pupal stage, therefore they must have a different wing morphogenesis. Insects such as those that are hemimetabolic have wings that start out as buds, which are found underneath the exoskeleton, and do not become exposed until the last instar of the nymph. The first indication of the wing buds is of a thickening of the hypodermis, which can be observed in insect species as early the embryo, and in the earliest stages of the life cycle. During the development of morphological features while in the embryo, or embryogenesis, a cluster of cells grow underneath the ectoderm which later in development, after the lateral ectoderm has grown dorsally to form wind imaginal disc. An example of wing bud development in the larvae, can be seen in those of White butterflies (Pieris). In the second instar the histoblast become more prominent, which now form a pocket-like structure. As of the third and fourth instars, the histoblast become more elongated. This greatly extended and evaginated, or protruding, part is what becomes the wing. By the close of the last instar, or fifth, the wing is pushed out of the wing-pocket, although continues to lie under the old larval cuticle while in its prepupal stage. It is not until the butterfly is in its pupal stage that the wing-bud becomes exposed, and shortly after eclosion, the wing begins to expand and form its definitive shape. The development of tracheation of the wings begin before the wing histoblast form, as it is important to note that they develop near a large trachea. During the fourth instar, cells from the epithelium of this trachea become greatly enlarged extend into the cavity of the wing bud, with each cell having developed a closely coiled tracheole. Each trachcole is of unicellular origin, and is at first intracellular in position; while tracheae are of multicellular origin and the lumen of each is intercellular in position. The development of tracheoles, each coiled within a single cell of the epithelium of a trachea, and the subsequent opening of communication between the tracheoles and the lumen of the trachea, and the uncoiling and stretching out of the tracheoles, so that they reach all parts of the wing. In the earlier stages of its development, the wing-bud is not provided with special organs of respiration such as tracheation, as it resembles in this respect the other portions of the hypodermis of which it is still a part. The histoblast is developed near a large trachea, a cross-section of which is shown in, which represents sections of these parts of the first, second, third and fourth instars respectively. At the same time the tracheoles uncoil, and extend in bundles in the forming vein-cavities of the wing-bud. At the molt that marks the beginning of the pupal stadium stage, they become functional. At the same time, the larval tracheoles degenerate; their function having been replaced by the wing tracheae. Nomenclature Most of the nomenclature of insect orders is based on the Ancient Greek word for wing, (), as the suffix -ptera. Adaptations Variation Insect wings are fundamental in identifying and classifying species as there is no other set of structures in studying insects more significant. Each order and insect family has distinctive wing shapes and features. In many cases, even species may be distinguished from each other by differences of color and pattern. For example, just by position one can identify species, albeit to a much lesser extent. Though most insects fold their wings when at rest, dragonflies and some damselflies rest with their wings spread out horizontally, while groups such as the caddisflies, stoneflies, alderflies, and lacewings hold their wings sloped roof-like over their backs. A few moths wrap their wings around their bodies, while many flies and most butterflies close their wings together straight upward over the back. Many times the shape of the wings correlates with the type of insect flight. The best-flying insects tend to have long, slender wings. In many species of Sphingidae (sphinx moths), the forewings are large and sharply pointed, forming with the small hindwings a triangle that is suggestive of the wings of fast, modern airplanes. Another, possibly more important correlation, is that of the size and power of the muscles to the speed and power of flight. In the powerfully flying insects, the wings are most adapted for the stresses and aerodynamics of flight. The veins are thicker, stronger, and closer together toward the front edge (or "leading edge") and thinner yet flexible toward the rear edge (or "trailing edge"). This makes the insect wing an excellently constructed airfoil, capable of exerting both propulsion and lift while minimizing drag. Variation of the wing beat may also occur, not just amongst different species, but even among individuals at different times. In general, the frequency is dependent upon the ratio between the power of the wing muscles and the resistance of the load. Large-winged, light-bodied butterflies may have a wing beat frequency of 4–20 per second whereas small-winged, heavy-bodied flies and bees beat their wings more than 100 times a second and mosquitoes can beat up to 988–1046 times a second. The same goes for flight; though it is generally difficult to estimate the speed of insects in flight, most insects can probably fly faster in nature than they do in controlled experiments. Coleoptera In species of Coleoptera (beetles), the only functional wings are the hindwings. The hindwings are longer than the elytra, folded longitudinally and transversely under the elytra. The wing is rotated forwards on its base into flight position. This action spread the wing and unfolded longitudinally and transversely. There is the spring mechanism in the wing structure, sometimes with the help of abdomen movement, to keep the wing in folded position. The beetle wing venation is reduced and modified due to the folding structure, which include: Costa (C), Subcosta posterior (ScP) – at the leading wing marginal, fused for most of the length. Radius anterior (RA) – divided into two branches beyond the middle of the wing. Radius posterior (RP) – basal connection is lost. Media posterior (MP) – branches, long and strong vein. Cubitus anterior (CuA) Anal veins (AA, AP) – veins behind the cubitus, separated by anal fold. In most species of beetles, the front pair of wings are modified and sclerotised (hardened) to form elytra and they protect the delicate hindwings which are folded beneath. The elytra are connected to the pterathorax; being called as such because it is where the wings are connected (pteron meaning "wing" in Greek). The elytra are not used for flight, but tend to cover the hind part of the body and protect the second pair of wings (alae). The elytra must be raised in order to move the hind flight wings. A beetle's flight wings are crossed with veins and are folded after landing, often along these veins, and are stored below the elytra. In some beetles, the ability to fly has been lost. These include some ground beetles (family Carabidae) and some "true weevils" (family Curculionidae), but also some desert and cave-dwelling species of other families. Many of these species have the two elytra fused together, forming a solid shield over the abdomen. In a few families, both the ability to fly and the elytra have been lost, with the best known example being the glow-worms of the family Phengodidae, in which the females are larviform throughout their lives. Lepidoptera The two pairs of wings are found on the middle and third segment, or mesothorax and metathorax respectively. In the more recent genera, the wings of the second segment are much more pronounced, however some more primitive forms have similarly sized wings of both segments. The wings are covered in scales arranged like shingles, forming the extraordinary variety seen in color. The mesothorax is evolved to have more powerful muscles to propel moth or butterfly through the air, with the wing of said segment having a stronger vein structure. The largest superfamily, Noctuidae, has the wings modified to act as tympanal or hearing organs Modifications in the wing's venation include: Costa (C) – not found in Butterflies. Subcosta (Sc) + Radius 1 (Sc+R1) – at the leading wing marginal, fused or very close for most of the length, in hindwing fused and well developed in the humeral area, subcosta never branches in butterfly. Radius (R2-R5) – radius divides into branches beyond the middle of the wing up to five branches in Papilionidae. On forewing, the last R is stalked in all butterflies except Hesperiidae is separated. Radius sector (Rs) – in hindwing. Media (M1-M3) – the basal section has been lost. Cubitus anterior (CuA1-CuA2) – CuP section has been lost. Anal veins (A, 1A+2A, 3A) – either one vein A, or two veins 1A+2A, 3A. Humeral vein – The hindwing of most butterflies has the humeral vein, except Lycaenidae There is the enlargement of the humeral area of the hindwing which is overlapped with the forewing. The humeral vein strengthened the hindwing overlapped area so that the two wings coupling better. The wings, head parts of thorax and abdomen of Lepidoptera are covered with minute scales, from which feature the order 'Lepidoptera' derives its names, the word "lepteron" in Ancient Greek meaning 'scale'. Most scales are lamellar, or blade-like and attached with a pedicel, while other forms may be hair-like or specialized as secondary sexual characteristics. The lumen or surface of the lamella, has a complex structure. It gives color either due to the pigmentary colors contained within or due to its three-dimensional structure. Scales provide a number of functions, which include insulation, thermoregulation, aiding gliding flight, amongst others, the most important of which is the large diversity of vivid or indistinct patterns they provide which help the organism protect itself by camouflage, mimicry, and to seek mates. Odonata Species of Odonata (Damselflies and dragonflies) both have two pairs of wings which are about equal in size and shape and are clear in color. There are five, if the R+M is counted as 1, main vein stems on dragonfly and damselfly wings, and wing veins are fused at their bases and the wings cannot be folded over the body at rest, which also include: Costa (C) – at the leading edge of the wing, strong and marginal, extends to the apex of the wing. Subcosta (Sc) – second longitudinal vein, it is unbranched, joins C at nodus. Radius and Media (R+M) – third and fourth longitudinal vein, the strongest vein on the wing, with branches, R1-R4, reach the wing margin, the media anterior (MA) are also reach the wing margin. IR2 and IR3 are intercalary veins behind R2 and R3 respectively. Cubitus (Cu) – fifth longitudinal vein, cubitus posterior (CuP) is unbranched and reach the wing margin. Anal veins (A1) – unbranched veins behind the cubitus. A nodus is formed where the second main vein meets the leading edge of the wing. The black pterostigma is carried near the wing tip. The main veins and the crossveins form the wing venation pattern. The venation patterns are different in different species. There may be very numerous crossveins or rather few. The Australian Flatwing Damselfly's wings are one of the few veins patterns. The venation pattern is useful for species identification. Almost all Anisoptera settle with the wings held out sideways or slightly downward, however most Zygoptera settle with the wings held together, dorsal surfaces apposed. The thorax of Zygoptera is so oblique that when held in this way the wings fit neatly along the top of the abdomen. They do not appear to be held straight up as in butterflies or mayflies. In a few zygopteran families the wings are held horizontally at rest, and in one anisopteran genus (e.g. Cordulephya, Corduliidae) the wings are held in the typical damselfly resting position. Adult species possess two pairs of equal or subequal wings. There appear to be only five main vein stems. A nodus is formed where the second main vein (subcosta) meets the leading edge of the wing. In most families a conspicuous pterostigma is carried near the wing tip. Identification as Odonata can be based on the venation. The only likely confusion is with some lacewings (order Neuroptera) which have many crossveins in the wings. Until the early years of the 20th century Odonata were often regarded as being related to lacewings and were given the ordinal name Paraneuroptera, but any resemblance between these two orders is entirely superficial. In Anisoptera the hindwing is broader than the forewing and in both wings a crossvein divides the discoidal cell into a Triangle and Supertriangle. Orthoptera Species of Orthoptera (grasshoppers and crickets) have forewings that are tough opaque tegmina, narrow which are normally covering the hindwings and abdomen at rest. The hindwings are board membranous and folded in fan-like manner, which include the following venation: Costa (C) – at the leading marginal of the forewing and hindwing, unbranched. Subcosta (Sc) – second longitudinal vein, unbranched. Radius (R) – third longitudinal vein, branched to Rs in forewing and hindwing. Media anterior (MA) – fourth longitudinal vein, branched in basal part as Media posterior (MP). Cubitus (Cu) – fifth longitudinal vein, on forewing and hindwing dividing near the wing base into branched CuA, and unbranched CuP. Anal veins (A) – veins behind the cubitus, unbranched, two in forewing, many in hindwing. Phasmatodea Costa (C) – at the leading marginal of the hindwing, unbranched, absent in forewing. Subcosta (Sc) – second longitudinal vein, unbranched. Radius (R) – third longitudinal vein, branched to Rs in hindwing, unbranched in forewing. Media anterior (MA) – fourth longitudinal vein, branched in basal part as Media posterior (MP). Cubitus (Cu) – fifth longitudinal vein, unbranched. Anal veins (A) – veins behind the cubitus, unbranched, two in forewing, many in hindwing 1A-7A in one group and the rest in another group. Stick insect have forewings that are tough, opaque tegmina, short and covering only the base part of the hindwings at rest. Hindwings from costa to Cubitus are tough and opaque like the forewings. The large anal area are membranous and folded in fan-like manner. There are no or very few branching in stick insect wing veins. Dermaptera Other orders such as the Dermaptera (earwigs), Orthoptera (grasshoppers, crickets), Mantodea (praying mantis) and Blattodea (cockroaches) have rigid leathery forewings that are not flapped while flying, sometimes called tegmen (pl. tegmina), elytron (pl. elytra), or pseudoelytron. Hemiptera In Hemiptera (true bugs), the forewings may be hardened, though to a lesser extent than in the beetles. For example, the anterior part of the front wings of stink bugs is hardened, while the posterior part is membranous. They are called hemelytron (pl. hemelytra). They are only found in the suborder Heteroptera; the wings of the Homoptera, such as the cicada, are typically entirely membranous. Both forewings and hindwings of Cicada are membranous. Most species are glass-like although some are opaque. Cicadas are not good fliers and most fly only a few seconds. When flying, forewing and hindwing are hooked together by a grooved coupling along the hindwing costa and forewing margin. Most species have a basic venation as shown in the following picture. Costa (C) – at the leading wing marginal, in forewing extends to the node and lies close to Sc+R. Subcosta + Radius (Sc+R) – in forewing Sc and R fused together to the node. Radial sector (Rs) arises near the node and unbranches. Radius anterior (RA) Radius posterior (RP) Media (M) – branches to M1 to M4. Cubitus anterior (CuA) – branches to CuA1 and CuA2. Cubitus posterior (CuP) – unbranches. Anal veins (A) – veins behind the cubitus, 1A and 2A fused in the forewing, CuP and 2A are folded. Also notice there are the ambient veins and peripheral membranes on the margin of both wings. Diptera In the Diptera (true flies), there is only one pair of functional wings, with the posterior pair of wings are reduced to halteres, which help the fly to sense its orientation and movement, as well as to improve balance by acting similar to gyroscopes. In Calyptratae, the very hindmost portion of the wings are modified into somewhat thickened flaps called calypters which cover the halteres. Costa (C) – not found in Diptera. Subcosta (Sc) – became the leading wing vein, unbranched. Radius (R) – branched to R1-R5. Media (M) – branched to M1-M4. Cubitus anterior(CuA)- unbranched, CuP is reduced in Diptera. Some species CuA and 1A are separated, some species meets when reaching the wing margin, some species fused. Anal veins (A) – only two anal veins 1A and 2A are present, 2A is not distinctive in some species. Discal Cell (dc) – well defined in most species. Blattodea Species of Blattodea (cockroaches) have a forewing, are also known as tegmen, that is more or less sclerotized. It is used in flight as well as a form of protection of the membranous hindwings. The veins of hindwing are about the same as front wing but with large anal lobe folded at rest between CuP and 1A. The anal lobe usually folded in a fan-like manner. Costa (C) – at the leading edge of the wing. Subcosta (Sc) – second longitudinal vein, it is relatively short. Radius (R) – third longitudinal vein, with many pectinate branches. Media (M) – fourth longitudinal vein, reach the wing margin. Cubitus anterior (CuA) – fifth longitudinal vein, with dichotomous branches occupy large part of tegmen. Cubitus posterior (CuP) – is unbranched, curved and reach the wing margin. Anal veins (A) – veins behind the cubitus. Hymenoptera Hymenoptera adults, including sawflies, wasps, bees, and non-worker ants, all have two pairs of membranous wings. Costa (C) – not found in Hymenoptera. Subcosta (Sc) – unbranched. Radius (R) – branched to R1-R5. Media (M) – M is unbranched, in forewing M is fused with Rs for part of its length. Cubitus (CuA) – unbranched, CuP is absent in Hymenoptera. Anal veins (A) – only two anal veins 1A and 2A are present, 2A is not distinctive in some species. Wing-coupling – Row of hooks on the leading edge of hindwing engage the hind margin of the forewing, strongly couple the wings in flight. Line of wing folding – Some species, including Vespidae, the forewing are longitudinally folded along the 'line of wing folding' at rest. Pterostigma – is present for some species. The forward margin of the hindwing bears a number of hooked bristles, or "hamuli", which lock onto the forewing, keeping them held together. The smaller species may have only two or three hamuli on each side, but the largest wasps may have a considerable number, keeping the wings gripped together especially tightly. Hymenopteran wings have relatively few veins compared with many other insects, especially in the smaller species. Other families Termites are relatively poor fliers and are readily blown downwind in wind speeds of less than 2 km/h, shedding their wings soon after landing at an acceptable site, where they mate and attempt to form a nest in damp timber or earth. Wings of most termites have three heavy veins along the basal part of the front edge of the forewing and the crossveins near the wing tip are angled, making trapezoidal cells. Although subterranean termite wings have just two major veins along the front edge of the forewing and the cross veins towards the wingtip are perpendicular to these veins, making square and rectangular cells. Species of Thysanoptera (thrips), Ptiliidae and other flying microinsects have slender front and hindwings with long fringes of hair, called fringed wings, also referred to as ptiloptery. While species of Trichoptera (caddisfly) have hairy wings with the front and hindwings clothed with setae. Male Strepsiptera also have halteres that evolved from the forewings instead of the hindwings. This means that only their hindwings are functional at flying, as opposed to Diptera which have functional forewings and halteres for hindwings. Also the hindwings in males of Coccidae are reduced to halteres (or are absent).
Biology and health sciences
External anatomy and regions of the body
Biology
27265528
https://en.wikipedia.org/wiki/Microplastics
Microplastics
Microplastics are fragments of any type of plastic less than in length, according to the U.S. National Oceanic and Atmospheric Administration (NOAA) and the European Chemicals Agency. US EPA researchers define microplastics, or MPs, as plastic particles ranging in size from 5 millimeters (mm), which is about the size of a pencil eraser, to 1 nanometer (nm). For comparison, a strand of human hair is about 80,000 nanometers wide. They cause pollution by entering natural ecosystems from a variety of sources, including cosmetics, clothing, construction, renovation, food packaging, and industrial processes. The term microplastics is used to differentiate from larger, non-microscopic plastic waste. Two classifications of microplastics are currently recognized. Primary microplastics include any plastic fragments or particles that are already 5.0 mm in size or less before entering the environment. These include microfibers from clothing, microbeads, plastic glitter and plastic pellets (also known as nurdles). Secondary microplastics arise from the degradation (breakdown) of larger plastic products through natural weathering processes after entering the environment. Such sources of secondary microplastics include water and soda bottles, fishing nets, plastic bags, microwave containers, tea bags and tire wear. Both types are recognized to persist in the environment at high levels, particularly in aquatic and marine ecosystems, where they cause water pollution. 35% of all ocean microplastics come from textiles/clothing, primarily due to the erosion of polyester, acrylic, or nylon-based clothing, often during the washing process. Microplastics also accumulate in the air and terrestrial ecosystems. Airborne microplastics have been detected in the atmosphere, as well as indoors and outdoors. Because plastics degrade slowly (often over hundreds to thousands of years), microplastics have a high probability of ingestion, incorporation into, and accumulation in the bodies and tissues of many organisms. The toxic chemicals that come from both the ocean and runoff can also biomagnify up the food chain. In terrestrial ecosystems, microplastics have been demonstrated to reduce the viability of soil ecosystems. As of 2023, the cycle and movement of microplastics in the environment was not fully known. Microplastics in surface sample ocean surveys might have been underestimated as deep layer ocean sediment surveys in China found that plastics are present in deposition layers far older than the invention of plastics. Microplastics are likely to degrade into smaller nanoplastics through chemical weathering processes, mechanical breakdown, and even through the digestive processes of animals. Nanoplastics, or NPs, are a subset of microplastics and they are smaller than 1 µm (1 micrometer or 1000 nm). Nanoplastics cannot be seen by the human eye. Classification The term "microplastics" was introduced in 2004 by Professor Richard Thompson, a marine biologist at the University of Plymouth in the United Kingdom. Microplastics are common in our world today. In 2014, it was estimated that there are between 15 and 51 trillion individual pieces of microplastic in the world's oceans, which was estimated to weigh between 93,000 and 236,000 metric tons. Under the influence of sunlight, wind, waves and other factors, plastic degrades into small fragments known as microplastics, or even nanoplastics. Primary microplastics Primary microplastics are small pieces of plastic that are purposefully manufactured. They are usually used in facial cleansers and cosmetics, or in air blasting technology. In some cases, their use in medicine as vectors for drugs was reported. Microplastic "scrubbers", used in exfoliating hand cleansers and facial scrubs, have replaced traditionally used natural ingredients, including ground almond shells, oatmeal, and pumice. Primary microplastics have also been produced for use in air-blasting technology. This process involves blasting acrylic, melamine, or polyester microplastic scrubbers at machinery, engines, and boat hulls to remove rust and paint. As these scrubbers are used repeatedly until they diminish in size and their cutting power is lost, they often become contaminated with heavy metals such as cadmium, chromium, and lead. Although many companies have committed to reducing the production of microbeads, there are still many bioplastic microbeads that also have a long degradation life cycle, for example in cosmetics. Secondary microplastics Secondary microplastics are small pieces of plastic derived from the physical breakdown and mechanical degradation of larger plastic debris, both at sea and on land. Over time, a culmination of physical, biological, and photochemical degradation, including photo-oxidation caused by sunlight exposure, can reduce the structural integrity of plastic debris to a size that is eventually undetectable to the naked eye. This process of breaking down large plastic material into much smaller pieces is known as fragmentation. It is considered that microplastics might further degrade to be smaller in size, although the smallest microplastic reportedly detected in the oceans in 2017 was 1.6 micrometres (6.3×10−5 in) in diameter. The prevalence of microplastics with uneven shapes suggests that fragmentation is a key source. One study suggested that more microplastics might be formed from biodegradable polymer than from non-biodegradable polymer in both seawater and fresh water. "It's actually classified as a very high priority high contaminant by the EPA... when they litter or put something in a landfill, the plastic will break down into smaller and smaller particles. And eventually, they become microplastics... They're in the air, they're in the water, they're in the soil." University of Tennessee professor Mike McKinney. Microplastic fibers enter the environment as a by-product during wear and tear and from the washing of synthetic clothing. Tires, composed partly of synthetic styrene-butadiene rubber, erode into tiny plastic and rubber particles as they are used and become dust particles. 2.0-5.0 mm plastic pellets, used to create other plastic products, enter ecosystems due to spillages and other accidents. A 2015 Norwegian Environment Agency review report about microplastics stated it would be beneficial to classify these sources as primary, as long as microplastics from these sources are added from human society since the "start of the pipe", and their emissions are inherently a result of human material and product use and not secondary to fragmentation in the nature Nanoplastics Depending on the definition used, nanoplastics are less than 1 μm (i.e. 1000 nm) or less than 100 nm in size. Speculations over nanoplastics in the environment range from it being a temporary byproduct during the fragmentation of microplastics to it being an invisible environmental threat at potentially high and continuously rising concentrations. The presence of nanoplastics in the North Atlantic Subtropical Gyre has been confirmed and recent developments in Raman spectroscopy coupled with optical tweezers (Raman Tweezers) as well as nano-fourier-transform infrared spectroscopy (nano-FTIR) or atomic force infrared (AFM-IR) are promising answers in the near future regarding the nanoplastic quantity in the environment. Fluorescence could represent a unique tool for the identification and quantification of nanoplastics, since it allows the development of fast, easy, cheap, and sensitive methods. However, the nanoplastic problem is complex and nanoscale properties as well as interaction with biomolecules need to be explored at the fundamental level with high spatial and temporal resolution. Nanoplastics are thought to be a risk to environmental and human health. Due to their small size, nanoplastics can cross cellular membranes and affect the functioning of cells. Nanoplastics are lipophilic and models show that polyethylene nanoplastics can be incorporated into the hydrophobic core of lipid bilayers. Nanoplastics are also shown to cross the epithelial membrane of fish accumulating in various organs including the gallbladder, pancreas, and the brain. Nanoplastics are believed to cause interruptions in bone cell activities, causing improper bone formation. Little is known on adverse health effects of nanoplastics in organisms including humans. In zebrafish (Danio rerio), polystyrene nanoplastics can induce a stress response pathway altering glucose and cortisol levels, which is potentially tied to behavioral changes in stress phases. In Daphnia, polystyrene nanoplastic can be ingested by the freshwater cladoceran Daphnia pulex and affect its growth and reproduction as well as induce stress defense, including the ROS production and MAPK-HIF-1/NF-κB-mediated antioxidant system. Nanoplastics can also adsorb toxic chemical pollutants, such as antibiotics, which enable the selective association with antibiotic-resistant bacteria, resulting in the dissemination of nanoplastics and antibiotic-resistant bacteria by bacterivorous nematode Caenorhabditis elegans across the soil. Sources A big portion of microplastic pollution comes from textiles, tires and city dust which account for over 80% of all microplastic in the seas and the environment. Microplastic is also a type of airborne particulates and is found to prevail in air. Paint appears as the largest source of microplastic leakage into the ocean and waterways (1.9 Mt/year), outweighing all other sources of microplastic leakage (e.g. textiles fibres and tyre dust). The existence of microplastics in the environment is often established through aquatic studies. These include taking plankton samples, analyzing sandy and muddy sediments, observing vertebrate and invertebrate consumption, and evaluating chemical pollutant interactions. Through such methods, it has been shown that there are microplastics from multiple sources in the environment. Microplastics could contribute up to 30% of the Great Pacific Garbage Patch polluting the world's oceans and, in many developed countries, are a bigger source of marine plastic pollution than the visible larger pieces of marine litter, according to a 2017 IUCN report. Microplastics floating in the world's oceans is a common source of heavy metals. The colours of plastics are commonly produced by the inclusion of heavy metals (or their compounds) such as chromium, manganese, cobalt, copper, zinc, zirconium, molybdenum, silver, tin, praseodymium, neodymium, erbium, tungsten, iridium, gold, lead, or uranium. Clothing Studies have shown that many synthetic fibers, such as polyester, nylon, acrylics, and spandex, can be shed from clothing and persist in the environment. Each garment in a load of laundry can shed more than 1,900 fibers of microplastics, with fleeces releasing the highest percentage of fibers, over 170% more than other garments. For an average wash load of , over 700,000 fibers could be released per wash. Washing machine manufacturers have also reviewed research into whether washing machine filters can reduce the amount of microfiber fibers that need to be treated by sewage treatment facilities. These microfibers have been found to persist throughout the food chain from zooplankton to larger animals such as whales. The primary fiber that persists throughout the textile industry is polyester which is a cheap cotton alternative that can be easily manufactured. However, these types of fibers contribute greatly to the persistence to microplastics in terrestrial, aerial, and marine ecosystems. The process of washing clothes causes garments to lose an average of over 100 fibers per liter of water. This has been linked with health effects possibly caused by the release of monomers, dispersive dyes, mordants, and plasticizers from manufacturing. The occurrence of these types of fibers in households has been shown to represent 33% of all fibers in indoor environments. Textile fibers have been studied in both indoor and outdoor environments to determine the average human exposure. The indoor concentration was found to be 1.0–60.0 fibers/m3, whereas the outdoor concentration was much lower at 0.3–1.5 fibers/m3. The deposition rate indoors was 1586–11,130 fibers per day/m3 which accumulates to around 190-670 fibers/mg of dust. The largest concern with these concentrations is that it increases exposure to children and the elderly, which can cause adverse health effects. Containers and packaging Plastic containers can shed microplastics and nanoparticles into foods and beverages. Bottled water In one study, 93% of the bottled water from 11 different brands showed microplastic contamination. Per liter, researchers found an average of 325 microplastic particles. Of the tested brands, Nestlé Pure Life and Gerolsteiner bottles contained the most microplastic with 930 and 807 microplastic particles per liter (MPP/L), respectively. San Pellegrino products showed the least quantity of microplastic densities. Compared to water from taps, water from plastic bottles contained twice as much microplastic. Another study capable of detecting nanoplastics found 240,000 fragments per liter: 10% between 5 mm and 1 μm and 90% under 1 μm in diameter. Some of the contamination likely comes from the process of bottling and packaging the water, and possibly from filters used to purify the water. Baby bottles In 2020 researchers reported that polypropylene infant feeding bottles with contemporary preparation procedures were found to cause microplastics exposure to infants ranging from 14,600 to 4,550,000 particles per capita per day in 48 regions. Microplastics release is higher with warmer liquids and similar with other polypropylene products such as lunchboxes. Unexpectedly, silicone rubber baby bottle nipples degrade over time from repeated steam sterilization, shedding micro- and nano-sized particles of silicone rubber, researchers found in 2021. They estimated that, using such heat-degraded nipples for a year, a baby will ingest more than 660,000 particles. Single-use plastic products Common single-use plastic products, such as plastic cups, or even paper coffee cups that are lined with a thin plastic film inside, release trillions of microplastic-nanoparticles per liter into water during normal use. Single-use plastic products enter aquatic environments and "[l]ocal and statewide policies that reduce single-use plastics were identified as effective legislative actions that communities can take to address plastic pollution". Construction and renovation Plastics are extensively used in the construction and renovation industry. Airborne microplastic dust is produced during renovation, building, bridge and road reconstruction projects and the use of power tools. Materials containing polyvinyl chloride (PVC), polycarbonate, polypropylene, and acrylic, can degrade overtime releasing microplastics. During the construction process single use plastic containers and wrappers are discarded adding to plastic waste. These plastics are difficult to recycle and end up in landfills where they break down over a long period of time causing potential leaching into the soil and the release of airborne microplastics. Airborne microplastic dust is also generated by deterioration of building materials Due to the environmental impact from plastic waste creation in the construction and renovation sectors waste management practices that address this issue are required. Although many researchers have investigated the use of wastes, such as plastic, in the construction process in an effort to reduce waste and increase sustainability, construction is not an environmentally-friendly activity by nature. Efforts have been made to reduce plastic waste by adding it to concrete as agglomerates. However, one solution for resolving the problem from the large amount of plastic wastes generated could bring another serious problem of leaching of microplastics. The unknown part of this area is huge and needs prompt investigation. Around twenty percent of all plastics and seventy percent of all polyvinyl chloride (PVC) produced in the world each year are used by the construction industry. It is predicted that much more will be produced and used in the future. "In Europe, approximately 20% of all plastics produced are used in the construction sector including different classes of plastics, waste and nanomaterials." Common types: Polyvinyl chloride (PVC) Polyethylene (PE) Polypropylene (PP) Expandable polystyrene (EPS) Polyurethane (PU) Indirect use (packaging of construction materials) examples: Foils and moisture barriers Covers Soft plastic wraps EPS and PP sacks Direct use (construction materials containing plastics) examples: Building products Insulation Damp-proofing Flooring Roofing Windows Laminated surfaces Building service installations Pipes Cabling Surface treatments Paints Varnishes Sealants Glues Resins Covers Shrink wrap Tarpaulins Cosmetics industry Some companies have replaced natural exfoliating ingredients with microplastics, usually in the form of "microbeads" or "micro-exfoliates". These products are typically composed of polyethylene, a common component of plastics, but they can also be manufactured from polypropylene, polyethylene terephthalate (PET), and nylon. They are often found in face washes, hand soaps, and other personal care products; the beads are usually washed into the sewage system immediately after use. Their small size prevents them from fully being retained by preliminary treatment screens at wastewater plants, thereby allowing some to enter rivers and oceans. Wastewater treatment plants only remove an average of 95–99.9% of microbeads because of their small design. This leaves an average of 0–7 microbeads per litre being discharged. Considering that the treatment plants of the world discharge 160 trillion liters of water per day, around 8 trillion microbeads are released into waterways every day. This number does not account for the sewage sludge that is reused as fertilizer after the waste water treatment that has been known to still contain these microbeads. Although many companies have committed to phasing out the use of microbeads in their products, there are at least 80 different facial scrub products that are still being sold with microbeads as a main component. This contributes to the 80 metric tons of microbead discharge per year by the United Kingdom alone, which not only has a negative impact upon the wildlife and food chain, but also upon levels of toxicity, as microbeads have been proven to absorb dangerous chemicals such as pesticides and polycyclic aromatic hydrocarbons. The restriction proposal by the European Chemicals Agency (ECHA) and reports by the United Nations Environment Programme (UNEP) and TAUW suggest that there are more than 500 microplastic ingredients that are widely used in cosmetics and personal care products. Even when microbeads are removed from cosmetic products, there are still harmful products being sold with plastics in them. For example, acrylate copolymers cause toxic effects for waterways and animals if they are polluted. Acrylate copolymers also can emit styrene monomers when used in body products which increases a person's chances of cancer. Countries like New Zealand which have banned microbeads often pass over other polymers such as acrylate copolymers, which can be just as toxic to people and the environment. After the Microbead-Free Waters Act of 2015, the use of microbeads in toothpaste and other rinse-off cosmetic products has been discontinued in the US, however since 2015 many industries have instead shifted toward using FDA-approved "rinse-off" metallized-plastic glitter as their primary abrasive agent. Fishing industry Recreational and commercial fishing, marine vessels, and marine industries are all sources of plastic that can directly enter the marine environment, posing a risk to biota both as macroplastics, and as secondary microplastics following long-term degradation. Marine debris observed on beaches also arises from beaching of materials carried on inshore and ocean currents. Fishing gear is a form of plastic debris with a marine source. Discarded or lost fishing gear, including plastic monofilament line and nylon netting (sometimes called ghost nets), is typically neutrally buoyant and can, therefore, drift at variable depths within the oceans. Various countries have reported that microplastics from the industry and other sources have been accumulating in different types of seafood. In Indonesia, 55% of all fish species had evidence of manufactured debris similar to America which reported 67%. However, the majority of debris in Indonesia was plastic, while in North America the majority was synthetic fibers found in clothing and some types of nets. The implication from the fact that fish are being contaminated with microplastic is that those plastics and their chemicals will bioaccumulate in the food chain. One study analyzed the plastic-derived chemical called polybrominated diphenyl ethers (PBDEs) in the stomachs of short-tailed shearwaters. It found that one-fourth of the birds had higher-brominated congeners that are not naturally found in their prey. However, the PBDE got into the birds' systems through plastic that was found in the stomachs of the birds. It is therefore not just the plastics that are being transferred through the food chain but the chemicals from the plastics as well. Manufacturing The manufacture of plastic products uses granules and small resin pellets as their raw material. In the United States, production increased from 2.9 million pellets in 1960 to 21.7 million pellets in 1987. In 2019, plastic world production was 368 million tonnes; 51% were produced in Asia. China, the world's largest producer, created 31% of the world total. Through accidental spillage during land or sea transport, inappropriate use as packing materials, and direct outflow from processing plants, these raw materials can enter aquatic ecosystems. In an assessment of Swedish waters using an 80 μm mesh, KIMO Sweden found typical microplastic concentrations of 150–2,400 microplastics per m3; in a harbor adjacent to a plastic production facility, the concentration was 102,000 per m3. Many industrial sites in which convenient raw plastics are frequently used are located near bodies of water. If spilled during production, these materials may enter the surrounding environment, polluting waterways. "More recently, Operation Cleansweep, a joint initiative of the American Chemistry Council and Society of the Plastics Industry, is aiming for industries to commit to zero pellet loss during their operations". Overall, there is a significant lack of research aimed at specific industries and companies that contribute to microplastics pollution. Personal protective equipment Since the emergence of the COVID-19 pandemic, the usage of medical face masks has sharply increased to reach approximately 89 million masks each. Single use face masks are made from polymers, such as polypropylene, polyurethane, polyacrylonitrile, polystyrene, polycarbonate, polyethylene, or polyester. The increase in production, consumption, and littering of face masks was added to the list of environmental challenges, due to the addition of plastic particles waste in the environment. After degrading, disposable face masks could break down into smaller size particles (under 5mm) emerging a new source of microplastic. A single surgical weathered face mask may release up to 173,000 fibers/ day. A report made in February 2020 by Oceans Asia, an organization committed to advocacy and research on marine pollution, confirms "the presence of face masks of different types and colors in an ocean in Hong Kong". Sewage treatment plants Sewage treatment plants, also known as wastewater treatment plants (WWTPs), remove contaminants from wastewater, primarily from household sewage, using various physical, chemical, and biological processes. Most plants in developed countries have both primary and secondary treatment stages. In the primary stage of treatment, physical processes are employed to remove oils, sand, and other large solids using conventional filters, clarifiers, and settling tanks. Secondary treatment uses biological processes involving bacteria and protozoa to break down organic matter. Common secondary technologies are activated sludge systems, trickling filters, and constructed wetlands. The optional tertiary treatment stage may include processes for nutrient removal (nitrogen and phosphorus) and disinfection. Microplastics have been detected in both the primary and secondary treatment stages of the plants. A groundbreaking 1998 study suggested that microplastic fibers would be a persistent indicator of sewage sludges and wastewater treatment plant outfalls. A study estimated that about one particle per liter of microplastics are being released back into the environment, with a removal efficiency of about 99.9%. A 2016 study showed that most microplastics are actually removed during the primary treatment stage where solid skimming and sludge settling are used. When these treatment facilities are functioning properly, the contribution of microplastics into oceans and surface water environments from WWTPs is not disproportionately large. Many studies show that while wastewater treatment plants certainly reduce the microplastic load on waterways, with current technological developments they are not able to clean the waters fully of this pollutant. Sewage sludge is used for soil fertilizer in some countries, which exposes plastics in the sludge to the weather, sunlight, and other biological factors, causing fragmentation. As a result, microplastics from these biosolids often end up in storm drains and eventually into bodies of water. In addition, some studies show that microplastics do pass through filtration processes at some WWTPs. According to a study from the UK, samples taken from sewage sludge disposal sites on the coasts of six continents contained an average one particle of microplastic per liter. A significant amount of these particles was of clothing fibers from washing machine effluent. Transportation Car and truck tires Wear and tear from tires significantly contributes to the flow of (micro-)plastics into the environment. Estimates of emissions of microplastics to the environment in Denmark are between per year. Secondary microplastics (e.g. from car and truck tires or footwear) are more important than primary microplastics by two orders of magnitude. The formation of microplastics from the degradation of larger plastics in the environment is not accounted for in the study. The estimated per capita emission ranges from 0.23 to 4.7 kg/year, with a global average of 0.81 kg/year. The emissions from car tires (wear reaching 100%) are substantially higher than those of other sources of microplastics, e.g., airplane tires (2%), artificial turf (wear 12–50%), brakes (wear 8%), and road markings (wear 5%). In the case of road markings, recent field study indicated that they were protected by a layer of glass beads and their contribution was only between 0.1 and 4.3 g/person/year, which would constitute approximately 0.7% of all of the secondary microplastics emissions; this value agrees with some emissions estimates. Emissions and pathways depend on local factors like road type or sewage systems. The relative contribution of tire wear and tear to the total global amount of plastics ending up in our oceans is estimated to be 5–10%. In air, 3–7% of the particulate matter (PM2.5) is estimated to consist of tire wear and tear, indicating that it may contribute to the global health burden of air pollution which has been projected by the World Health Organization at 3 million deaths in 2012. Pollution from tire wear and tear also enters the food chain, but further research is needed to assess human health risks. Shipping Shipping has significantly contributed to marine pollution. Some statistics indicate that in 1970, commercial shipping fleets around the world dumped over 23,000 tons of plastic waste into the marine environment. In 1988, an international agreement (MARPOL 73/78, Annex V) prohibited the dumping of waste from ships into the marine environment. In the United States, the Marine Plastic Pollution Research and Control Act of 1987 prohibits discharge of plastics in the sea, including from naval vessels. However, shipping remains a dominant source of plastic pollution, having contributed around 6.5 million tons of plastic in the early 1990s. Research has shown that approximately 10% of the plastic found on the beaches in Hawaii are nurdles. In one incident on 24 July 2012, 150 tonnes of nurdles and other raw plastic material spilled from a shipping vessel off the coast near Hong Kong after a major storm. This waste from the Chinese company Sinopec was reported to have piled up in large quantities on beaches. While this is a large incident of spillage, researchers speculate that smaller accidents also occur and further contribute to marine microplastic pollution. Exposure pathways Air Airborne microplastics have been detected in the atmosphere, as well as indoors and outdoors. Microplastic can be atmospherically transported to remote areas by the wind. A 2017 study found indoor airborne microfiber concentrations between 1.0 and 60.0 microfibers per cubic meter (33% of which were found to be microplastics). Another study looked at microplastic in the street dust of Tehran and found 2,649 particles of microplastic within 10 samples of street dust, with ranging samples concentrations from 83 particle – 605 particles (±10) per 30.0 g of street dust. Microplastics and microfibers were also found in snow samples, and high up in "clean" air in high mountains at vast distances from their source. Much like freshwater ecosystems and soil, more studies are needed to understand the full impact and significance of airborne microplastics. Water Oceans Ice cores Plastic pollution has previously been recorded in Antarctic surface waters and sediments as well as in the Arctic sea ice, but in 2009, for the first time, plastic was found in Antarctic sea ice, with 96 microplastic particles from 14 different types of polymers in an ice core sampled from east Antarctica. Relatively large particle sizes in Antarctic sea ice suggest local pollution sources. Freshwater Microplastics have been widely detected in the world's aquatic environments. The first study on microplastics in freshwater ecosystems was published in 2011 that found an average of 37.8 fragments per square meter of Lake Huron sediment samples. Additionally, studies have found MP (microplastic) to be present in all of the Great Lakes with an average concentration of 43,000 MP particle km−2. Microplastics have also been detected in freshwater ecosystems outside of the United States, for example in 2019 study conducted in Poland showed that microplastic was present in all 30 studied lakes of the Masurian Lakeland with density from 0.27 to 1.57 particles per liter. In Canada, a three-year study found a mean microplastic concentration of 193,420 particles km−2 in Lake Winnipeg. None of the microplastics detected were micro-pellets or beads and most were fibers resulting from the breakdown of larger particles, synthetic textiles, or atmospheric fallout. The highest concentration of microplastic ever discovered in a studied freshwater ecosystem was recorded in the Rhine river at 4000 MP particles kg−1. Soil A substantial portion of microplastics are expected to end up in the world's soil, yet very little research has been conducted on microplastics in soil outside of aquatic environments. In wetland environments microplastic concentrations have been found to exhibit a negative correlation with vegetation cover and stem density. There exists some speculation that fibrous secondary microplastics from washing machines could end up in soil through the failure of water treatment plants to completely filter out all of the microplastic fibers. Furthermore, geophagous soil fauna, such as earthworms, mites, and collembolans could contribute to the amount of secondary microplastic present in soil by converting consumed plastic debris into microplastic via digestive processes. Further research, however, is needed. There is concrete data linking the use of organic waste materials to synthetic fibers being found in the soil; but most studies on plastics in soil merely report its presence and do not mention origin or quantity. Controlled studies on fiber-containing land-applied wastewater sludges (biosolids) applied to soil reported semiquantitative recoveries of the fibers a number of years after application. Salt and seafood A 2015 review of 15 brands of table salts commercially available in China found microplastics were much more prevalent in sea salts compared to lake, rock, or well salts, attributing this to sea salts being contaminated by ocean water pollution while the rock/well salts were more likely contaminated during the production stages of collecting, wind drying, and packaging. According to a 2017 estimate, a person who consumes seafood will ingest 11,000 bits of microplastics per year. A 2019 study found a kilo of sugar had 440 microplastic particles, a kilo of salt contained 110 particles, and a litre of bottled water contained 94 particles. Composition The composition of microplastics are complex. A study in 2023 tested some fish species and found that “about 80% of the MPs detected were fibrous in shape and were made of polyethylene (25%), polyester (20%), and polyamide (10%). Most microplastic particles observed were black (61%) or blue (27%) in color.” Effects on the environment In 2008, an International Research Workshop at the University of Washington at Tacoma concluded that microplastics were a problem in the marine environment, based on their documented occurrence, the long residence times of these particles, their likely buildup in the future, and their demonstrated ingestion by marine organisms. According to a comprehensive review of scientific evidence published by the European Union's Scientific Advice Mechanism in 2019, microplastics were present in every part of the environment. While there was no evidence of widespread ecological risk from microplastic pollution yet, risks were likely to become widespread within a century if pollution continued at its current rate. As of 2020 microplastics had been detected in freshwater systems including marshes, streams, ponds, lakes, and rivers in Europe, North America, South America, Asia, and Australia. Samples collected across 29 Great Lakes tributaries from six states in the United States were found to contain plastic particles, 98% of which were microplastics ranging in size from 0.355mm to 4.75mm. Likewise, they have been found in high mountains, at great distances from their source. Deep layer ocean sediment surveys in China (2020) show the presence of plastics in deposition layers far older than the invention of plastics, leading to suspected underestimation of microplastics in surface sample ocean surveys. In September 2021 Hurricane Larry deposited, during the storm peak, 113,000 particles/m2/day as it passed over Newfoundland, Canada. Back-trajectory modelling and polymer type analysis indicated that those microplastics may have been ocean-sourced as the hurricane traversed the North Atlantic garbage patch of the North Atlantic Gyre. As of 2023 there was rapid growth of microplastic pollution research, with marine and estuarine environments most frequently studied. Researchers have called for better sharing of research data that might lead to effective solutions. Consequences of plastic degradation and pollution release over long term have mostly been overlooked. The large amounts of plastic in the environment, exposed to degradation, with years of decay and release of toxic compounds to follow was referred to as toxicity debt. Marine and freshwater organisms Microplastics are inconspicuous, being less than 5 mm. Particles of this size are available to every species, enter the food chain at the bottom, and become embedded in animal tissue. Micro- and nanoplastics can become embedded in animals' tissue through ingestion or respiration. The initial demonstration of bioaccumulation of these particles in animals was conducted under controlled conditions by exposing them to high concentrations of microplastics over extended periods, accumulating these particles in their gut and gills due to ingestion and respiration, respectively. Various annelid species, such as deposit-feeding lugworms (Arenicola marina), have been shown to accumulate microplastics embedded in their gastrointestinal tract. Similarly, many crustaceans, like the shore crab Carcinus maenas, have been seen to integrate microplastics into both their respiratory and digestive tracts. Plastic particles are often mistaken by fish for food, which can block their digestive tracts, sending incorrect feeding signals to the brains of the animals. However, research in 2021 revealed that fish ingest microplastics inadvertently rather than intentionally. The first occurrence of bioaccumulation of micro and nanoplastics in wild animals was documented in the skin mucosa of salmon, and it was attributed to the resemblance between nanoplastics and the outer shell of the viruses that the mucosa traps. This discovery was entirely serendipitous, as the research team had developed a detailed molecular separation process for the components of fish skin with the primary objective of isolating chitin from a vertebrate for the first time. A study done at the Argentinean coastline of the Rio de la Plata estuary, found the presence of microplastics in the guts of 11 species of coastal freshwater fish. These 11 species of fish represented four different feeding habits: detritivore, planktivore, omnivore and ichthyophagous. This study is one of the few so far to show the ingestion of microplastics by freshwater organisms. It can take up to 14 days for microplastics to pass through an animal (as compared to a normal digestion period of 2 days), but enmeshment of the particles in animals' gills can prevent elimination entirely. When microplastic-laden animals are consumed by predators, the microplastics are then incorporated into the bodies of higher trophic-level feeders. For example, scientists have reported plastic accumulation in the stomachs of lantern fish which are small filter feeders and are the main prey for commercial fish like tuna and swordfish. Microplastics also absorb chemical pollutants that can be transferred into the organism's tissues. Small animals are at risk of reduced food intake due to false satiation and resulting starvation or other physical harm from the microplastics. Zooplankton ingest microplastics beads (1.7–30.6 μm) and excrete fecal matter contaminated with microplastics. Along with ingestion, the microplastics stick to the appendages and exoskeleton of the zooplankton. Zooplankton, among other marine organisms, consume microplastics because they emit similar infochemicals, notably dimethyl sulfide, just as phytoplankton do. Plastics such as high-density polyethylene (HDPE), low-density polyethylene (LDPE), and polypropylene (PP) produce dimethyl sulfide odors. These types of plastics are commonly found in plastic bags, food storage containers, and bottle caps. Green and red filaments of plastics are found in the planktonic organisms and in seaweeds. Bottom feeders, such as benthic sea cucumbers, who are non-selective scavengers that feed on debris on the ocean floor, ingest large amounts of sediment. It has been shown that four species of sea cucumber (Thyonella gemmate, Holothuria floridana, H. grisea and Cucumaria frondosa) ingested between 2- and 20-fold more PVC fragments and between 2- and 138-fold more nylon line fragments (as much as 517 fibers per organism) based on plastic-to-sand grain ratios from each sediment treatment. These results suggest that individuals may be selectively ingesting plastic particles. This contradicts the accepted indiscriminate feeding strategy of sea cucumbers, and may occur in all presumed non-selective feeders when presented with microplastics. Bivalves, important aquatic filter feeders, have also been shown to ingest microplastics and nanoplastics. Upon exposure to microplastics, bivalve filtration ability decreases. Multiple cascading effects occur as a result, such as immunotoxicity and neurotoxicity. Decreased immune function occurs due to reduced phagocytosis and NF-κB gene activity. Impaired neurological function is a result of the inhibition of ChE and suppression of neurotransmitter regulatory enzymes. When exposed to microplastics, bivalves also experience oxidative stress, indicating an impaired ability to detoxify compounds within the body, which can ultimately damage DNA. Bivalve gametes and larvae are also impaired when exposed to microplastics. Rates of developmental arrest, and developmental malformities increase, while rates of fertilization decrease. When bivalves have been exposed to microplastics as well as other pollutants such as POPs, mercury or hydrocarbons in lab settings, toxic effects were shown to be aggravated. Not only fish and free-living organisms can ingest microplastics. Some corals such as Pocillopora verrucosa have also been found to ingest microplastics. Scleractinian corals, which are primary reef-builders, have been shown to ingest microplastics under laboratory conditions. While the effects of ingestion on these corals has not been studied, corals can easily become stressed and bleach. Microplastics have been shown to stick to the exterior of the corals after exposure in the laboratory. The adherence to the outside of corals can potentially be harmful, because corals cannot handle sediment or any particulate matter on their exterior and slough it off by secreting mucus, expending energy in the process, increasing the likelihood of mortality. Marine biologists in 2017 discovered that three-quarters of the underwater seagrass in the Turneffe Atoll off the coast of Belize had microplastic fibers, shards, and beads stuck to it. The plastic pieces had been overgrown by epibionts (organisms that naturally stick themselves to seagrass). Seagrass is part of the barrier reef ecosystem and is fed on by parrotfish, which in turn are eaten by humans. These findings, published in Marine Pollution Bulletin, may be "the first discovery of microplastics on aquatic vascular plants... [and] only the second discovery of microplastics on marine plant life anywhere in the world." Research published in 2023 demonstrated that microplastic exposure impaired the cognitive performance of hermit crabs, which could potentially impact their survivability. Microbes, soil ecosystems and terrestrial plants Microplastics can affect the soil ecosystem and stunt the growth of terrestrial plants due to the increased uptake of toxic metals such as cadmium. Microplastics can reduce weight of earthworms. Microbes also live on the surface of microplastics, and can form a biofilm which, according to a 2019 study, has a unique structure and possesses a special risk, because microplastic biofilms have been proven to provide a novel habitat for colonization that increases overlap between different species, thus spreading pathogens and antibiotic resistant genes through horizontal gene transfer. Then, due to rapid movement through waterways, these pathogens can be moved from their origin to another location where a specific pathogen may not be naturally present, spreading potential disease. There is concern microplastic pollutants may act as a vector for antibiotic resistant genes and bacteria. Clinically important bacterial genus like Eggerthella were more than three times enriched on riverine microplastics compared to water. Animals In 2019, the first European records of microplastic items in amphibians' stomach content was reported in specimens of the common European newt (Triturus carnifex). This also represented the first evidence for Caudata worldwide, highlighting that the emerging issue of plastics is a threat even in remote high-altitude environments. The microplastic has also been found in common blackbirds (Turdus merula) and song thrushes (Turdus philomelos) which shows a ubiquity of microplastics in terrestrial environments. In 2023, plasticosis, a new disease caused solely by plastics, was discovered in seabirds who had scarred digestive tracts from ingesting plastic waste. "When birds ingest small pieces of plastic, [...]it inflames the digestive tract. Over time, the persistent inflammation causes tissues to become scarred and disfigured, affecting digestion, growth and survival." Persistent organic pollutants and emerging organic contaminants Plastic particles may highly concentrate and transport synthetic organic compounds (e.g. persistent organic pollutants and emerging organic contaminants), commonly present in the environment and ambient seawater, on their surface through adsorption. Microplastics can act as carriers for the transfer of POPs from the environment to organisms, also termed as the Trojan Horse effect. Recent articles have also shown that microplastics can sorb emerging organic chemicals such as pharmaceuticals and personal care products. The sorption potential is affected by water matrix, pH, ionic strength and aging of microparticles. Additives added to plastics during manufacture may leach out upon ingestion, potentially causing serious harm to the organism. Endocrine disruption by plastic additives may affect the reproductive health of humans and wildlife alike. Geophysics Microplastics can increase the stability of breaking waves or sea foam, potentially affecting sea albedo or atmosphere-ocean gas exchange. Microplastics in the ocean may re-enter the atmosphere via sea spray. Human health In 2009, mean/median intake of microplastics in humans were at levels considered by some to be safe in humans; however, some individuals would sometimes exceed these limits; the effects of this, if any, were unknown. As of 2010, the degree of absorption and retention from microplastics exposure from air, water, and food that humans ate as the end of the food chain was unclear. Up to 2018 it was unknown whether and to what degree microplastics bioaccumulated in humans. According to a 2019 comprehensive review of scientific evidence published by the European Union's Scientific Advice Mechanism, little was known with respect to the human health risks of nano- and microplastics, and what was known was surrounded by considerable uncertainty. The main limitations were the quality or methodology of the research to date. The review concluded that "there is a need to understand the potential modes of toxicity for different size-shape-type NMP [nano- (< 0.1 mm) and microplastic] combinations in carefully selected human models, before robust conclusions about 'real' human risks can be made". Also in 2019, scientists estimated annual microplastic consumption was about 39,000 to 52,000 plastic particles for the average person. This varied among different ages and gender. In 2020, it was suggested that ingestion of microplastics via food might be relatively minor, with humans predicted to be exposed to more microplastics in household dust than by consuming mussels. As of 2022 and 2023, the quantities of microplastic entering the human body from the environment were still not well understood, as was the potential risks of microplastic to human health; the field is difficult to research because of the potentially long time between exposure to the contaminant and any associated health effect becoming evident. Microplastic pollution has been associated with various adverse human health conditions, including respiratory disease and inflammation, but it was not known whether this was a causative effect. Prevention Dust control Some of the suggested dust control measures include “lining cutting areas with tarps, cutting inside a protective tent, and using vacuum bags on power tool” when cutting materials like Trex and Azek. The cost of these measures is low.” Street sweeping may also inhibited the spread of pollutants by gathering significant amounts of dirty materials from the extensive construction, renovation and reconstruction projects of road tunnels, bridges, roads and buildings. Treatment Some researchers have proposed incinerating plastics to use as energy, which is known as energy recovery. As opposed to losing the energy from plastics into the atmosphere in landfills, this process turns some of the plastics back into energy that can be used. However, as opposed to recycling, this method does not diminish the amount of plastic material that is produced. Therefore, recycling plastics is considered a more efficient solution. Biodegradation is another possible solution to large amounts of microplastic waste. In this process, microorganisms consume and decompose synthetic polymers by means of enzymes. These plastics can then be used in the form of energy and as a source of carbon once broken down. The microbes could potentially be used to treat sewage wastewater, which would decrease the amount of microplastics that pass through into the surrounding environments. Filtering Efficient removal of microplastics via waste water treatment plants is critical to prevent the transfer of microplastics from society to natural water systems. The captured microplastics in the treatment plants become part of the sludge produced by the plants. The problem is that this sludge is often used as farm fertilizer meaning the plastics enter waterways through runoff. Fionn Ferreira, winner of the 2019 Google Science Fair, is developing a device for the removal of microplastic particles from water using a ferrofluid. Collection devices Computer modelling done by The Ocean Cleanup, a Dutch foundation, has suggested that collection devices placed nearer to the coasts could remove about 31% of the microplastics in the area. On 9 September 2018, The Ocean Cleanup launched the world's first ocean cleanup system, 001 aka "Wilson", which is being deployed to the Great Pacific Garbage Patch. System 001 is 600 meters long that acts as a U-shaped skiff that uses natural oceanic currents to concentrate plastic and other debris on the ocean's surface into a confined area for extraction by vessels. The project has been met with criticism from oceanographers and plastic pollution experts, though it has seen wide public support. In addition, some bacteria have adapted to eat plastic, and some bacteria species have been genetically modified to eat (certain types of) plastics. Other than degrading microplastics, microbes had been engineered in a novel way to capture microplastics in their biofilm matrix from polluted samples for easier removal of such pollutants. The microplastics in the biofilms can then be released with an engineered 'release' mechanism via biofilm dispersal to facilitate with microplastics recovery. Absorption devices include sponges made of cotton and squid bones, which may be scalable for water remediation projects. Education and recycling Increasing education through recycling campaigns is another proposed solution for microplastic contamination. While this would be a smaller-scale solution, education has been shown to reduce littering, especially in urban environments where there are often large concentrations of plastic waste. If recycling efforts are increased, a cycle of plastic use and reuse would be created to decrease our waste output and production of new raw materials. In order to achieve this, states would need to employ stronger infrastructure and investment around recycling. Some advocate for improving recycling technology to be able to recycle smaller plastics to reduce the need for production of new plastics. In April 2013, Italian artist Maria Cristina Finucci founded The Garbage Patch State in order to create awareness, under the patronage of UNESCO and the Italian Ministry of the Environment. In February 2013 the U.S. Environmental Protection Agency (EPA) launched its "Trash-Free Waters" initiative to prevent single-use plastic wastes from ending up in waterways and ultimately the ocean. As of 2018, EPA collaborated with the United Nations Environment Programme–Caribbean Environment Programme (UNEP-CEP) and the Peace Corps to reduce and remove trash in the Caribbean Sea. EPA also funded various projects in the San Francisco Bay Area including one that is aimed at reducing the use of single-use plastics such as disposable cups, spoons and straws, from three University of California campuses. The Florida Microplastic Awareness Project (FMAP), a group of volunteers who search for microplastics in coastal water samples Many organizations advocate action to counter microplastic, spreading microplastic awareness. Global advocacy aimed at achieving the target of the United Nations Sustainable Development Goal 14 hopes to prevent and significantly reduce all forms of marine pollution by 2025. Funding The Clean Oceans Initiative is a project launched in 2018 by the public institutions European Investment Bank, Agence Française de Développement and KfW Entwicklungsbank. Their goal was to provide up to €2 billion in lending, grants and technical assistance until 2023 to develop projects that removed pollution from waterways (with a focus on macroplastics and microplastics) before it reached the oceans. The effort focuses on initiatives that demonstrate efficient methods of minimising plastic waste and microplastics output, emphasising on riverine and coastal areas. Cassa Depositi e Prestiti (CDP), the Italian national promotional institution and financial institution for development cooperation, and the Instituto de Crédito Oficial (ICO), the Spanish promotional bank, became new partners in October 2020. As of December 2023, The Clean Oceans Initiative had funded almost €3.2 billion, exceeding 80% of its €4 billion objective. Over 20 million people were supposed to benefit from the signed project proposals, which include better wastewater treatment in Sri Lanka, China, Egypt, and South Africa, solid waste management in Togo and Senegal, and stormwater management and flood protection in Benin, Morocco, and Ecuador. In February 2022, the initiative stated that it would increase its financing aim to €4 billion by the end of 2025. At the same time, the European Bank for Reconstruction and Development (EBRD) became the Clean Oceans Initiative's sixth member. By February 2023, the program had met 65% of its goal, with €2.6 billion spent in 60 projects benefiting more than 20 million people across Africa, Asia, Latin America, and Europe. By the beginning of 2022, more than 80% of this target was achieved, with €1.6 billion being used in long-term financing for public and private sector initiatives that minimise the discharge of plastics, microplastics, and other pollutants through enhanced solid waste, wastewater, and storm water management. In January 2021, the European Investment Bank and the Asian Development Bank had formed the Clean and Sustainable Ocean Partnership to promote cooperative projects for a clean and sustainable ocean and blue economy in the Asia-Pacific region. Policy and legislation With increasing awareness of the detrimental effects of microplastics on the environment, groups are now advocating for the removal and ban of microplastics from various products. One such campaign is "Beat the Microbead", which focuses on removing plastics from personal care products. The Adventurers and Scientists for Conservation run the Global Microplastics Initiative, a project to collect water samples to provide scientists with better data about microplastic dispersion in the environment. UNESCO has sponsored research and global assessment programs due to the trans-boundary issue that microplastic pollution constitutes. These environmental groups will keep pressuring companies to remove plastics from their products in order to maintain healthy ecosystems. China In 2018, China banned the import of recyclables from other countries, forcing those other countries to re-examine their recycling schemes. The Yangtze River in China contributes 55% of all plastic waste going to the seas. Including microplastics, the Yangtze bears an average of 500,000 pieces of plastic per square kilometer. Scientific American reported that China dumps 30% of all plastics in the ocean. Hong Kong In 2024, the Hong Kong government implemented the first phase of its plastic restriction regulation. Promotional videos have also been produced to encourage citizens to bring their own utensils when dining out, to refrain from using disposable utensils, and to bring their own shopping bags when shopping. Merchants are prohibited from providing related plastic products to customers. United States In the US, some states have taken action to mitigate the negative environmental effects of microplastics. Illinois was the first US state to ban cosmetics containing microplastics. At the federal level, the Microbead-Free Waters Act 2015 was enacted after being signed by President Barack Obama on 28 December 2015. The law bans "rinse-off" cosmetic products that perform an exfoliating function, such as toothpaste or face wash. It does not apply to other products such as household cleaners. The act took effect on 1 July 2017, with respect to manufacturing, and 1 July 2018, with respect to introduction or delivery for introduction into interstate commerce. On 16 June 2020, California adopted a definition of 'microplastics in drinking water', setting the foundation for a long-term approach to studying their contamination and human health effects. On 25 July 2018, a microplastic reduction amendment was passed by the U.S. House of Representatives. The legislation, as part of the Save Our Seas Act designed to combat marine pollution, aims to support the NOAA's Marine Debris Program. In particular, the amendment is geared towards promoting NOAA's Great Lakes Land-Based Marine Debris Action Plan to increase testing, cleanup, and education around plastic pollution in the Great Lakes. President Donald Trump signed the re-authorization and amendment bill into effect on 11 October 2018. Japan On 15 June 2018, the Japanese government passed a bill with the goal of reducing microplastic production and pollution, especially in aquatic environments. Proposed by the Environment Ministry and passed unanimously by the Upper House, this is also the first bill to pass in Japan that is specifically targeted at reducing microplastic production, specifically in the personal care industry with products such as face wash and toothpaste. This law is revised from previous legislation, which focused on removing plastic marine debris. It also focuses on increasing education and public awareness surrounding recycling and plastic waste. The Environment Ministry has also proposed a number of recommendations for methods to monitor microplastic quantities in the ocean (Recommendations, 2018). However, the legislation does not specify any penalties for those who continue manufacturing products with microplastics. European Union The European Commission has noted the increased concern about the impact of microplastics on the environment. In April 2018, the European Commission's Group of Chief Scientific Advisors commissioned a comprehensive review of the scientific evidence on microplastic pollution through the EU's Scientific Advice Mechanism. The evidence review was conducted by a working group nominated by European academies and delivered in January 2019. A Scientific Opinion based on the SAPEA report was presented to the Commission in 2019, on the basis of which the commission will consider whether policy changes should be proposed at a European level to curb microplastic pollution. In January 2019, the European Chemicals Agency (ECHA) proposed to restrict intentionally added microplastics. The European Union participates with 10% of the global total, around 150 000 tonnes of microplastics each year. This is 200 grams per person per year, with significant regional variance in per-capita microplastic creation. The European Commission's Circular Economy Action Plan sets out mandatory requirements for the recycling and waste reduction of key products e.g. plastic packaging. The plan starts the process to restrict addition of microplastics in products. It mandates measures for capturing more microplastics at all stages of the lifecycle of a product. E.g. the plan would examine different policies which aim to reduce release of secondary microplastics from tires and textiles. The European Commission plans to update the Urban Waste Water Treatment Directive to further address microplastic waste and other pollution. They aim to protect the environment from industrial and urban waste water discharge. A revision to the EU Drinking Water Directive was provisionally approved to ensure microplastics are regularly monitored in drinking water. It would require countries must propose solutions if a problem is found. The REACH restriction on synthetic polymer microparticles entered into force on 17 October 2023. United Kingdom The Environmental Protection (Microbeads) (England) Regulations 2017 ban the production of any rinse-off personal care products (such as exfoliants) containing microbeads. This particular law denotes specific penalties when it is not obeyed. Those who do not comply are required to pay a fine. In the event that a fine is not paid, product manufacturers may receive a stop notice, which prevents the manufacturer from continuing production until they have followed regulation preventing the use of microbeads. Criminal proceedings may occur if the stop notice is ignored. Haiti Haiti has no collective system for waste collection and treatment, and thus plastic is often disposed of in urban water evacuation canals, which then degrade to form microplastics. Due to tropical temperatures and the plastics present in urban waterways could degrade more rapidly. Their discharge into Port-au-Prince Bay exposes this ecosystem to a number of environmental hazards pollutants contained in the waste, and to climatic hazards, particularly ocean acidification. On August 9, 2012, the Haitian government published a decree prohibiting the production, importation, marketing and use, of polyethylene bags and expanded polystyrene objects for foodstuffs. However, 14 Caribbean countries (more than a third) have banned single-use plastic bags and/or polystyrene containers. On July 10, 2013, a second decree was published to once again prohibit "the importation, production or sale of expanded polystyrene articles for food use." In support of the second decree, the ministries of the Environment, Justice and Public Security, Trade and Industry as well as the Economy and Finance announced in a note published in January 2018 that specialists from the brigade will be deployed on the territory to force the application of the said decree.
Technology
Materials
null
8416361
https://en.wikipedia.org/wiki/Sanitary%20engineering
Sanitary engineering
Sanitary engineering, also known as public health engineering or wastewater engineering, is the application of engineering methods to improve sanitation of human communities, primarily by providing the removal and disposal of human waste, and in addition to the supply of safe potable water. Traditionally a branch of civil engineering and now a subset of environmental engineering, in the mid-19th century, the discipline concentrated on the reduction of disease, then thought to be caused by miasma. This was accomplished mainly by the collection and segregation of sewerage flow in London specifically, and Great Britain generally. These and later regulatory improvements were reported in the United States as early as 1865. It is also concerned with environmental factors that do not have an immediate and clearly understood effect on public health. Areas outside the purview of sanitary engineering include aesthetic concerns such as landscaping, and environmental conservation as it pertains to plants and animals. Skills within this field are usually employed for the primary goal of disease prevention within human beings by assuring a supply of healthy drinking water, treatment of waste water, and removal of garbage from inhabited areas. Compared to (for example) electrical engineering or mechanical engineering which are concerned primarily with closed systems, sanitary engineering is a very interdisciplinary field which may involve such elements as plumbing, fire protection, hydraulics, life safety, constructive modelling, information technology, project design, microbiology, pathology and the many divisions within environmental science and environmental technology. In some cases, considerations that fall within the field of social sciences and urban planning must be factored in as well. Although sanitary engineering may be most associated with the design of sewers, sewage treatment and wastewater treatment facilities, recycling centers, public landfills and other things which are constructed, the term applies equally to a plan of action to reverse the effects of water pollution or soil contamination in a specific area. History Irrigation systems were invented five to seven thousand years ago as a means of supplying water to agriculture-based societies. Aqueducts and irrigation systems were among the first forms of wastewater engineering. As population centers became more dense, they were used to remove sewage from settlements. The Romans were among the first to demonstrate the effectiveness of the aqueduct. The Dark Ages marked a period where progress in water management came to a halt. As populations grew, the management of human waste became a growing concern and a public health threat. By the 1850s in London, more than 400,000 tons of sewage were flushed into the River Thames each day - around 150 million tons per year. Diseases such as smallpox, diphtheria, measles, scarlet fever, typhus, cholera, and typhoid were spread via the contaminated water supply. During the 19th century, major cities started building sewage systems to remove human waste out of cities and into rivers. Sanitation in the 1900's During the 1900s, the activated sludge process was invented. The activated sludge process is a form of water purification that uses bacteria to consume human feces. Chlorine is used later in the process to kill off the bacteria. In the 1950s, the public health reports provided plans for supplying clean water for the public by first looking at potential hazards. The organization looked carefully at water contamination as well as how drinking water was being treated. They also prioritized finding methods that were effective, yet not too costly. Sanitation cost is the main issue for many foreign (not the United States) countries. The average cost of home water and sanitation systems start at $50 a month, when many citizens don't make enough money to use on non-necessities. Over the centuries, much has changed in the field of wastewater engineering. Advancements in microbiology, chemistry, and engineering have drastically changed the field. Today, wastewater engineers also work on the collection of clean water for drinking, chemically treating it, and using UV light to kill off micro-organisms. They also treat water pollution in wastewater (blackwater and greywater) so that this water may be made safe for use without endangering the population and environment around it. Wastewater treatment and water reclamation are areas of concern in this field. Harm Huizenga Prior to modern forms of sanitation in neighborhoods and cities, people would simply leave their trash on the street. In 1892, it was such an issue, that a man named Harm Huizenga volunteered to clean up the mess by himself. The Dutch man went around the streets in his wagon, picking up the garbage of the city of Chicago. Little efforts like that were present throughout the early 1900s, until around 1968. Huizenga's grandson, Wayne Huizenga, made his grandfather's idea into a business, Waste Management. By the seventies, waste management as a whole was seen as a necessary practice by the public. Sanitation in the United States California/Counties In the early 1940s, many counties in the state had problems with their disposal of waste, especially in the Lake Tahoe area. Citizens of these towns feared that their city's poor sewage systems would cause outbreaks in illnesses, like poliomyelitis, cholera, and hepatitis, to name a few. Cholera in particular is the biggest health risk attached to waste management. The illness is caused by bacteria, especially when a person ingests food or water that contain the bacteria. In poorer areas, this is extremely likely due to the cross contamination of waste and drinking water. Counties El Dorado county has numerous garbage collection facilities, some private companies. In residential areas, the main source of waste is oil. Since then, many waste management facilities have been built in El Dorado county, reducing the risk of these illnesses. Since the fifties, the county has been utilizing the contacts from the companies to provide a low-cost and successful method of keeping the towns clean. Today, there are 7 franchises assigned to the county with different areas of pickup, such as El Dorado Disposal and American River Disposal. San Joaquin valley is very recycling focused. The website for the San Joaquin county's waste management shows many tips for how to recycle all recyclable items, in hopes that their county will comply. One of the tips is to verify that all items in the recycling bins are recyclable, because the load might not get recycled at all. The website is very helpful for the public for to help with waste management in residential areas. Education Engineering Wastewater engineering is not usually its own degree course, but a specialization from degrees such as environmental and sanitary engineering, sanitary engineering, civil engineering, environmental engineering, bio-chemical engineering, or chemical engineering. Formal education for wastewater engineers begins in high school with students taking classes such as chemistry, biology, physics, and higher mathematics including calculus. After high school most jobs require certification from a state agency. Those wanting to advance in the industry should pursue a sanitary engineering, environmental and sanitary engineering, civil engineering, mechanical engineering, environmental engineering, or a facilities engineering degree. Gaining experience through internships and working while in college is a common pathway toward advancement. Education about waste treatment requires course work in systems design, machinery design principles, water chemistry, and similar coursework. Other classes may include Chemistry of Plant Processes, and various plant operations courses. Wastewater engineers may advance in their careers through additional education and experience. With additional knowledge and experience one can become the manager of an entire plant. The accreditation body certifying the education for the degree and license is the Accreditation Board for Engineering and Technology (ABET). Over time, some companies may require the wastewater engineer to continue their education to keep up with any changes in technology. Obtaining one's master's degree is encouraged since many companies list it as a preference in selection. In this field 76 percent of those employed have a bachelor's degree, 17 percent have a master's degree and three percent have a post-doctoral degree as of 2013. The average annual salary is approximately $83,360. Plant Operations Initial employment in wastewater engineering can be obtained by those with and without advanced formal education. The California State Water Resources Control Board (SWRCB), for example, shows how individuals can advance through a progression of certifications as Waste Water Treatment Operators. The Board uses a five level classification system to classify water treatment facilities into categories I-V according to the population served and the complexity of the treatment system. The Operator Certification requirements for water treatment operators and waste water treatment operators are described in detail by State law. To meet certification requirements, operators must submit an application to SWRCB, have the necessary work experience, meet the educational requirements, and pass an examination based on the knowledge, skill, and abilities described in the regulations. Operators are required to renew their certificates every three years. To be eligible for renewal, certified operators must complete a specified number of continuing education hours after the previous issuance of a certificate. Job description and typical tasks Important job types working in sanitary engineering include sanitation workers, waste collectors and wastewater engineers. Wastewater engineers use a variety of skills and must have knowledge of mechanical and environmental engineering. They are required to perform tasks and demonstrate knowledge in design, mathematics, English, construction, physics, chemistry, biology, management, and personnel. Wastewater engineers must have skills in complex problem solving, critical thinking, mathematics, active listening, judgement, reading comprehension, speaking, writing, science, and system analysis. Typical work activities include problem solving, communication with management and staff, gathering information, analyzing data, evaluating standards and complying with them, and communicating with others in the field. Wastewater engineers perform these activities by combining their knowledge and skills to perform tasks. These tasks are to understand computer-aided design programs, and to conduct studies for the construction of facilities, water supply systems and collection systems. They may design systems for wastewater collection machinery, as well as system components. They may perform water flow analysis, then select designs and equipment based on government and industry standards. Some are involved with a specific area of concern such as waste collection or the maintenance of waste water facilities and stormwater drainage systems within an area. Others cover a broader scope of activities that might include maintenance of the public water supply, collection of residential yard waste program, disposal of hazardous waste, recycling strategies and even community programs where individuals or businesses "adopt" an area and either maintain it themselves or donate funds for doing so. Wastewater engineers may also map out topographical and geographical features of Earth to determine the best means of collection, design pipe and pumped collection systems, and design treatment processes for collected wastewater. Typical employers Wastewater engineers work for private companies, state and local governments, and special districts. Modern challenges Water scarcity Water managers confront new challenges and the need for new technology as water levels decrease due to increasingly frequent and extended droughts. Technologies such as sonar mapping are being used in wells to determine the volume of water that they can hold. For example, the United States Geological Survey and the State of New York worked together to map underground aquifers since the 1980s. Today they have thorough maps of these aquifers to assist in water management. Desalination plants may be required in the future for those regions hardest hit by water scarcity. Desalination is a process of cleaning water by means of evaporation. Water is evaporated and it passes through membranes. The water is then cooled and condenses allowing it to flow either back into the main water line or out to sea. Smart Sanitation Smart Sanitation: Advances in sensor technology, data analytics, and automation are enabling the development of smart sanitation systems that can monitor water quality, detect leaks, optimize treatment processes, and improve overall efficiency. Sanitary engineers need to leverage these technologies to enhance the performance and reliability of sanitation infrastructure. Climate change Wastewater treatment contributes to global warming in many ways. One of the factors that contributes to global warming is wastewater treatment facilities and their emissions of greenhouse gases. Some of those gases are carbon dioxide, methane, and nitrous oxide. These gases occur because of the decomposition of organic material from the anaerobic bacteria. These bacteria clean the leftover waste. Even if the anaerobic bacteria decomposition produces these gases, the percentage of greenhouse gases that other equipment produce is still greater than the contribution of the anaerobic bacteria. Also, the power usage from those machinery is very high. That is why many facilities are undergoing renovation to use higher levels of anaerobic bacteria compared to other types of equipment. Impacts of climate change on sanitary engineering vary based on region and the sanitation solutions employed there. In the Arctic, permafrost melting has caused damage to pipes and other infrastructure. In the Northeastern United States, increased precipitation has overwhelmed aging infrastructure not equipped to handle the massive volume of water from heavy precipitation. In the Western United States, prolonged drought has decreased water availability. This has led some wastewater facilities to expand recycled and reclaimed water programs. Climate change has also affected water distribution pipes. Physical stress from climate change-related conditions such as extreme rainfall or drought increases the rate of pipe corrosion, adding to facility cost.
Technology
Disciplines
null
8417064
https://en.wikipedia.org/wiki/Occupational%20medicine
Occupational medicine
Occupational and Environmental Medicine (OEM), previously called industrial medicine, is a board certified medical specialty under the American Board of Preventative Medicine that specializes in the prevention and treatment of work-related illnesses and injuries. OEM physicians are trained in both clinical medicine and public health. They may work in a clinical capacity providing direct patient care to workers through worker's compensation programs or employee health programs and performing medical screening services for employers. Corporate medical directors are typically occupational medicine physicians who often have specialized training in the hazards relevant to their industry. OEM physicians are employed by the US military in light of the significant and unique exposures faced by this population of workers. Public health departments, the Occupational Safety and Health Administration (OSHA) and the National Institute of Occupational Safety and Health (NIOSH) commonly employ physicians specialized in occupational medicine. They often advise international bodies, governmental and state agencies, organizations, and trade unions. The specialty of Occupational Medicine rose in prominence following the industrial revolution. Factory workers and laborers in a broad host of emergent industries at the time were becoming profoundly ill and often dying due to work exposures which prompted formal efforts to better understand, recognize, treat and prevent occupational injury and disease. More recently occupational medicine gained visibility during the COVID-19 Pandemic as spread of the illness was intricately linked to the workplace necessitating dramatic adjustments in workplace health, safety and surveillance practices. In the United States, the American College of Preventive Medicine oversees board certification of physicians in Occupational and Environmental Medicine Mission Occupational medicine aims to prevent diseases and promote wellness among workers. Occupational health physicians must: Have knowledge of potential hazards in the workplace including toxic properties of materials used. Be able to evaluate employee fitness for work. Be able to diagnose and treat occupational disease and injury. Know about rehabilitation methods, health education, and government laws and regulations concerning workplace and environmental health. Be able to manage health service delivery. OM can be described as: work that combines clinical medicine, research, and advocacy for people who need the assistance of health professionals to obtain some measure of justice and health care for illnesses they suffer as a result of companies pursuing the biggest profits they can make, no matter what the effect on workers or the communities they operate in. History The first textbook of occupational medicine, De Morbis Artificum Diatriba (Diseases of Workers), was written by Italian physician Bernardino Ramazzini in 1700. Notable Occupational Medicine Physicians Dr. Alice Hamilton Dr. Stephen M Levin Dr. Archibald Cochrane (Preventative Medicine) Governmental bodies United States National Institute for Occupational Safety and Health (NIOSH) Occupational Safety and Health Administration (OSHA) Russian Federation Research Institute of Occupational Medicine of the Russian Academy of Sciences (Moscow) Non-governmental organizations International International Commission on Occupational Health (ICOH) Institute of Occupational Medicine (IOM) Canadian Occupational Medicine Specialists of Canada Japan Japan Society of Occupational Health United Kingdom Faculty of Occupational Medicine United States American College of Occupational and Environmental Medicine (ACOEM) American Osteopathic College of Occupational & Preventive Medicine (AOCOPM) Europe European Society for Environmental and Occupational Medicine (EOM) Australasia ANZSOM Australia https://www.anzsom.org.au/ ANZSOM New Zealand https://anzsom.org.nz/
Biology and health sciences
Fields of medicine
Health
8421646
https://en.wikipedia.org/wiki/Open-end%20spinning
Open-end spinning
Open-end spinning is a technology for creating yarn without using a spindle. It was invented and developed in Czechoslovakia in Výzkumný ústav bavlnářský / Cotton Research Institute in Ústí nad Orlicí in 1963. Method It is also known as break spinning or rotor spinning. The principle behind open-end spinning is similar to that of a clothes dryer spinning full of sheets. If you could open the door and pull out a sheet, it would spin together as you pulled it out. Sliver from the card goes into the rotor, is spun into yarn and comes out, wrapped up on a bobbin, all ready to go to the next step. There is no roving stage or re-packaging on an auto-coner. This system is much less labour-intensive and faster than ring spinning with rotor speeds up to 140,000 rpm. The Rotor design is the key to the operation of the open-ended spinners. Each type of fibre may require a different rotor design for optimal product quality and processing speed. The first open-end machines in the United Kingdom were placed, under great secrecy, by Courtaulds into Maple Mill, Oldham in 1967. One disadvantage of open-end spinning is that it is limited to coarser counts, another is the structure of the yarn itself with fibres less in parallel compared to ring-spun yarns, for example, consequently cloth made from open-end yarn has a "fuzzier" feel and poorer wear resistance. History The global demand for spun fibre is huge. Converting raw fibre to yarn is a complicated process. Many manufacturers compete to provide the spinning machines that are essential to meeting the demand by delivering increases in spinning productivity and additional improvements in yarn quality. Over the past three centuries spinning technology has been continuously improved through thousands of minor innovations, and occasional major advances that have collectively increased the quality and lowered the cost of producing yarn dramatically. Major technology advances have included: Hand spinning Mule spinning Ring spinning Rotor spinning Dref Friction Spinning Open-end spinning The number of manufacturers who can successfully compete has been reduced, as the technical complexity of the spinning machines has increased. However, there are many competent companies serving the global market for spinning machines who continue to pursue innovative ways to increase spinning productivity and yarn quality. Characteristics A good open-end machine should have: Higher productivity This is a major criterion, as productivity reduces the cost of manufacturing. The O.E. machines that are now in market boasts of many a basic needs like, longer length of machine, higher speeds, able to process coarser hank, fewer changes for count, easy access to parts (less downtime for cleaning), longer production time between cleaning schedules, computerized controls for less power consumption and lower downtime and complete report generation giving leads to problem area are some points to discuss. High-content sliver cans (up to 18”) In early days large machines were equipped with less distance between rotors (gauge of machine). This led to creeling of very small cans, which required frequent can changes. All major manufacturers currently allow cans up to 18” diameter leading to less breakage, less joining of yarn, hence better quality and higher productivity. Originally round cans were used. Rectangular cans are used because they double sliver capacity in the same sliver can footprint. Larger packages of yarn (4 to 5 kg) The final package size has continued to increase. The final package size is important because it reduces tube change frequency and thus reduces idle time for creeling. Current yarn packages typically weigh 4–5 kg. The Savio Super Spinner 3000 currently has the largest package size at 6 kg. Less power consumption Using individual motors and electronic controls for each of the various drives of the machine maximizes energy efficiency and minimizes downtime. Automation All spinning machines, whether ring or open-end, need yarn joining to repair breaks or start new sliver cans. Joining the yarn has historically been a labor-intensive activity and a source of quality defects. Autopiecing units are robots that automate this process. Market leaders like Schlafhorst, Rieter, Savio, have machines that incorporate good quality autopiecers and autodoffing. This automation leads to less material handling costs and helps improve quality of the final product. Flexibility of spinning components Many vendors are offering machines that can be programmed to produce many different types of yarns. The ability to rapidly change production results in the flexibility to serve multiple markets. A contemporary spinning mill should be able to produce a range of products: denim, knitting, towels, structured fabrics, construction fabrics, and various other products like core spun, multi count, etc. Handling count range. Machines need to be easily programmed to spin yarns from 4sNe to 60sNe. This ability allows a single machine to produce yarns that cater to many different end-user requirements. Advantages Disappearance of simplex frame. Under certain circumstances, elimination of the second passage draw frame. In some cases, with the use of auto-leveller at the cards, elimination of even the draw frame passage. Bigger supply of cans to open-end and bigger packages to weaving. Elimination of winding. Less labor and power cost per kilogram of yarn. Higher productivity almost 7 times in the case of 10s and high efficiency. Fully automated mill a reality. Disadvantages Restricted only coarse counts. High capital cost. Usage restricted in case yarn is weak. Yarn realization in the case of waste mixing will be poor, resulting in increased mixing cost. Wear and tear of rotors, combing rollers, and navels are very high when high trash content mixing is used resulting in heavy replacement cost. In case reeling is done additional reeling cost is involved resulting in higher manufacturing cost. Products Linen / Flax yarns Cotton Yarns Polyester Cotton Blended yarn Tencel 100% Polyester 100% Polyester / Cotton / Linen / Viscose Multi blend Dyed yarn (and fibre) Acrylic/Rayon Recycle Polyester 100% and different Blends
Technology
Spinning
null
138214
https://en.wikipedia.org/wiki/Five%20lemma
Five lemma
In mathematics, especially homological algebra and other applications of abelian category theory, the five lemma is an important and widely used lemma about commutative diagrams. The five lemma is not only valid for abelian categories but also works in the category of groups, for example. The five lemma can be thought of as a combination of two other theorems, the four lemmas, which are dual to each other. Statements Consider the following commutative diagram in any abelian category (such as the category of abelian groups or the category of vector spaces over a given field) or in the category of groups. The five lemma states that, if the rows are exact, m and p are isomorphisms, l is an epimorphism, and q is a monomorphism, then n is also an isomorphism. The two four-lemmas state: Proof The method of proof we shall use is commonly referred to as diagram chasing. We shall prove the five lemma by individually proving each of the two four lemmas. To perform diagram chasing, we assume that we are in a category of modules over some ring, so that we may speak of elements of the objects in the diagram and think of the morphisms of the diagram as functions (in fact, homomorphisms) acting on those elements. Then a morphism is a monomorphism if and only if it is injective, and it is an epimorphism if and only if it is surjective. Similarly, to deal with exactness, we can think of kernels and images in a function-theoretic sense. The proof will still apply to any (small) abelian category because of Mitchell's embedding theorem, which states that any small abelian category can be represented as a category of modules over some ring. For the category of groups, just turn all additive notation below into multiplicative notation, and note that commutativity of abelian group is never used. So, to prove (1), assume that m and p are surjective and q is injective. Let c′ be an element of C′. Since p is surjective, there exists an element d in D with p(d) = t(c′). By commutativity of the diagram, u(p(d)) = q(j(d)). Since im t = ker u by exactness, 0 = u(t(c′)) = u(p(d)) = q(j(d)). Since q is injective, j(d) = 0, so d is in ker j = im h. Therefore, there exists c in C with h(c) = d. Then t(n(c)) = p(h(c)) = t(c′). Since t is a homomorphism, it follows that t(c′ − n(c)) = 0. By exactness, c′ − n(c) is in the image of s, so there exists b′ in B′ with s(b′) = c′ − n(c). Since m is surjective, we can find b in B such that b′ = m(b). By commutativity, n(g(b)) = s(m(b)) = c′ − n(c). Since n is a homomorphism, n(g(b) + c) = n(g(b)) + n(c) = c′ − n(c) + n(c) = c′. Therefore, n is surjective. Then, to prove (2), assume that m and p are injective and l is surjective. Let c in C be such that n(c) = 0. t(n(c)) is then 0. By commutativity, p(h(c)) = 0. Since p is injective, h(c) = 0. By exactness, there is an element b of B such that g(b) = c. By commutativity, s(m(b)) = n(g(b)) = n(c) = 0. By exactness, there is then an element a′ of A′ such that r(a′) = m(b). Since l is surjective, there is a in A such that . By commutativity, . Since m is injective, f(a) = b. So c = g(f(a)). Since the composition of g and f is trivial, c = 0. Therefore, n is injective. Combining the two four lemmas now proves the entire five lemma. Applications The five lemma is often applied to long exact sequences: when computing homology or cohomology of a given object, one typically employs a simpler subobject whose homology/cohomology is known, and arrives at a long exact sequence which involves the unknown homology groups of the original object. This alone is often not sufficient to determine the unknown homology groups, but if one can compare the original object and sub object to well-understood ones via morphisms, then a morphism between the respective long exact sequences is induced, and the five lemma can then be used to determine the unknown homology groups.
Mathematics
Category theory
null
138301
https://en.wikipedia.org/wiki/Dwarfism
Dwarfism
Dwarfism is a condition of people and animals marked by unusually small size or short stature. In humans, it is sometimes defined as an adult height of less than , regardless of sex; the average adult height among people with dwarfism is . Disproportionate dwarfism is characterized by either short limbs or a short torso. In cases of proportionate dwarfism, both the limbs and torso are unusually small. Intelligence is usually normal, and most people with it have a nearly normal life expectancy. People with dwarfism can usually bear children, although there are additional risks to the mother and child depending upon the underlying condition. The most common and recognizable form of dwarfism in humans (comprising 70% of cases) is achondroplasia, a genetic disorder whereby the limbs are diminutive. Growth hormone deficiency is responsible for most other cases. There are many other less common causes. Treatment of the condition depends on the underlying cause. Those with genetic disorders such as osteochondrodysplasia can sometimes be treated with surgery or physical therapy. Hormone disorders can also be treated with growth hormone therapy before the child's growth plates fuse. Individual accommodations such as specialized furniture, are often used by people with dwarfism. Many support groups provide services to aid individuals and the discrimination they may face. In addition to the medical aspect of the condition there are social aspects. For a person with dwarfism, height discrimination can lead to ridicule in childhood and discrimination in adulthood. In the United Kingdom, United States, Canada, Australia, and other English-speaking countries, labels that some people with dwarfism accept include dwarf (plural: dwarfs), little person (LP), or person of short stature (see terminology). Historically, the term midget was used to describe dwarfs (primarily proportionate); however, some now consider this term offensive. Signs and symptoms A defining characteristic of dwarfism is an adult height less than 2.3% of the CDC standard growth charts. There is a wide range of physical characteristics. Variations in individuals are identified by diagnosing and monitoring the underlying disorders. There may not be any complications outside adapting to their size. Short stature is a common replacement of the term 'dwarfism', especially in a medical context. Short stature is clinically defined as a height within the lowest 2.3% of those in the general population. However, those with mild skeletal dysplasias may not be affected by dwarfism. In some cases of untreated hypochondroplasia, males grow up to . Though that is short in a relative context, it does not fall into the extreme ranges of the growth charts. Disproportionate dwarfism is characterized by shortened limbs or a shortened torso. In achondroplasia one has an average-sized trunk with short limbs and a larger forehead. Facial features are often affected and individual body parts may have problems associated with them. Spinal stenosis, ear infection, and hydrocephalus are common. In case of spinal dysostosis, one has a small trunk, with average-sized limbs. Proportionate dwarfism is marked by a short torso with short limbs, thus leading to a height that is significantly below average. There may be long periods without any significant growth. Sexual development is often delayed or impaired into adulthood. This dwarfism type is caused by an endocrine disorder and not a skeletal dysplasia. Physical effects of malformed bones vary according to the specific disease. Many involve joint pain caused by abnormal bone alignment, or from nerve compression. Early degenerative joint disease, exaggerated lordosis or scoliosis, and constriction of spinal cord or nerve roots can cause pain and disability. Reduced thoracic size can restrict lung growth and reduce pulmonary function. Some forms of dwarfism are associated with disordered function of other organs, such as the brain or liver, sometimes severely enough to be more of an impairment than the unusual bone growth. Mental effects also vary according to the specific underlying syndrome. In most cases of skeletal dysplasia, such as achondroplasia, mental function is not impaired. However, there are syndromes which can affect the cranial structure and growth of the brain, severely impairing mental capacity. Unless the brain is directly affected by the underlying disorder, there is little to no chance of mental impairment that can be attributed to dwarfism. The psycho-social limitations of society may be more disabling than the physical symptoms, especially in childhood and adolescence, but people with dwarfism vary greatly in the degree to which social participation and emotional health are affected. Social prejudice against extreme shortness may reduce social and marital opportunities. Numerous studies have demonstrated reduced employment opportunities. Severe shortness is associated with lower income. Self-esteem may decline and family relationships may be affected. Extreme shortness (in the range) can, if not accommodated for, interfere with activities of daily living, like driving or using countertops built for taller people. Other common attributes of dwarfism such as bowed knees and unusually short fingers can lead to back problems and difficulty in walking and handling objects. Children with dwarfism are particularly vulnerable to teasing and ridicule from classmates. Because dwarfism is relatively uncommon, children may feel isolated from their peers. Causes Dwarfism can result from many medical conditions, each with its own separate symptoms and causes. Extreme shortness in humans with proportional body parts usually has a hormonal cause, such as growth hormone deficiency, once called pituitary dwarfism. Achondroplasia is responsible for the majority of human dwarfism cases, followed by spondyloepiphyseal dysplasia and diastrophic dysplasia. Achondroplasia The most recognizable and most common form of dwarfism in humans is achondroplasia, which accounts for 70% of dwarfism cases, and occurs in 4 to 15 out of 100,000 live births. It produces rhizomelic short limbs, increased spinal curvature, and distortion of skull growth. In achondroplasia the body's limbs are proportionately shorter than the trunk (abdominal area), with a larger head than average and characteristic facial features. Achondroplasia is an autosomal dominant disorder caused by the presence of an altered allele in the genome. If a pair of achondroplasia alleles are present, the result is fatal, usually perinatally. Achondroplasia is a mutation in the fibroblast growth factor receptor 3. In the context of achondroplasia, this mutation causes FGFR3 to become constitutively active, inhibiting bone growth. Growth hormone deficiency Growth hormone deficiency (GHD) is a medical condition in which the body produces insufficient growth hormone. Growth hormone, also called somatotropin, is a polypeptide hormone which stimulates growth and cell reproduction. If this hormone is lacking, stunted or even halted growth may become apparent. Children with this disorder may grow slowly and puberty may be delayed by several years or indefinitely. Growth hormone deficiency has no single definite cause. It can be caused by mutations of specific genes, damage to the pituitary gland, Turner's syndrome, poor nutrition, or even stress (leading to psychogenic dwarfism). Laron syndrome (growth hormone insensitivity) is another cause. Those with growth hormone issues tend to be proportionate. Metatropic dysplasia Metatropic means "changing form" and refers to this form of skeletal dysplasia as there is an abnormality in the growth plates. Skeletal changes continue over time and may need surgical intervention to help protect the lungs. Symptoms starting at birth may be mild or can be fatal. There are recognizable features in individuals with this genetic disorder. Some are short stature, narrow chest, "facial features such as a prominent forehead, underdevelopment of the upper jaw, cheekbones and eye sockets (midface hypoplasia), and a squared-off jaw." It is considered a more severe skeletal dysplasia, but is very rare, with the exact number of those affected unknown. Prognosis is largely on a case-by-case basis depending on the severity, and life expectancy may not be impacted unless there are respiratory complications. Other Other causes of dwarfism are spondyloepiphyseal dysplasia congenita, diastrophic dysplasia, pseudoachondroplasia, hypochondroplasia, Noonan syndrome, primordial dwarfism, Cockayne syndrome, Kniest dysplasia, Turner syndrome, osteogenesis imperfecta (OI), and hypothyroidism. Severe shortness with skeletal distortion also occurs in several of the mucopolysaccharidoses and other storage disorders. Hypogonadotropic hypogonadism may cause proportionate, yet temporary, dwarfism. NPR2 disproportionate dwarfism was discovered recently and is caused by a mutant gene. Serious chronic illnesses may produce dwarfism as a side effect. Harsh environmental conditions, such as malnutrition, may also produce dwarfism. These types of dwarfism are indirect consequences of the generally unhealthy or malnourished condition of the individual, and not of any specific disease. The dwarfism often takes the form of simple short stature, without any deformities, thus leading to proportionate dwarfism. In societies where poor nutrition is widespread, the average height of the population may be reduced below its genetic potential by the lack of proper nutrition. Sometimes there is no definitive cause of short stature. Diagnosis Dwarfism is often diagnosed in childhood on the basis of visible symptoms. A physical examination can usually suffice to diagnose certain types of dwarfism, but genetic testing and diagnostic imaging may be used to determine the exact condition. In a person's youth, growth charts that track height can be used to diagnose subtle forms of dwarfism that have no other striking physical characteristics. Short stature or stunted growth during youth is usually what brings the condition to medical attention. Skeletal dysplasia is usually suspected because of obvious physical features (e.g., unusual configuration of face or shape of skull), because of an obviously affected parent, or because body measurements (arm span, upper to lower segment ratio) indicate disproportion. Bone X-rays are often key to diagnosing a specific skeletal dysplasia, but are not the sole diagnostic tool. Most children with suspected skeletal dysplasias are referred to a genetics clinic for diagnostic confirmation and genetic counseling. Since about the year 2000, genetic tests for some of the specific disorders have become available. During an initial medical evaluation of shortness, the absence of disproportion and other clues listed above usually indicates causes other than bone dysplasias. Classification In men and women, the sole requirement for being considered a dwarf is having an adult height under and it is almost always sub-classified with respect to the underlying condition that is the cause of the short stature. Dwarfism is usually caused by a genetic variant; achondroplasia is caused by a mutation on chromosome 4. If dwarfism is caused by a medical disorder, the person is referred to by the underlying diagnosed disorder. Disorders causing dwarfism are often classified by proportionality. Disproportionate dwarfism describes disorders that cause unusual proportions of the body parts, while proportionate dwarfism results in a generally uniform stunting of the body. Disorders that cause dwarfism may be classified according to one of hundreds of names, which are usually permutations of the following roots: location rhizomelic = root, i.e., bones of the upper arm or thigh mesomelic = middle, i.e., bones of the forearm or lower leg acromelic = end, i.e., bones of hands and feet. micromelic = entire limbs are shortened source chondro = of cartilage osteo = of bone spondylo = of the vertebrae plasia = form trophy = growth Examples include achondroplasia and chondrodystrophy. Prevention Many types of dwarfism are currently impossible to prevent because they are genetically caused. Genetic conditions that cause dwarfism may be identified with genetic testing, by screening for the specific variations that result in the condition. However, due to the number of causes of dwarfism, it may be impossible to determine definitively if a child will be born with dwarfism. Dwarfism resulting from malnutrition or a hormonal abnormality may be treated with an appropriate diet or hormonal therapy. Growth hormone deficiency may be remedied via injections of human growth hormone (HGH) during early life. Management Genetic mutations of most forms of dwarfism caused by bone dysplasia cannot be altered yet, so therapeutic interventions are typically aimed at preventing or reducing pain or physical disability, increasing adult height, or mitigating psychosocial stresses and enhancing social adaptation. Forms of dwarfism associated with the endocrine system may be treated using hormonal therapy. If the cause is prepubescent hyposecretion of growth hormone, supplemental growth hormone may correct the abnormality. If the receptor for growth hormone is itself affected, the condition may prove harder to treat. Hypothyroidism is another possible cause of dwarfism that can be treated through hormonal therapy. Injections of thyroid hormone can mitigate the effects of the condition, but lack of proportion may be permanent. Pain and disability may be ameliorated by physical therapy, braces or other orthotic devices, or by surgical procedures. The only simple interventions that increase perceived adult height are dress enhancements, such as shoe lifts or hairstyle. Growth hormone is rarely used for shortness caused by bone dysplasias, since the height benefit is typically small (less than ) and the cost high. The most effective means of increasing adult height by several inches is distraction osteogenesis, though availability is limited and the cost is high in terms of money, discomfort, and disruption of life. Most people with dwarfism do not choose this option, and it remains controversial. For other types of dwarfism, surgical treatment is not possible. Society and culture Terminology The appropriate term for describing a person of particularly short stature (or with the genetic condition achondroplasia) has developed euphemistically. The noun dwarf stems from , originally referring to a being from Germanic mythology—a dwarf—that dwells in mountains and in the earth, and is associated with wisdom, smithing, mining, and crafting. The etymology of the word dwarf is contested, and scholars have proposed varying theories about the origins of the being, including that dwarfs may have originated as nature spirits or as beings associated with death, or as a mixture of concepts. Competing etymologies include a basis in the Indo-European root (meaning ), the Indo-European root (whence modern Dutch and ), and comparisons have been made with the Old Indian dhvaras (a type of demonic being). The being may not have gained associations with small stature until a later period. The terms "little person", "LP" and "person of short stature" are the preferred terms of many of those with this disorder, and while some are uncomfortable with "dwarf" it remains a common term in some areas. However, the plural "dwarfs" as opposed to "dwarves" is generally preferred in the medical context, possibly because the plural "dwarves" was popularized by author J. R. R. Tolkien, describing a race of characters in his The Lord of the Rings books resembling Norse dwarfs. "Midget", whose etymology indicates a "tiny biting insect", came into prominence in the mid-19th century after Harriet Beecher Stowe used it in her novels Sunny Memories of Foreign Lands and Oldtown Folks where she described children and an extremely short man, respectively. Later some people of short stature considered the word to be offensive because it was the descriptive term applied to P. T. Barnum's dwarfs used for public amusement during the freak show era. It is also not considered accurate as it is not a medical term or diagnosis, though it is sometimes used as a slang term to describe those who are particularly short, whether or not they have dwarfism. Participation Individuals with dwarfism are capable of actively participating in various aspects of society. They have access to education, sports, and can pursue careers, engaging in a wide range of professions. Acceptance Individuals with dwarfism often face prejudice and stereotypes. Research by Klein (2019) has demonstrated that awareness of the stigmatization of this group can promote full participation in society. The research by Green and Pinter (2018) in the field of humor and social psychology can provide insights to reduce stereotypes and promote a more objective perception. Accommodation In daily life, little people face numerous obstacles because the environment is tailored to average-sized individuals. Some little people can only use ATMs, kitchens, toilets, and sinks with aids. Low stools and step stools play a special role, as they can be used in various ways to bridge the height difference. Stools are also useful as footrests while sitting, as the legs of short people dangle in the air when sitting on an average chair, which can be painful and uncomfortable in the long run and may hinder fine motor skills during work. To be mobile, some individuals use customized scooters or bicycles, as it can be problematic, depending on the type of short stature, to walk longer distances. With specially adapted vehicles, most individuals of short stature can drive without further hindrances. Generally, pedal extensions and an individually adjusted seat at the correct height are required. Some little people are tall enough to drive without pedal extensions. Usually, patients with skeletal dysplasia with limited mobility can receive allowances or grants for vehicle assistance through governmental help or rehabilitation providers. Dwarf sports Dwarfs have support and compete in sport by a number of organizations nationally and internationally. They are included in some events in the athletics at the Summer Paralympics. The Dwarf Athletic Association of America and the Dwarf Sports Association UK provide opportunities for dwarfs to compete nationally and internationally in the Americas and Europe, respectively. The World Dwarf Games (WDG) are a multi-sport event for athletes of short stature. The WDG have been held every four years since 1993 and are the world's largest sporting event exclusively for athletes with dwarfism. The Dwarf Sports Association UK organizes between 5 and 20 events per month for athletes with restricted growth conditions in the UK. For instance, swimming and bicycling are often recommended for people with skeletal dysplasias, since those activities put minimal pressure on the spine. Since its early days, professional wrestling has had the involvement of dwarf athletes. "Midget wrestling" had its heyday in the 1950s–'70s, when wrestlers such as Little Beaver, Lord Littlebrook, and Fuzzy Cupid toured North America, and Sky Low Low was the first holder of the National Wrestling Alliance's World Midget Championship. In the next couple of decades, more wrestlers became prominent in North America including foreign wrestlers like Japan's Little Tokyo. Although the term is seen by some as pejorative, many past and current midget wrestlers including Hornswoggle said they take pride in the term due to its history in the industry and its marketability. Art and media depictions In art, literature, and movies, dwarfs are rarely depicted as ordinary people who are very short but rather as a species apart. Novelists, artists, and moviemakers may attach special moral or aesthetic significance to their "apartness" or misshapenness. Artistic representations of dwarfism are found on Greek vases and other ancient artifacts, including ancient Egyptian art in which dwarfs are likely to have been seen as a divine manifestation, with records indicating that they were able to reach high positions in society at the time. The ancient Hindu text Bhagavat Purana devotes nine chapters to the adventures of Vamana, a dwarf avatar of Lord Vishnu. Depictions of dwarfism are also found in European paintings and many illustrations. Many European paintings (especially Spanish) of the 16th–19th centuries depict dwarfs by themselves or with others. In the Talmud, it is said that the second born son of the Egyptian Pharaoh of the Bible was a dwarf. Recent scholarship has suggested that ancient Egyptians held dwarfs in high esteem. Several important mythological figures of the North American Wyandot nation are portrayed as dwarfs. As popular media has become more widespread, the number of works depicting dwarfs have increased dramatically. Dwarfism is depicted in many books, films, and TV series such as Willow, The Wild Wild West, The Man with the Golden Gun (and later parodied in Austin Powers), Gulliver's Travels by Jonathan Swift, The Wizard of Oz, Willy Wonka & the Chocolate Factory, Bad Santa, A Son of the Circus, Little People, Big World, The Little Couple, A Song of Ice and Fire (and its TV adaptation Game of Thrones), Seinfeld, The Orator, In Bruges, The Tin Drum by Günter Grass, the short-lived reality show The Littlest Groom, and the films The Station Agent and Zero. The Animal Planet TV series Pit Boss features dwarf actor Shorty Rossi and his talent agency, "Shortywood Productions", which Rossi uses to provide funding for his pit bull rescue operation, "Shorty's Rescue". Rossi's three full-time employees, featured in the series, are little people and aspiring actors. In September 2014, Creative Business House along with Donnons Leur Une Chance, created the International Dwarf Fashion Show to raise awareness and boost self-confidence of people living with dwarfism. A number of reality television series on Lifetime, beginning with Little Women: LA in 2014, focused on showing the lives of women living with dwarfism in various cities around the United States.
Biology and health sciences
Disabilities
Health
138484
https://en.wikipedia.org/wiki/Commutative%20diagram
Commutative diagram
In mathematics, and especially in category theory, a commutative diagram is a diagram such that all directed paths in the diagram with the same start and endpoints lead to the same result. It is said that commutative diagrams play the role in category theory that equations play in algebra. Description A commutative diagram often consists of three parts: objects (also known as vertices) morphisms (also known as arrows or edges) paths or composites Arrow symbols In algebra texts, the type of morphism can be denoted with different arrow usages: A monomorphism may be labeled with a or a . An epimorphism may be labeled with a . An isomorphism may be labeled with a . The dashed arrow typically represents the claim that the indicated morphism exists (whenever the rest of the diagram holds); the arrow may be optionally labeled as . If the morphism is in addition unique, then the dashed arrow may be labeled or . If the morphism acts between two arrows (such as in the case of higher category theory), it's called preferably a natural transformation and may be labelled as (as seen below in this article). The meanings of different arrows are not entirely standardized: the arrows used for monomorphisms, epimorphisms, and isomorphisms are also used for injections, surjections, and bijections, as well as the cofibrations, fibrations, and weak equivalences in a model category. Verifying commutativity Commutativity makes sense for a polygon of any finite number of sides (including just 1 or 2), and a diagram is commutative if every polygonal subdiagram is commutative. Note that a diagram may be non-commutative, i.e., the composition of different paths in the diagram may not give the same result. Examples Example 1 In the left diagram, which expresses the first isomorphism theorem, commutativity of the triangle means that . In the right diagram, commutativity of the square means . Example 2 In order for the diagram below to commute, three equalities must be satisfied: Here, since the first equality follows from the last two, it suffices to show that (2) and (3) are true in order for the diagram to commute. However, since equality (3) generally does not follow from the other two, it is generally not enough to have only equalities (1) and (2) if one were to show that the diagram commutes. Diagram chasing Diagram chasing (also called diagrammatic search) is a method of mathematical proof used especially in homological algebra, where one establishes a property of some morphism by tracing the elements of a commutative diagram. A proof by diagram chasing typically involves the formal use of the properties of the diagram, such as injective or surjective maps, or exact sequences. A syllogism is constructed, for which the graphical display of the diagram is just a visual aid. It follows that one ends up "chasing" elements around the diagram, until the desired element or result is constructed or verified. Examples of proofs by diagram chasing include those typically given for the five lemma, the snake lemma, the zig-zag lemma, and the nine lemma. In higher category theory In higher category theory, one considers not only objects and arrows, but arrows between the arrows, arrows between arrows between arrows, and so on ad infinitum. For example, the category of small categories Cat is naturally a 2-category, with functors as its arrows and natural transformations as the arrows between functors. In this setting, commutative diagrams may include these higher arrows as well, which are often depicted in the following style: . For example, the following (somewhat trivial) diagram depicts two categories and , together with two functors , : → and a natural transformation : ⇒ : There are two kinds of composition in a 2-category (called vertical composition and horizontal composition), and they may also be depicted via pasting diagrams (see 2-category#Definition for examples). Diagrams as functors A commutative diagram in a category C can be interpreted as a functor from an index category J to C; one calls the functor a diagram. More formally, a commutative diagram is a visualization of a diagram indexed by a poset category. Such a diagram typically includes: a node for every object in the index category, an arrow for a generating set of morphisms (omitting identity maps and morphisms that can be expressed as compositions), the commutativity of the diagram (the equality of different compositions of maps between two objects), corresponding to the uniqueness of a map between two objects in a poset category. Conversely, given a commutative diagram, it defines a poset category, where: the objects are the nodes, there is a morphism between any two objects if and only if there is a (directed) path between the nodes, with the relation that this morphism is unique (any composition of maps is defined by its domain and target: this is the commutativity axiom). However, not every diagram commutes (the notion of diagram strictly generalizes commutative diagram). As a simple example, the diagram of a single object with an endomorphism (), or with two parallel arrows (, that is, , sometimes called the free quiver), as used in the definition of equalizer need not commute. Further, diagrams may be messy or impossible to draw, when the number of objects or morphisms is large (or even infinite).
Mathematics
Category theory
null
19918814
https://en.wikipedia.org/wiki/Landscape%20photography
Landscape photography
Landscape photography (often shortened to landscape photos) shows the spaces within the world, sometimes vast and unending, but other times microscopic. Landscape photographs typically capture the presence of nature but can also focus on human-made features or disturbances of landscapes. Landscape photography is done for a variety of reasons. Perhaps the most common is to recall a personal observation or experience while in the outdoors, especially when traveling. Others pursue it particularly as an outdoor lifestyle, to be involved with nature and the elements, some as an escape from the artificial world. Many landscape photographs show little or no human activity and are created in the pursuit of a pure, unsullied depiction of nature, devoid of human influence—instead featuring subjects such as strongly defined landforms, weather, and ambient light. As with most forms of art, the definition of a landscape photograph is broad and may include rural or urban settings, industrial areas or nature photography. Environmentalism Some of the most important and celebrated landscape photographers have been motivated by an appreciation of the beauty of the natural environment and a desire to see it preserved. The work of William Henry Jackson in the mid-19th century was instrumental in convincing Congress in 1872 to create Yellowstone, the first national park in the United States. Photography produced by Philip Hyde for the Sierra Club found extensive use in promoting the preservation of natural places in the Western United States during the 20th century. Renowned landscape photographer Ansel Adams received both a Conservation Service Award and a Presidential Medal of Freedom in recognition of the influence of his work on the preservation of wilderness and fostering of environmental consciousness. Subjects Landscape photography commonly involves daylight photography of natural features of land, sky and waters, at a distance—though some landscapes may involve subjects in a scenic setting nearby, even close-up, and sometimes at night. Photography of artificial scenery, such as farm fields, orchards, gardens and architecture, may be considered "landscape" photography as well. Even the presence of human-made structures (buildings, roads and bridges, etc.) or art (such as sculpture) may be considered "landscape" if presented in artistic settings or appearing (or photographed) in artistic style. Further, landscape photography is typically of relatively stationary subjects—arguably a form of "still life." This tends to simplify the task, as opposed to photography of kinetic or live subjects. However, landscape photography often overlaps the activity of wildlife photography and the two terms are used somewhat interchangeably; both wildlife and landscapes may be elements of the same picture or body of work. Tourism Landscape photography has become a big part of local economies throughout the world. Countries such as Scotland, Iceland, Faroe Islands, New Zealand, USA, Canada and countries in the European Alpine region are very popular with photography tourists and welcome people from all over the world. As a result of this, landscape photography workshops and tours have become big business in these countries. Apps have formed part of this experience telling users how to find the most popular locations and how to photograph them once there. Methods (technical) Landscape photography typically requires relatively simple photographic equipment, though more sophisticated equipment can give a wider range of possibilities to the art. An artist's eye for the subject can yield attractive and impressive results even with modest equipment. Camera Any ordinary (or sophisticated) camera—film camera or digital camera—can be readily used for common landscape photography. Higher-resolution and larger-format digital cameras (or larger-format film cameras) permit a greater amount of detail and a wider range of artistic presentation. However, a larger-format camera yields a more limited depth of field (range of the scene that is in focus) for a given aperture value, requiring greater care in focusing (see: "Shutter Speed and Aperture", below). A camera with "panorama" function or frame can permit very wide images suitable for capturing a panoramic view. Lens For "wide open spaces," a wide-angle lens is generally the preferred lens, allowing a broad angle of view. However, medium-range to telephoto lenses can achieve satisfying imagery, as well, and can enable the capture of detailed scenery of smaller areas at greater distances. Telephoto lenses can also facilitate limited ranges of focus, to enable the photographer to emphasize a specific area, at a fairly specific distance, in sharp focus, with the foreground and background blurred (see: depth of field). A big difference between a wide-angles lens and a telephoto lens is the compression of the landscape; the wider the angle the more distance will appear between the foreground and background elements; however, a telephoto lens will make the same elements appear closer to each other. Other lenses that can help include the fisheye lens for extremely wide angles and dramatic effect, and the macro/micro lens for extreme close-up work. While variable-range zoom lenses are widely used, some landscape photographers prefer fixed-range prime lenses to provide higher clarity and quality in the image. Medium: film or digital sensor The sensitivity to light, of the medium—the film or the digital camera sensor—is important in landscape photography, especially where great detail is required. In bright daylight, a "slow film" (low-ISO film), or low-ISO digital camera sensor sensitivity setting (typically ISO 100, or perhaps 200), is generally preferred, allowing maximum precision and evenness of image. However, if there is movement in the scene, and the scene is in lower light—as with cloudy days, twilight, night, or in shaded areas—a higher ISO (up to the limits of the film or camera sensor, depending upon the shortage of light) may be desirable, to ensure that fast shutter speeds can be used to "freeze" the motion. Lighting and flash Normally, landscape photography—being focused primarily on natural beauty—tends to be done with only naturally occurring ambient light. In some cases, however, artificial light is recommended or unavoidable. Careful use of flash, continuous artificial lighting or reflective surfaces (e.g.: reflectors) for "fill" in shadowy areas is often used in close-up landscape photography (e.g.: garden spaces, small areas of dark forests, etc.). However, given the broad expanses of open space that tend to dominate in landscape photography, artificial lighting is typically ineffective, or even destructive (causing the foreground to be wildly over-lit, and the background to become overly dark). Light at dawn or dusk, or just before or after those times (especially at sunrise, or during the "golden hour" just before sunset), is often considered the best for capturing detail, showing scenes in the best colors of light, or otherwise generating impressive and attractive images. Shutter speed and aperture With cameras that allow a variety of shutter speeds and lens apertures, landscape photographers tend to prefer settings that allow all of the viewed area to be in sharp focus. This typically requires a small aperture (a high f-stop, usually between 11 and 13 is best for clarity and depth of field), which creates only a small hole for the light to come into the camera from the lens, ensuring that as much of the field of view is in focus as possible (see: depth of field). With a small aperture, however, a slower shutter speed (longer exposure) may be required to compensate for the limited amount of light squeezing in through the small aperture. This can be a problem if there are kinetic elements in the picture, such as moving animals (especially birds), people or vehicles. It can also be a problem if the environment is kinetic (in motion), such as wind blowing and shaking all the trees and plants in the scene, or if water is flowing. Slow shutter speeds can also be a problem if the photographer is in motion (such as shooting a scene from a moving vehicle). Consequently, some compromise between shutter speed and aperture may be necessary, or advisable. To some extent, a higher-ISO film or digital camera setting can compensate without the need to alter shutter speed or aperture. However, higher ISO settings ("fast film") can result in grainy pictures and poor capture of details, especially at a distance. In some cases, a slow shutter speed is desired to show movement of the subjects, particularly moving water or the effects of wind. Filters Filters can serve a wide range of purposes in landscape photography. For instance, a polarizing filter can darken the sky, while allowing surface features to be shown in relatively sharper clarity. Polarizing filters also help with cutting glare from water reducing reflections, snow and ice—even facilitating greater transparency of water and ice. Neutral density filters are darkened with a neutral (colorless) gray tint which reduces the amount of light entering the camera lens. These filters are used to lengthen shutter speeds without the need to alter aperture or film/sensor sensitivity, or alternatively, to use large apertures without exceeding the maximum shutter speed of a camera. A variation of this filter, termed the graduated neutral density filter or simply 'ND grad', transitions from dark, neutral gray on one side to clear on the opposite side. Photographers use these filters to lower natural contrasts by reducing light transmission from the brightest portion of the subject landscape, while letting light from the darker portion of the landscape enter the lens unobstructed. UV-Zero haze filters reduce "purple fringing" caused by ultraviolet light, especially in digital situations. They are also recommended by some professional photographers as protection for the vulnerable lens, especially when outdoors or in dynamic situations. Color filters can create other effects, or compensate for the appearance of unnatural lighting due to camera characteristics. Other accommodations In order to mitigate shaking associated with hand-holding a camera, landscape photography oftentimes requires a firm camera footing which affords the potential for sharper imagery. Tripods are specifically designed for stabilizing cameras and are widely regarded as essential equipment for landscape photography. However, any firm surface unaffected by vibration, wind or human contact may offer similar benefits. The use of a timer, remote control or cable release allows the shutter to be tripped without the introduction of vibration that might result from manually depressing the shutter button. Some modern, high-quality cameras also provide image stabilization, which compensates for vibration by moving inner workings of the camera, or electronically correcting the photograph. Because landscape photography is normally outdoors photography, protection from the elements can be helpful. Shooting from inside a sheltering structure or stationary vehicle (engine off, occupants stationary) can be helpful. Use of an umbrella or other shield to keep camera and photographer dry can also be helpful. A waterproof container for the camera, with drying agent inside (e.g.: dry cloth) may be advised, and experts advise that the camera should be shielded from blowing dust, snow, and rain, and extremely harsh direct sunlight.
Technology
Photography
null
19926944
https://en.wikipedia.org/wiki/Hills%20cloud
Hills cloud
In astronomy, the Hills cloud (also called the inner Oort cloud and inner cloud) is a theoretical vast circumstellar disc, interior to the Oort cloud, whose outer border would be located at around 20,000 to 30,000 astronomical units (AU) from the Sun, and whose inner border, less well defined, is hypothetically located at , well beyond planetary and Kuiper Belt object orbits—but distances might be much greater. If it exists, the Hills cloud likely contains roughly 5 times as many comets as the Oort cloud. Overview The need for the Hills cloud hypothesis is intimately connected with the dynamics of the Oort cloud: Oort cloud comets are continually perturbed in their environment. A non-negligible fraction leave the Solar System, or tumble into the inner system where they evaporate, fall into the Sun, or collide with or are ejected by the giant planets. Hence, the Oort cloud should have been depleted long ago, but it is still well supplied with comets. The Hills cloud hypothesis addresses the persistence of the Oort cloud by postulating a densely populated, inner-Oort region—the "Hills cloud". Objects ejected from the Hills cloud are likely to end up in the classical Oort cloud region, maintaining the Oort cloud. It is likely that the Hills cloud has the largest concentration of comets in the whole Solar System. The existence of the Hills cloud is plausible, since many bodies have been found there already. It should be denser than the Oort cloud. Gravitational interaction with the closest stars and tidal effects from the galaxy have given circular orbits to the comets in the Oort cloud, which may not be the case for the comets in the Hills cloud. The Hills cloud's total mass is unknown; some scientists think it would be many times more massive than the outer Oort cloud. History Original Oort cloud model Between 1932 and 1981, astronomers believed that the Oort cloud proposed by Ernst Öpik and Jan Oort, and the Kuiper belt were the only reserves of comets in the Solar System. In 1932, Estonian astronomer Ernst Öpik hypothesized that comets were rooted in a cloud orbiting the outer boundary of the Solar System. In 1950, this idea was revived independently by Dutch astronomer Jan Oort to explain an apparent contradiction: Comets are destroyed after several passes through the inner Solar System, so if any had existed for several billion years (since the beginning of the Solar System), no more could be observed now. Oort selected 46 comets for his study that were best observed between 1850 and 1952. The distribution of the reciprocal of the semi-major axes showed a maximum frequency which suggested the existence of a reservoir of comets between away. This reservoir, located at the limits of the Sun's sphere of influence (astrodynamics), would be subject to stellar disturbances, likely to expel cloud comets outwards or impel them inwards. New model In the 1980s, astronomers realized that the main cloud could have an internal section that would start at about 3,000 AU from the Sun and continue up to the classic cloud at 20,000 AU. Most estimates place the population of the Hills cloud at about 20 trillion (about five to ten times that of the outer cloud), although the number could be ten times greater than that. The main model of an "inner cloud" was proposed in 1981 by the astronomer Jack G. Hills, from the Los Alamos Laboratory, who gave the region its name. He calculated that the passage of a star near the Solar System could have triggered a "comet rain," thereby causing extinctions on Earth. His research suggested that the orbits of most cloud comets have a semi-major axis of 10,000 AU, much closer to the Sun than the proposed distance of the Oort cloud. Moreover, the influence of the surrounding stars and that of the galactic tide should have sent the Oort cloud comets either closer to the Sun or outside of the Solar System. To account for these issues, Hills proposed the presence of an inner cloud, which would have tens or hundreds of times as many comet nuclei as the outer halo. Thus, it would be a possible source of new comets to resupply the tenuous outer cloud. In the following years other astronomers searched for the Hills cloud and studied long-period comets. This was the case with Sidney van den Bergh and Mark E. Bailey, who each suggested the Hills cloud's structure in 1982 and 1983, respectively. In 1986, Bailey stated that the majority of comets in the Solar System were located not in the Oort cloud area, but closer and in an internal cloud, with an orbit with a semi-major axis of 5,000 AU. The research was further expanded upon by studies of Victor Clube and Bill Napier (1987), and by R. B. Stothers (1988). However, the Hills cloud gained major interest in 1991, when scientists resumed Hills' theory. Characteristics Structure and composition Oort cloud comets are constantly disturbed by their surroundings and distant objects. A significant number either leave the Solar System or go much closer to the Sun. The Oort cloud should therefore have broken apart long ago, but it still remains intact. The Hills cloud proposal could provide an explanation; J. G. Hills and other scientists suggest that it could replenish the comets in the outer Oort cloud. It is also likely that the Hills cloud is the largest concentration of comets across the Solar System. The Hills cloud should be much denser than the outer Oort cloud: If it exists, it is somewhere between 5,000 and 20,000 AU in size. In contrast, the Oort cloud is between in size. The mass of the Hills cloud is not known. Some scientists believe it could be five times more massive than the Oort cloud. Mark E. Bailey estimates the mass of the Hills cloud to be 13.8 Earth masses, if the majority of the bodies are located at 10,000 AU. If the analyses of comets are representative of the whole, the vast majority of Hills cloud objects consists of various ices, such as water, methane, ethane, carbon monoxide and hydrogen cyanide. However, the discovery of the object 1996 PW, an asteroid on a typical orbit of a long-period comet, suggests that the cloud may also contain rocky objects. The carbon analysis and isotopic ratios of nitrogen firstly in the comets of the families of the Oort cloud and the other in the body of the Jupiter area shows little difference between the two, despite their distinctly remote areas. This suggests that both come from a protoplanetary disk, a conclusion also supported by studies of comet cloud sizes and the recent impact study of comet Tempel 1. Formation Many scientists think that the Hills cloud formed from a close (800 AU) encounter between the Sun and another star within the first 800 million years of the Solar System, which could explain the eccentric orbit of 90377 Sedna, which should not be where it is, being neither influenced by Jupiter nor Neptune, nor tidal effects. It is then possible that the Hills cloud would be "younger" than the Oort cloud. However, only Sedna and two other sednoids ( and 541132 Leleākūhonua) bear those irregularities; for and this theory is not necessary, because both orbit close to the Solar System's gas giants. Possible Hills cloud objects Bodies in the Hills cloud are made mostly of water ice, methane and ammonia. Astronomers suspect many long-period comets originate from the Hills cloud, such as Comet Hyakutake. In their article announcing the discovery of Sedna, Mike Brown and his colleagues asserted that they observed the first Oort cloud object. They observed that, unlike scattered disc objects like Eris, Sedna's perihelion (76 AU) was too remote for the gravitational influence of Neptune to have played a role in its evolution. The authors regarded Sedna as an "inner Oort cloud object", located along the Ecliptic and placed between the Kuiper belt and the more spherical part of the Oort cloud. However, Sedna is much closer to the Sun than expected for objects in the Hills cloud and its inclination is close to that of the planets and the Kuiper belt. Considerable mystery surrounds , with its retrograde orbit that could make it originate from the Hills cloud or perhaps the Oort cloud. The same goes for damocloids, whose origins are doubtful, such as the namesake for this category, 5335 Damocles. Comets Astronomers suspect that several comets come from the same region as the Hills cloud; in particular, they focus on those with aphelia greater than 1,000 AU (which are thus from a farther region than the Kuiper belt), but less than 10,000 AU (or they would otherwise be too close to the outer Oort cloud). Some famous comets reach great distances and are candidates for Hills cloud objects. For example, Comet Lovejoy, discovered on 15 March 2007 by Australian astronomer Terry Lovejoy, had an inbound aphelion distance of around 1,800 AU. Comet Hyakutake, discovered in 1996 by amateur astronomer Yuji Hyakutake, has an outbound aphelion of 3,500 AU. Comet McNaught, discovered on 7 August 2006 in Australia by Robert H. McNaught, became one of the brightest comets of recent decades, with an aphelion of 4,100 AU. Comet Machholz, discovered on 27 August 2004 by amateur astronomer Donald Machholz, came from about 5,000 AU. Sedna Sedna is a dwarf planet discovered by Michael E. Brown, Chad Trujillo and David L. Rabinowitz on 14 November 2003. Spectroscopic measures show that its surface composition is similar to that of other trans-Neptunian objects: It is mainly composed of a mixture of water ices, methane, and nitrogen with tholins. Its surface is one of the reddest in the Solar System. This may be the first detection of an object originating from the Hills cloud, depending on the definition used. The area of the Hills cloud is defined as any objects with orbits measuring between 1,500 and 10,000 AU. Sedna is, however, much closer than the supposed distance of the Hills cloud. The planetoid discovered at a distance of about from the Sun, travels in an elliptical orbit of 11,400 years with a perihelion point of only 76 AU from the Sun during its closest approach (the next to occur in 2076), and travels out to 936 AU at its farthest point. However, Sedna is not considered a Kuiper belt object, because its orbit does not bring it into the region of the Kuiper belt at 50 AU. Sedna is a "detached object", and thus is not in a resonance with Neptune. The Trans-Neptunian object was announced on 26 March 2014 and has a similar orbit to Sedna with a perihelion point significantly detached from Neptune. Its orbit lies between 80 and 400 AU from the Sun.
Physical sciences
Solar System
Astronomy
3635265
https://en.wikipedia.org/wiki/Brontosaurus
Brontosaurus
Brontosaurus (; meaning "thunder lizard" from the Greek words , "thunder" and , "lizard") is a genus of herbivorous sauropod dinosaur that lived in present-day United States during the Late Jurassic period. It was described by American paleontologist Othniel Charles Marsh in 1879, the type species being dubbed B. excelsus, based on a partial skeleton lacking a skull found in Como Bluff, Wyoming. In subsequent years, two more species of Brontosaurus were named: B. parvus in 1902 and B. yahnahpin in 1994. Brontosaurus lived about 156 to 146 million years ago (mya) during the Kimmeridgian and Tithonian ages in the Morrison Formation of what is now Utah and Wyoming. For decades, the animal was thought to have been a taxonomic synonym of its close relative Apatosaurus, but a 2015 study by Emmanuel Tschopp and colleagues found it to be distinct. It has seen widespread representation in popular culture, being the archetypal "long-necked" dinosaur in general media. The anatomy of Brontosaurus is well known, with fossils demonstrating that it was large, long-necked, and quadrupedal with a long tail terminating in a whip-like structure. The cervical vertebrae are notably extremely robust and heavily-built, in contrast to its lightly built relatives Diplodocus and Barosaurus. The forelimbs were short and stout whereas the hindlimbs were elongated and thick, supported respectively by a heavily built shoulder girdle and pelvis. Several size estimates have been made, with the largest species B. excelsus reaching up to from head to tail and weighing in at , whereas the smaller B. parvus only got up to long. Juvenile specimens of Brontosaurus are known, with younger individuals growing rapidly to adult size in as little as 15 years. Brontosaurus has been classified within the family Diplodocidae, which was a group of sauropods that had shorter necks and longer tails compared to other families like brachiosaurs and mamenchisaurs. Diplodocids first evolved in the Middle Jurassic but peaked in diversity during the Late Jurassic with forms like Brontosaurus before becoming extinct in the Early Cretaceous. Brontosaurus is a genus in the subfamily Apatosaurinae, which includes only it and Apatosaurus, which are distinguished by their firm builds and thick necks. Although Apatosaurinae was named in 1929, the group was not used validly until an extensive 2015 paper, which found Brontosaurus to be valid. However, the status of Brontosaurus is still uncertain, with some paleontologists still considering it a synonym of Apatosaurus. Being from the Morrison Formation, Brontosaurus coexisted with a menagerie of other taxa such as the sauropods Diplodocus, Barosaurus, and Brachiosaurus; herbivorous ornithischians Stegosaurus, Dryosaurus, and Nanosaurus; as well as the carnivorous theropods Allosaurus, Marshosaurus and Ceratosaurus. This formation was a hotspot of sauropod biodiversity, with over 16 recognized genera, which resulted in niche partitioning between different sauropods. History of discovery The discovery of a large and fairly complete sauropod skeleton was announced in 1879 by Othniel Charles Marsh, a professor of paleontology at Yale University. The specimen was collected from Morrison Formation rocks at Como Bluff, Wyoming by William Harlow Reed. He identified it as belonging to an entirely new genus and species, which he named Brontosaurus excelsus, meaning "thunder lizard", from the Greek / meaning "thunder" and / meaning "lizard", and from the Latin excelsus, "noble" or "high". By this time, the Morrison Formation had become the center of the Bone Wars, a fossil-collecting rivalry between Marsh and another early paleontologist, Edward Drinker Cope. Because of this, the publications and descriptions of taxa by Marsh and Cope were rushed at the time. Brontosaurus excelsus type specimen (YPM 1980) was one of the most complete sauropod skeletons known at the time, preserving many of the characteristic but fragile cervical vertebrae. Marsh believed that Brontosaurus was a member of the Atlantosauridae, a clade of sauropod dinosaurs he named in 1877 that also included Atlantosaurus and Apatosaurus. A year later in 1880, another partial postcranial Brontosaurus skeleton was collected near Como Bluff by Reed, including well-preserved limb elements. Marsh named this second skeleton Brontosaurus amplus ("large thunder lizard") in 1881, but it was considered a synonym of B. excelsus in 2015. In August 1883, Marshall P. Felch collected a disarticulated partial skull (USNM V 5730) of a sauropod further south in the Felch Quarry at Garden Park, Colorado and sent the specimen to Yale. Marsh referred the skull to B. excelsus, later featuring it in a skeletal reconstruction of the B. excelsus type specimen in 1891 and the illustration was featured again in Marsh's landmark publication, The Dinosaurs of North America, in 1896. At the Yale Peabody Museum, the skeleton of Brontosaurus excelsus was mounted in 1931 with a skull based on the Marsh reconstruction of the Felch Quarry skull. While at the time most museums were using Camarasaurus casts for skulls, the Peabody Museum sculpted a completely different skull based on Marsh's recon. Marsh's skull was inaccurate for several other reasons: it included forward-pointing nasals, something truly different to any other dinosaur, and fenestrae differing from the drawing and other skulls. The mandible was based on a Camarasaurus'. In 1998, the Felch Quarry skull that Marsh included in his 1896 skeletal restoration was suggested to belong to Brachiosaurus instead and this was supported in 2020 with a redescription of the brachiosaurid material found at the Felch Quarry. Second Dinosaur Rush and skull issue During a Carnegie Museum expedition to Wyoming in 1901, William Harlow Reed collected another Brontosaurus skeleton, a partial postcranial skeleton of a young juvenile (CM 566), including partial limbs. However, this individual was found intermingled with a fairly complete skeleton of an adult (UW 15556). The adult skeleton specifically was very well-preserved, bearing many cervical (neck) and caudal (tail) vertebrae, and is the most complete definite specimen of the species. The skeletons were granted a new genus and species name, Elosaurus parvus ("little field lizard"), by Olof A. Peterson and Charles Gilmore in 1902. Both of the specimens came from the Brushy Basin Member of the Morrison Formation. The species was later transferred to Apatosaurus by several authors In 2008, a nearly complete postcranial skeleton of an apatosaurine was collected in Utah by crews working for Brigham Young University (BYU 1252-18531) where some of the remains are currently on display. The skeleton is undescribed, but many of the features of the skeleton are shared with A. parvus. The species was placed in Brontosaurus Tschopp et al. in 2015 during their comprehensive study of Diplodocidae.In the 1903 edition of Geological Series of the Field Columbian Museum, Elmer Riggs argued that Brontosaurus was not different enough from Apatosaurus to warrant a separate genus, so he created the new combination Apatosaurus excelsus for it. Riggs stated that "In view of these facts the two genera may be regarded as synonymous. As the term 'Apatosaurus' has priority, Brontosaurus' will be regarded as a synonym". Nonetheless, before the mounting of the American Museum of Natural History specimen, Henry Fairfield Osborn chose to label the skeleton "Brontosaurus", though he was a strong opponent of Marsh and his taxa. In 1905, the American Museum of Natural History (AMNH) unveiled the first-ever mounted skeleton of a sauropod, a composite specimen (mainly made of bones from AMNH 460) that they referred to as Brontosaurus excelsus. The AMNH specimen was very complete, only missing the feet, from the specimen AMNH 592 were added to the mount, lower leg and shoulder bones, added from AMNH 222, and tail bones, added from AMNH 339. To finish the mount, the rest of the tail was fashioned to appear as Marsh believed it should, which meant it had too few vertebrae. In addition, a sculpted model of what the museum felt the skull of this massive creature might have looked like was placed on the skeleton. This was not a delicate skull like that of Diplodocus, which would later turn out to be more accurate, but was based on "the biggest, thickest, strongest skull bones, lower jaws, and tooth crowns from three different quarries". These skulls were likely those of Camarasaurus, the only other sauropod of which good skull material was known at the time. The mount construction was overseen by Adam Hermann, who failed to find Brontosaurus skulls. Hermann was forced to sculpt a stand-in skull by hand. Henry Fairfield Osborn noted in a publication that the skull was "largely conjectural and based on that of Morosaurus" (now Camarasaurus). In 1909, an Apatosaurus skull was found, during the first expedition to what would become the Carnegie Quarry at Dinosaur National Monument, led by Earl Douglass. The skull was found a few meters away from a skeleton (specimen CM 3018) identified as the new species Apatosaurus louisae. The skull was designated CM 11162 and was very similar to the skull of Diplodocus. It was accepted as belonging to the Apatosaurus specimen by Douglass and Carnegie Museum director William J. Holland, although other scientists, most notably Osborn, rejected this identification. Holland defended his view in 1914 in an address to the Paleontological Society of America, yet he left the Carnegie Museum mount headless. While some thought Holland was attempting to avoid conflict with Osborn, others suspected that Holland was waiting until an articulated skull and neck were found to confirm the association of the skull and skeleton. After Holland's death in 1934, a cast of a Camarasaurus skull was placed on the mount by museum staff. Skull correction, resurgent discoveries, and reassessment No apatosaurine skull was mentioned in the literature until the 1970s when John Stanton McIntosh and David Berman redescribed the skulls of Diplodocus and Apatosaurus in 1975. They found that though he never published his opinion, Holland was almost certainly correct in that Apatosaurus and Brontosaurus had a Diplodocus-like skull. According to them, many skulls long thought to belong to Diplodocus might instead be those of Apatosaurus. They reassigned multiple skulls to Apatosaurus based on associated and closely associated vertebrae. Though they supported Holland, Apatosaurus was falsely theorized to possibly have possessed a Camarasaurus-like skull, based on a disarticulated Camarasaurus-like tooth found at the precise site where an Apatosaurus specimen was found years before. However, this tooth does not come from Apatosaurus. On October 20, 1979, after the publications by McIntosh and Berman, the first skull of an Apatosaurus was mounted on a skeleton in a museum, that of the Carnegie. In 1995, the American Museum of Natural History followed suit, and unveiled their remounted skeleton (now labelled Apatosaurus excelsus) with a corrected tail and a new skull cast from A. louisae. In 1998, Robert T. Bakker referred a skull and mandible of an apatosaurine from Como Bluff to Brontosaurus excelsus (TATE 099-01), though the skull is still undescribed. In 2011, the first specimen of Apatosaurus where a skull was found articulated with its cervical vertebrae was described. This specimen, CMC VP 7180, was found to differ in both skull and neck features from A. louisae, and the specimen was found to have a majority of features related to those of A. ajax. Another specimen of an Apatosaurine now referred to Brontosaurus was discovered in 1993 by the Tate Geological Museum, also from the Morrison Formation of central Wyoming. The specimen consisted of a partial postcranial skeleton, including a complete manus and multiple vertebrae, and was described by James Filla and Pat Redman a year later. Filla and Redman named the specimen Apatosaurus yahnahpin ("yahnahpin-wearing deceptive lizard"), but Robert T. Bakker gave it the genus name Eobrontosaurus in 1998. Bakker believed that Eobrontosaurus was the direct predecessor to Brontosaurus, although Tschopp et al.'s phylogenetic analysis placed B. yahnahpin as the basalmost species of Brontosaurus.Almost all 20th-century paleontologists agreed with Riggs that all Apatosaurus and Brontosaurus species should be classified in a single genus. According to the rules of the ICZN, which governs the scientific names of animals, the name Apatosaurus, having been published first, had priority; Brontosaurus was considered a junior synonym and was therefore discarded from formal use. Despite this, at least one paleontologist—Robert T. Bakker—argued in the 1990s that A. ajax and A. excelsus are sufficiently distinct that the latter continues to merit a separate genus. In 2015, an extensive study of diplodocid relationships by Emanuel Tschopp, Octavio Mateus, and Roger Benson concluded that Brontosaurus was indeed a valid genus of sauropod distinct from Apatosaurus. The scientists developed a statistical method to more objectively assess differences between fossil genera and species and concluded that Brontosaurus could be "resurrected" as a valid name. They assigned two former Apatosaurus species, A. parvus, and A. yahnahpin, to Brontosaurus, as well as the type species B. excelsus. The publication was met with some criticism from other paleontologists, including Michael D'Emic, Donald Prothero, who criticized the mass media reaction to this study as superficial and premature, and many others below. Some paleontologists, such as John and Rebecca Foster, continue to consider Brontosaurus as a synonym of Apatosaurus. DescriptionBrontosaurus was a large, long-necked, quadrupedal animal with a long, whip-like tail, and forelimbs that were slightly shorter than its hindlimbs. The largest species, B. excelsus, measured up to long from head to tail and weighed up to ; other species were smaller, measuring long and weighing . The skull of Brontosaurus has not been found but was probably similar to the skull of the closely related Apatosaurus. Several skulls of Apatosaurus have been found, all of which are very small in proportion to the body. Their snouts were squared off and low, in contrast to macronarians'. Jaws of Apatosaurus and other diplodocids were lined with spatulate (chisel-like) teeth which were adapted for herbivory. Vertebrae Like those of other diplodocids, the vertebrae of the neck were deeply bifurcated on the dorsal side; that is, they carried paired spines, resulting in a wide and deep neck. The spine and tail consisted of 15 cervicals, ten dorsals, five sacrals, and about 82 caudals, based on Apatosaurus. The number of caudal vertebrae has been noted to vary, even within a species. Vertebrae in the neck, torso, and sacrum of sauropods bore large pneumatic foramina on their lateral sides. These are used to lighten the bones which aided in keeping the animal lighter. Within the vertebrae as well, smooth bone walls in addition to diverticula would make pockets of air to keep the bones light. Similar structures are observable in birds and large mammals. The cervical vertebrae were stouter than those of other diplodocids, as in Apatosaurus. On the lateral sides of the cervicals, apatosaurines had well-developed and thick parapophyses (extensions on the lateral sides of the vertebrae that attached to cervical ribs) which would point ventrally under the centrum. These parapophyses in conjunction with dense diapophyses and cervical ribs were strong anchors for neck muscles, which could sustain extreme force. The cervicals were also more boxy than in other sauropods due to their truncated zygapophyses and tall build. These vertebrae are triangular in anterior view, whereas they most often are rounded or square in genera like Camarasaurus. Despite its pneumaticy, the neck of Brontosaurus is thought to have been double the mass of that of other diplodocids due to the former’s sturdiness. Brontosaurus differs from Apatosaurus in that the base of the posterior dorsal vertebrae's neural spines are longer than they are wide. The cervicals of species within Brontosaurus also vary, such as the lack of tubercules on the neural spines of B. excelsus and the lateral expansion of unbifurcated neural spines in B. parvus.Its dorsal vertebrae had short centra with large fossae (shallow excavations) on their lateral sides, though not as extensively as the cervicals’. Neural canals, which contain the spinal cord of the vertebral column, are ovate and large in the dorsals. The diapophyses protrude outward and curve downward in a hook-shape. Neural spines are thick in anterior-posterior view with a bifurcate top. The neural spines of the dorsals would increase in height further towards the tail, creating an arched back. Apatosaurine neural spines compose more than half the height of the vertebrae. Medial surfaces of neural spines are gently rounded in B. yahnahpin, whereas in other B. spp. they are not. The dorsal ribs are not fused or tightly attached to their vertebrae, instead being loosely articulated. Ten dorsal ribs are on either side of the body. Expanded excavations within the sacrum are present making it into a hollow cylinder-shape. Sacral neural spines are fused together into a thin plate. The posteriormost caudal vertebra was lightly fused to the sacral vertebrae, becoming part of the plate. Internally, the neural canal was enlarged. The shape of the tail was typical of diplodocids, being comparatively slender, due to the vertebral spines rapidly decreasing in height the farther they are from the hips. As in other diplodocids, the last portion of the tail of Brontosaurus possessed a whip-like structure. The tail also bears an extensive air-sac system to lighten its weight, as observed in specimens of B. parvus. Limbs Several scapulae are known from Brontosaurus, all of which are long and thin with relatively elongated shafts. One of traits that distinguishes Brontosaurus and Apatosaurus is the presence of a depression on the posterior face of the scapula, which the latter lacks. The scapula of Brontosaurus also has a rounded extension off of its edge, a characteristic unique to Brontosaurus among Apatosaurinae. The coracoid anatomy is closely akin to that of Apatosaurus, with a quadratic outline in dorsal view. Sterna have been preserved in some specimens of Brontosaurus, which display an oval outline. The hip bones include robust ilia and the fused pubes and ischia. The limb bones were also very robust, with the humerus resembling that of Camarasaurus, and those of B. excelsus being nearly identical to those of Apatosaurus ajax. The humerus had a thin bone shaft and larger transverse ends. Its anterior end bears a large deltopectoral crest, which was on the extremities of the bone. Charles Gilmore in 1936 noted that previous reconstructions erroneously proposed that the radius and ulna could cross, when in life they would have remained parallel. Brontosaurus had a single large claw on each forelimb which faced towards the body, whereas the rest of the phalanges lacked unguals. Even by 1936, it was recognized that no sauropod had more than one hand claw preserved, and this one claw is now accepted as the maximum number throughout the entire group. The metacarpals are elongated and thinner than the phalanges, bearing boxy articular ends on its proximal and distal faces. The single front claw bone is slightly curved and squarely shortened on the front end. The phalangeal formula is 2-1-1-1-1, meaning the innermost finger (phalanx) on the forelimb has two bones and the next has one. The single manual claw bone (ungual) is slightly curved and squarely truncated on the anterior end. Proportions of the manus bones vary within Apatosaurinae as well, with B. yahnahpin's ratio of longest metacarpal to radius length around 0.40 or greater compared to a lower value in Apatosaurus louisae. The femora of Brontosaurus are very stout and represent some of the most robust femora of any member of Sauropoda. The tibia and fibula bones are different from the slender bones of Diplodocus but are nearly indistinguishable from those of Camarasaurus. The fibula is longer and slenderer than the tibia. The foot of Brontosaurus has three claws on the innermost digits; the digit formula is 3-4-5-3-2. The first metatarsal is the stoutest, a feature shared among diplodocids. B. excelsus astragalus differs from other species in that it lacks a laterally directed ventral shelf. Classification Brontosaurus is a member of the family Diplodocidae, a clade of gigantic sauropod dinosaurs. The family includes some of the longest and largest creatures ever to walk the earth, including Diplodocus, Supersaurus, and Barosaurus. Diplodocids first evolved during the Middle Jurassic in what is now Georgia, spreading to North America during the Late Jurassic. Brontosaurus is classified in the subfamily Apatosaurinae, which also includes Apatosaurus and possibly one or more unnamed genera. Othniel Charles Marsh described Brontosaurus as being allied to Atlantosaurus, within the now defunct group Atlantosauridae. In 1878, Marsh raised his family to the rank of suborder, including Apatosaurus, Brontosaurus, Atlantosaurus, Morosaurus (=Camarasaurus), and Diplodocus. He classified this group within Sauropoda. In 1903, Elmer S. Riggs mentioned that the name Sauropoda would be a junior synonym of earlier names, and grouped Apatosaurus within Opisthocoelia. Most authors still use Sauropoda as the group name. Originally named by its discoverer Othniel Charles Marsh in 1879, Brontosaurus had long been considered a junior synonym of Apatosaurus; its type species, Brontosaurus excelsus, was reclassified as A. excelsus in 1903. However, an extensive study published in 2015 by a joint British-Portuguese research team concluded that Brontosaurus was a valid genus of sauropod distinct from Apatosaurus. Nevertheless, not all paleontologists agree with this division. The same study classified two additional species that had once been considered Apatosaurus and Eobrontosaurus as Brontosaurus parvus and Brontosaurus yahnahpin respectively. Cladogram of the Diplodocidae after Tschopp, Mateus, and Benson (2015): Species Brontosaurus excelsus, the type species of Brontosaurus, was first named by Marsh in 1879. Many specimens have been assigned to the species, such as FMNH P25112, the skeleton mounted at the Field Museum of Natural History, which has since been found to represent an unknown species of apatosaurine. Brontosaurus amplus, is a junior synonym of B. excelsus. B. excelsus therefore only includes its type specimen and the type specimen of B. amplus. The largest of these specimens is estimated to have weighed up to 15 tonnes and measured up to long from head to tail. The known definitive B. excelsus fossils have been reported from Reed's Quarries 10 and 11 of the Morrison Formation Brushy Basin member in Albany County, Wyoming, dated to the late Kimmeridgian age, about 152 million years ago. Brontosaurus parvus, first described as Elosaurus in 1902 by Peterson and Gilmore, was reassigned to Apatosaurus in 1994, and to Brontosaurus in 2015. Specimens assigned to this species include the holotype, CM 566 (a partial skeleton of a juvenile found in Sheep Creek Quarry 4 in Albany County, WY), BYU 1252-18531 (a nearly complete skeleton found in Utah and mounted at Brigham Young University), and the partial skeleton UW 15556. It dates to the middle Kimmeridgian. Adult specimens are estimated to have weighed up to 14 tonnes and measured up to long from head to tail. Brontosaurus yahnahpin is the oldest species, known from a single site from the lower Morrison Formation, Bertha Quarry, in Albany County, Wyoming, dating to about 155 million years ago. It grew up to long. It was described by James Filla and Patrick Redman in 1994 as a species of Apatosaurus (A. yahnahpin). The specific name is derived from Lakota mah-koo yah-nah-pin, "breast necklace", a reference to the pairs of sternal ribs that resemble the hair pipes traditionally worn by the tribe. The holotype specimen is TATE-001, a relatively complete postcranial skeleton found in the lower Morrison Formation of Wyoming. More fragmentary remains have also been referred to the species. A re-evaluation by Robert T. Bakker in 1998 found it to be more primitive, so Bakker coined the new generic name Eobrontosaurus, derived from Greek , "dawn", and Brontosaurus. The cladogram below is the result of an analysis by Tschopp, Mateus, and Benson (2015). The authors analyzed most diplodocid type specimens separately to deduce which specimen belonged to which species and genus. Palaeobiology When Brontosaurus was described in 1879, the widespread notion in the scientific community was that sauropods were semi-aquatic, lethargic reptiles that were inactive. In Othniel Marsh's publication The Dinosaurs of North America, he described the dinosaur as "more or less amphibious, and its food was probably aquatic plants or other succulent vegetation". This is unsupported by fossil evidence. Instead, sauropods were active and had adaptations for dwelling on land. Marsh also noted the animal's supposed lack of intellect based on the small braincase of the Felch Quarry skull and slender neural cord. Recent research has found signs of intelligence in dinosaurs, akin to modern birds, though sauropods had relatively small brains. Various uses for the single claw on the forelimb of sauropods have been proposed. One suggestion is that they were used for defense, but their shape and size make this unlikely. It was also possible they were for foraging, but the most probable use for the claw was grasping objects such as tree trunks when rearing. Trackways of sauropods like Brontosaurus show that the average range for them was around per day, and they could potentially reach a top speed of . The slow locomotion of sauropods may be due to the minimal muscling or recoil after strides. A possible bipedal trackway of a juvenile Apatosaurus is known, but it is disputed if it was possible for the sauropod. Diet and energy requirements Being a diplodocid sauropod, Brontosaurus was herbivorous and fed on ferns, cycadeoids, seed ferns, and horsetails, eating at ground height as a nonselective browser. The replacement method and physiology of Apatosaurus''' teeth is unique, with the entire tooth row being replaced at once and up to 60% more often than Diplodocus. The teeth of Apatosaurus are thick, lack denticles, and are strongly cylindrical in cross-section whereas they are long, slender, and elliptical in cross-section in Diplodocus. These characteristics imply that Apatosaurus, and likely Brontosaurus, consumed tougher vegetation than Diplodocus. Diplodocids in general also have shorter necks than the long-necked, vertically inclined macronarians. This would result in niche partitioning, the various taxa thus avoiding direct competition with each other due to feeding on different plants and at different heights. Hypotheses of the food requirements of Brontosaurus have been made, though predicting this is difficult due to the lack of modern analogues. Endotherms (mammals) and ectotherms (reptiles) require a specific amount of nutrition to survive which correlates with their metabolism as well as body size. Estimations of the dietary necessities of Brontosaurus were made in 2010, with a guess of 2•10^4 to 50•10^4 kilojoules needed daily. This led to hypotheses on the distributions of Brontosaurus to meet this requirement, though they varied on whether it was an ectotherm or endotherm. If Brontosaurus was an endotherm, fewer adult individuals could be sustained than if it were an ectotherm, which could have tens of animals per square kilometer.Farlow JO (1976) A consideration of the trophic dynamics of a Late Cretaceous large-dinosaur community (Oldman formation). Ecology 57:841–857 Due to this, it has been theorized that Brontosaurus and other sauropods living within the arid environment of the Morrison Formation participated in migrations between feeding sites. James Farlow (1987) calculates that a Brontosaurus-sized dinosaur about would have possessed of fermentation contents. Assuming Apatosaurus had an avian respiratory system and a reptilian resting-metabolism, Frank Paladino etal. (1997) estimate the animal would have needed to consume only about of water per day. Posture Historically, sauropods like Brontosaurus were believed to have been too massive to support their weight on dry land, so theoretically, they must have lived partly submerged in water, perhaps in swamps. Recent findings do not support this, and sauropods are thought to have been fully terrestrial animals. Diplodocids like Brontosaurus are often portrayed with their necks held high up in the air, allowing them to browse on tall trees. Though some studies have suggested that diplodocid necks were less flexible than previously believed, other studies have found that all tetrapods appear to hold their necks at the maximum possible vertical extension when in a normal, alert posture, and argue that the same would hold true for sauropods barring any unknown, unique characteristics that set the soft tissue anatomy of their necks apart from that of other animals. Physiology James Spotila et al. (1991) suggest that the large body size of Brontosaurus and other sauropods would have made them unable to maintain high metabolic rates, as they would not be able to release enough heat. However, temperatures in the Jurassic were 3 degrees Celsius higher than present. Furthermore, they assumed that the animals had a reptilian respiratory system. Matt Wedel found that an avian system would have allowed them to dump more heat. Some scientists have also argued that the heart would have had trouble sustaining sufficient blood pressure to oxygenate the brain. Given the large body mass and long neck of sauropods like Brontosaurus, physiologists have encountered problems determining how these animals breathed. Beginning with the assumption that, like crocodilians, Brontosaurus did not have a diaphragm, the dead-space volume (the amount of unused air remaining in the mouth, trachea, and air tubes after each breath) has been estimated at for a specimen. Paladino calculates its tidal volume (the amount of air moved in or out during a single breath) at with an avian respiratory system, if mammalian, and if reptilian. Based on this, its respiratory system would likely have consisted of parabronchi, with multiple pulmonary air sacs as in avian lungs, and a flow-through lung. An avian respiratory system would need a lung volume of about compared with a mammalian requirement of , which would exceed the space available. The overall thoracic volume of the same-sized Apatosaurus has been estimated at , allowing for a , four-chambered heart and a lung capacity. That would allow about for the necessary tissue. Evidence for the avian system in Brontosaurus and other sauropods is also present in the pneumaticity of the vertebrae. Though this plays a role in reducing the weight of the animal, Wedel (2003) states they are also likely connected to air sacs, as in birds. Juveniles A 1999 microscopic study of Apatosaurus and Brontosaurus bones concluded the animals grew rapidly when young and reached near-adult sizes in about 10years. In 2008, a study on the growth rates of sauropods was published by biologists Thomas Lehman and Holly Woodward. They said that by using growth lines and length-to-mass ratios, Apatosaurus would have grown to 25t (25 long tons; 28 short tons) in 15years, with growth peaking at in a single year. An alternative method, using limb length and body mass, found Brontosaurus and Apatosaurus grew per year, and reached their full mass before it was about 70years old. These estimates have been called unreliable because the calculation methods are not sound; old growth lines would have been obliterated by bone remodeling. One of the first identified growth factors of Apatosaurus was the number of sacral vertebrae, which increased to five by the time of the creature's maturity. This was first noted in 1903 and again in 1936. Juvenile Brontosaurus material is known based on the type specimen of B. parvus. The material of this specimen, CM 566, includes vertebrae from various regions, one pelvic bone, and some bones of the hindlimb. When describing B. parvus, Peterson and Gilmore noted that the neural spines were sutured, the sacral vertebrae were unfused, and the coracoid was missing. All of these features are signs of immaturity in other archosaurs, showing that sauropods had these traits too. Peterson and Gilmore also theorized that sauropods never stopped growing, which supposedly helped in attaining their massive size, a concept unsupported by fossils. Tail An article that appeared in the November 1997 issue of Discover magazine reported research into the mechanics of diplodocid tails by Nathan Myhrvold, a computer scientist from Microsoft. Myhrvold carried out a computer simulation of the tail, which in diplodocids like Brontosaurus was a very long, tapering structure resembling a bullwhip. This computer modeling suggested that sauropods were capable of producing a whip-like cracking sound of over 200 decibels, comparable to the volume of a cannon. There is some circumstantial evidence supporting this as well: a number of diplodocids have been found with fused or damaged tail vertebrae, which may be a symptom of cracking their tails: these are particularly common between the 18th and the 25th caudal vertebra, a region the authors consider a transitional zone between the stiff muscular base and the flexible whiplike section. However, Rega (2012) notes that Camarasaurus while lacking a tailwhip, displays a similar level of caudal co-ossification and that Mamenchisaurus while having the same pattern of vertebral metrics, lacks a tailwhip and does not display fusion in any "transitional region". Also, the crush fractures which would be expected if the tail was used as a whip have never been found in diplodocids. More recently, Baron (2020) has considered the use of the tail as a bullwhip unlikely because of the potentially catastrophic muscle and skeletal damage such speeds could cause on the large and heavy tail. Instead, he proposes that the tails might have been used as a tactile organ to keep in touch with the individuals behind and to the sides of the animal in a group, which could have augmented cohesion and allowed communication among individuals while limiting more energetically demanding activities like stopping to search for dispersed individuals, turning to visually check on others behind, or communicating vocally. Neck combat The cervical vertebrae of Brontosaurus and Apatosaurus are robust, which has led to speculation on the use of these structures. These structures had expensive energy requirements, so the reason for their evolution must have been important to the animal. Notable features include dense cervical ribs and diapophyses, ribs that are angled ventrally, and an overall subtriangular cross-section. These traits are in contrast to the more fragile cervicals of diplodocines. Cervical ribs acted as anchors for the longus colli ventralis and flexer colli lateralis muscles, which are used in the downward motion of the neck. Stronger muscles for ventral motions allowed more force to be exerted downward. The cervical ribs formed a "V"-shape, which could be used to shelter the softer underlying tissues of the neck from damage. Ventral sides of the cervical ribs were capped by round, protruding processes. These have been suggested to have been attachment points for bosses or keratinous spikes. A preprint by Wedel et al (2015) thought that due to the combination of these traits, Brontosaurus would use its neck for combat between individuals through the use of striking necks. Behavior like this has been observed in other animals like giraffes and large tortoises. Paleoecology The Morrison Formation is a sequence of shallow marine and alluvial sediments which, according to radiometric dating, ranges between 156.3 million years old (Mya) at its base, and 146.8 Mya at the top, which places it in the late Oxfordian, Kimmeridgian, and early Tithonian stages of the Late Jurassic period. This formation is interpreted as a semiarid environment with distinct wet and dry seasons. The Morrison Basin, where dinosaurs lived, stretched from New Mexico to Alberta and Saskatchewan and was formed when the precursors to the Front Range of the Rocky Mountains started pushing up to the west. The deposits from their east-facing drainage basins were carried by streams and rivers and deposited in swampy lowlands, lakes, river channels, and floodplains. This formation is similar in age to the Lourinhã Formation in Portugal and the Tendaguru Formation in Tanzania.Brontosaurus may have been a more solitary animal than other Morrison Formation dinosaurs. As a genus, Brontosaurus existed for a long interval, and was found in most levels of the Morrison. B. excelsus fossils have been reported from only the Brushy Basin Member, dating to the late Kimmeridgian age, about 151 Mya. Older Brontosaurus remains have also been identified from the middle Kimmeridgian, and are assigned to B. parvus. Fossils of these animals have been found in Nine Mile Quarry and Bone Cabin Quarry in Wyoming and at sites in Colorado, Oklahoma, and Utah, present in stratigraphic zones 2–6 according to John Foster’s model. The Morrison Formation records an environment and time dominated by gigantic sauropod dinosaurs. Dinosaurs known from the Morrison include the theropods Ceratosaurus, Ornitholestes, and Allosaurus, the sauropods Apatosaurus, Brachiosaurus, Camarasaurus, and Diplodocus, and the ornithischians Camptosaurus, Dryosaurus, and Stegosaurus. Other vertebrates that shared this paleoenvironment included ray-finned fishes, frogs, salamanders, turtles, sphenodonts, lizards, terrestrial and aquatic crocodylomorphs, and several species of pterosaurs. Shells of bivalves and aquatic snails, are also common. The flora of the period has been revealed by fossils of green algae, mosses, horsetails, cycads, ginkgoes, and several families of conifers. Vegetation varied from river-lining forests of tree ferns and ferns (gallery forests), to fern savannas with occasional trees such as the Araucaria-like conifer Brachyphyllum. In popular culture The length of time taken for Riggs's 1903 reclassification of Brontosaurus as Apatosaurus to be brought to public notice, as well as Osborn's insistence that the Brontosaurus name be retained despite Riggs's paper, meant that Brontosaurus became one of the most famous dinosaurs. Brontosaurus has often been depicted in cinema, beginning with Winsor McCay's 1914 classic Gertie the Dinosaur, one of the first animated films. McCay based his unidentified dinosaur on the apatosaurine skeleton in the American Museum of Natural History. The 1925 silent film The Lost World featured a battle between a Brontosaurus and an Allosaurus, using special effects by Willis O'Brien. The 1933 film King Kong featured a Brontosaurus chasing Carl Denham, Jack Driscoll and the terrified sailors on Skull Island. In 1938 the assembling of a Brontosaurus skeleton was a major plot point in the Katharine Hepburn and Cary Grant film Bringing Up Baby. These, and other early uses of the animal as a major representative of the group, helped cement Brontosaurus as a quintessential dinosaur in the public consciousness. Sinclair Oil Corporation has long been a fixture of American roads (and briefly in other countries) with its green dinosaur logo and mascot, a Brontosaurus. While Sinclair's early advertising included a number of different dinosaurs, eventually only Brontosaurus was used as the official logo, due to its popular appeal. As late as 1989, the U.S. Postal Service faced controversy when it issued four "dinosaur" stamps: Tyrannosaurus, Stegosaurus, Pteranodon, and Brontosaurus. The use of the term Brontosaurus in place of Apatosaurus led to complaints of "fostering scientific illiteracy." The Postal Service defended itself (in Postal Bulletin 21742) by saying, "Although now recognized by the scientific community as Apatosaurus, the name Brontosaurus was used for the stamp because it is more familiar to the general population." Indeed, the Postal Service even implicitly rebuked the somewhat inconsistent complaints by adding that "[s]imilarly, the term 'dinosaur' has been used generically to describe all the animals [i.e., all four of the animals represented in the given stamp set], even though the Pteranodon was a flying reptile [rather than a true 'dinosaur']," a distinction left unmentioned in the numerous correspondence regarding the Brontosaurus/Apatosaurus issue. Palaeontologist Stephen Jay Gould supported this position. In the essay from which the title of the 1991 collection Bully for Brontosaurus is taken, Gould wrote: "Touché and right on; no one bitched about Pteranodon, and that's a real error." His position, however, was not one suggesting the exclusive use of the popular name; he echoed Riggs' original argument that Brontosaurus is a synonym for Apatosaurus. Nevertheless, he noted that the former has developed and continues to maintain an independent existence in the popular imagination. The more vociferous denunciations of the usage have elicited sharply defensive statements from those who would not wish to see the name be struck from official usage. Tschopp's study has generated a very high number of responses from many, often opposed, of editorial, news staff, and personal blog nature (both related and not), from both sides of the debate, from related and unrelated contexts, and from all over the world. Since Wedel et al 2015 preprint, various reconstructions of Brontosaurus'' individuals engaging in intraspecific combat based on their study have been made. The art typically depicts the neck-battling hypothesis stipulated by their research. Many of these works are published online under the hashtag "#BrontoSmash".
Biology and health sciences
Sauropods
Animals
3636124
https://en.wikipedia.org/wiki/Common%20wheat
Common wheat
Common wheat (Triticum aestivum), also known as bread wheat, is a cultivated wheat species. About 95% of wheat produced worldwide is common wheat; it is the most widely grown of all crops and the cereal with the highest monetary yield. Taxonomy Numerous forms of wheat have evolved under human selection. This diversity has led to confusion in the naming of wheats, with names based on both genetic and morphological characteristics. List of common cultivars Albimonte Manital Shirley Hilliard Phylogeny Bread wheat is an allohexaploid a combination of six sets of chromosomes from different species. Of the six sets of chromosomes, four come from emmer (Triticum turgidum, itself a tetraploid) and two from Aegilops tauschii (a wild diploid goatgrass). Wild emmer arose from an even earlier ploidy event, a tetraploidy between two diploids, wild einkorn (T. urartu) and A. speltoides (another wild goatgrass). Free-threshing wheat is closely related to spelt. As with spelt, genes contributed from Aegilops tauschii give bread wheat greater cold hardiness than most wheats, and it is cultivated throughout the world's temperate regions. Cultivation History Common wheat was first domesticated in West Asia during the early Holocene, and spread from there to North Africa, Europe and East Asia in the prehistoric period. Naked wheats (including Triticum aestivum, T. durum, and T. turgidum) were found in Roman burial sites ranging from 100 BCE to 300 CE. Wheat first reached North America with Spanish missions in the 16th century, but North America's role as a major exporter of grain dates from the colonization of the prairies in the 1870s. As grain exports from Russia ceased during World War I, grain production in Kansas doubled. Worldwide, bread wheat has proved well adapted to modern industrial baking, and has displaced many of the other wheat, barley, and rye species that were once commonly used for bread making, particularly in Europe. Plant breeding Modern wheat varieties have been selected for short stems, the result of RHt dwarfing genes that reduce the plant's sensitivity to gibberellic acid, a plant hormone that lengthens cells. RHt genes were introduced to modern wheat varieties in the 1960s by Norman Borlaug from Norin 10 cultivars of wheat grown in Japan. Short stems are important because the application of high levels of chemical fertilizers would otherwise cause the stems to grow too high, resulting in lodging (collapse of the stems). Stem heights are also even, which is important for modern harvesting techniques. Other forms of common wheat Compact wheats (e.g., club wheat Triticum compactum, but in India T. sphaerococcum) are closely related to common wheat, but have a much more compact ear. Their shorter rachis segments lead to spikelets packed closer together. Compact wheats are often regarded as subspecies rather than species in their own right (thus T. aestivum subsp. compactum).
Biology and health sciences
Poales
null
13065889
https://en.wikipedia.org/wiki/Combat%20helmet
Combat helmet
A combat helmet or battle helmet is a type of helmet designed to serve as a piece of personal armor intended to protect the wearer's head during combat. Modern combat helmets are mainly designed to protect from shrapnel and fragments, offer some protection against small arms, and offer a mounting point for devices such as night-vision goggles and communications equipment. History Helmets are among the oldest forms of personal protective equipment and are known to have been worn by the Akkadians/Sumerians in the 23rd century BC, Mycenaean Greeks since the 17th century BC, the Assyrians around 900 BC, ancient Greeks and Romans, throughout the Middle Ages, and up to the end of the 17th century by many combatants. Their materials and construction became more advanced as weapons became more and more powerful. Initially constructed from leather and brass, and then bronze and iron during the Bronze and Iron Ages, they soon came to be made entirely from forged steel in many societies after about 950 AD. At that time, they were purely military equipment, protecting the head from cutting blows with swords, flying arrows, and low-velocity musketry. Iron helmets were deployed into the cavalry of the Mali Empire to protect the cavalrymen and their mount. Military use of helmets declined after 1670, and rifled firearms ended their use by foot soldiers after 1700 but the Napoleonic era saw ornate cavalry helmets reintroduced for cuirassiers and dragoons in some armies which continued to be used by French forces during World War I as late as 1915. During the French Revolutionary Wars and the Napoleonic Wars, the Austrian Imperial Army saw extensive usage of helmets. In the line infantry, mainly within the fusilier companies, helmets were worn from 1798 to 1806, which was true even for officers. Although they were officially replaced by the shako in 1806, most line infantry regiments continued to wear helmets up until the Austrian defeat at Battle of Wagram in July 1809. Dragoons and cuirassiers also wore the helmets more extensively than the line infantry, with them continuing to wear them well past the Napoleonic Wars. World War I and its increased use of artillery renewed the need for steel helmets, with the French Adrian helmet and the British Brodie helmet being the first modern steel helmets used on the battlefield, soon followed by the adoption of similar steel helmets, such as the Stahlhelm by the other warring nations. Such helmets offered protection for the head from shrapnel and fragments. Today's militaries often use high quality helmets made of ballistic materials such as Kevlar and Twaron, which offer improved protection. Some helmets also have good non-ballistic protective qualities, against threats such as concussive shock waves from explosions. Many of today's combat helmets have been adapted for modern warfare requirements and upgraded with STANAG rails to act as a platform for mounting cameras, video cameras and VAS Shrouds for the mounting of night-vision devices. Beginning in the early 20th century, combat helmets have often been equipped with helmet covers to offer greater camouflage. There have been two main types of covers—mesh nets were earlier widely used, but most modern combat helmets use camouflage cloth covers instead. By the late 20th century, starting in the 1970s and 1980s, new materials such as Kevlar and Twaron began replacing steel as the primary material for combat helmets, in an effort to improve weight reduction and ballistic protection, and protection against traumatic brain injury. This practice still continues into the 21st century, with further advancement and refinements in the fibers used, design and shape of the helmet, and increased modularity. Early helmet systems of this new design are the American PASGT, the Spanish MARTE, the Italian SEPT-2 PLUS, and British Mk 6. Padding Cushioning is used to negate concussive injuries. Researchers at the Lawrence Livermore National Laboratory published a study in 2011 that concluded that the addition of of cushion decreased the impact force to the skull by 24%.
Technology
Armour
null
13068017
https://en.wikipedia.org/wiki/Surveyor%27s%20wheel
Surveyor's wheel
A surveyor's wheel, also called a clickwheel, hodometer, waywiser, trundle wheel, measuring wheel or perambulator is a device for measuring distance. Origin The origins of the surveyor's wheel are connected to the origins of the odometer. While the latter is derived to measure distances travelled by a vehicle, the former is specialized to measure distances. In the 17th century, the surveyor's wheel was re-introduced and used to measure distances. A single wheel is attached to a handle and the device can be pushed or pulled along by a person walking. Early devices were made of wood and may have had an iron rim to provide strength. The wheels themselves would be made in the same manner as wagon wheels and often by the same makers. The measuring devices would be made by makers of scientific instruments and the device and handles would be attached to the wheel by them. The device to read the distance travelled would be mounted either near the hub of the wheel or at the top of the handle. In some cases, double-wheel hodometers were constructed. Francis Ronalds extended the concept in 1827 to create a device that recorded the distances travelled in graphical form as a survey plan. The apparatus had a worm on the axle of the two wheels that meshed with a toothed wheel to drive another transverse screw that carried a slider. A pencil on the slider recorded the distance travelled along the screw on an attached drawing board at a chosen scale. Modern surveyor's wheels are constructed primarily of aluminium, with solid or pneumatic tyres on the wheel. Some can fold for transport or storage. Principle The surveyor's wheel is marked in fractional increments of revolution from a reference position. Thus its current position can be represented as a fraction of a revolution from this reference. If the wheel rotated a full turn (360 angular degrees), the distance traveled would be equal to the circumference of the wheel. Otherwise, the distance the wheel traveled is the circumference of the wheel multiplied by the fraction of a full turn. In the figure on the right, the blue line is the reference starting point. As the wheel turns during measurement, it is seen that the wheel sweeps out an angle of radians, which is equal to 135 degrees or of a full turn. Usage Each revolution of the wheel measures a specific distance, such as a yard, meter or half-rod. Thus counting revolutions with a mechanical device attached to the wheel measures the distance directly. Surveyor's wheels will provide a measure of good accuracy on a smooth surface, such as pavement. On rough terrain, wheel slippage and bouncing can reduce the accuracy. Soft sandy or muddy soil can also affect the rolling of the wheel. As well, obstacles in the way of the path may have to be accounted for separately. Good surveyors will keep track of any circumstance on the path that can influence the accuracy of the distance measured and either measure that portion with an alternative, such as a surveyor's tape or measuring tape, or make a reasonable estimate of the correction to apply. Surveyor's wheels are used primarily for lower accuracy surveys. They are often used by road maintenance or underground utility workers and by farmers for fast measures over distances too inconvenient to measure with a surveyor's tape. The surveyor's wheel measures the distance along a surface, whereas in normal land surveying, distances between points are usually measured horizontally with vertical measurements indicated in differences in elevation. Thus conventionally surveyed distances will be less than those measured by a surveyor's wheel. Trundle wheel The trundle wheel is a simplified form of a surveyor's wheel. It is commonly used by people who need an easy way to find the rough distance from one place to another. The trundle wheel is composed of a wheel, a handle which is attached to the axle allowing the trundle wheel to be held easily, and a clicking device which is triggered once per revolution of the wheel. Trundle wheels are not as accurate as other methods of measuring distance but are a good way to get a rough estimation of a fairly long distance over a good surface. It works by having a wheel which has a circumference of exactly 1 metre (or other known unit of measure), hence one revolution of the wheel equates to 1 unit of distance traveled on the ground if there is no slip. Every time the wheel makes a rotation, the wheel produces an audible click which is then counted and therefore the number of clicks that are counted by the user is approximately the number of distance units traveled. Due to the design of the trundle wheel, it has the potential to not always travel in a straight line which may add extra distance to the final reading.
Technology
Surveying tools
null
13069246
https://en.wikipedia.org/wiki/Atmosphere%20of%20Mercury
Atmosphere of Mercury
Mercury, being the closest to the Sun, with a weak magnetic field and the smallest mass of the recognized terrestrial planets, has a very tenuous and highly variable atmosphere (surface-bound exosphere) containing hydrogen, helium, oxygen, sodium, calcium, potassium and water vapor, with a combined pressure level of about 10−14 bar (1 nPa). The exospheric species originate either from the Solar wind or from the planetary crust. Solar light pushes the atmospheric gases away from the Sun, creating a comet-like tail behind the planet. The existence of a Mercurian atmosphere was contentious until 1974, although by that time a consensus had formed that Mercury, like the Moon, lacked any substantial atmosphere. This conclusion was confirmed in 1974 when the unmanned Mariner 10 spaceprobe discovered only a tenuous exosphere. Later, in 2008, improved measurements were obtained by the MESSENGER spacecraft, which discovered magnesium in the Mercurian exosphere. Composition The Mercurian exosphere consists of a variety of species originating either from the Solar wind or from the planetary crust. The first constituents discovered were atomic hydrogen (H), helium (He) and atomic oxygen (O), which were observed by the ultraviolet radiation photometer of the Mariner 10 spaceprobe in 1974. The near-surface concentrations of these elements were estimated to vary from 230 cm−3 for hydrogen to 44,000 cm−3 for oxygen, with an intermediate concentration of helium. In 2008 the MESSENGER probe confirmed the presence of atomic hydrogen, although its concentration appeared higher than the 1974 estimate. Mercury's exospheric hydrogen and helium are believed to come from the Solar wind, while the oxygen is likely to be of crustal origin. The fourth species detected in Mercury's exosphere was sodium (Na). It was discovered in 1985 by Drew Potter and Tom Morgan, who observed its Fraunhofer emission lines at 589 and 589.6 nm. The average column density of this element is about 1 cm−2. Sodium is observed to concentrate near the poles, forming bright spots. Its abundance is also enhanced near the dawn terminator as compared to the dusk terminator. Some research has claimed a correlation of the sodium abundance with certain surface features such as Caloris or radio bright spots; however these results remain controversial. A year after the sodium discovery, Potter and Morgan reported that potassium (K) is also present in the exosphere of Mercury, though with a column density two orders of magnitude lower than that of sodium. The properties and spatial distribution of these two elements are otherwise very similar. In 1998 another element, calcium (Ca), was detected with column density three orders of magnitude below that of sodium. Observations by the MESSENGER probe in 2009 showed that calcium is concentrated mainly near the equator—opposite to what is observed for sodium and potassium. Further observations by Messenger reported in 2014 note the atmosphere is supplemented by materials vaporized off the surface by meteors both sporadic and in a meteor shower associated with Comet Encke. In 2008 the MESSENGER probe's Fast Imaging Plasma Spectrometer (FIPS) discovered several molecular and different ions in the vicinity of Mercury, including H2O+ (ionized water vapor) and H2S+ (ionized hydrogen sulfide). Their abundances relative to sodium are about 0.2 and 0.7, respectively. Other ions such as H3O+ (hydronium), OH (hydroxyl), O2+ and Si+ are present as well. During its 2009 flyby, the Ultraviolet and Visible Spectrometer (UVVS) channel of the Mercury Atmospheric and Surface Composition Spectrometer (MASCS) on board the MESSENGER spacecraft first revealed the presence of magnesium in the Mercurian exosphere. The near-surface abundance of this newly detected constituent is roughly comparable to that of sodium. Properties Mariner 10 ultraviolet observations have established an upper bound on the exospheric surface density at about 105 particles per cubic centimeter. This corresponds to a surface pressure of less than 10−12 bar (1 nPa). The temperature of Mercury's exosphere depends on species as well as geographical location. For exospheric atomic hydrogen, the temperature appears to be about 420 K, a value obtained by both Mariner 10 and MESSENGER. The temperature for sodium is much higher, reaching 750–1,500 K on the equator and 1,500–3,500 K at the poles. Some observations show that Mercury is surrounded by a hot corona of calcium atoms with temperature between 12,000 and 20,000 K. In the early 2000s, a simulation of Mercury's Na exosphere and its temporal variation was conducted to identify the source process that supplied crustal species to the exosphere. Processes like; evaporation, diffusion from the interior, sputtering by photons and energetic ions, chemical sputtering by photons, and meteoritic vaporization were tested. However, evaporation provides the strongest match when comparing the changes in the sodium exosphere with solar distance and time of day to the 2001 observations of Mercury's sodium tail. Tails Because of Mercury's proximity to the Sun, the pressure of solar light is much stronger than near Earth. Solar radiation pushes neutral atoms away from Mercury, creating a comet-like tail behind it. The main component in the tail is sodium, which has been detected beyond 24 million km (1000 RM) from the planet. This sodium tail expands rapidly to a diameter of about 20,000 km at a distance of 17,500 km. In 2009, MESSENGER also detected calcium and magnesium in the tail, although these elements were only observed at distances less than 8 RM. Observation difficulties Mercury is the least explored planet of the inner Solar System due to the many difficulties of observation. The position of Mercury as seen from Earth is always very close to the Sun, which causes challenges when trying to observe it. The Hubble Space Telescope and other Earth-based space imaging systems have highly sensitive sensors so they can observe deep space objects. They must not be directed toward the Sun, lest its powerful radiation destroy the sensors. Instead, flyby and orbital missions to Mercury can study the planet and receive accurate data. Even though Mercury is closer to Earth than Pluto is, the transfer orbit from Earth to Mercury requires more energy. Mercury being so close to the Sun, space probes going there are accelerating as they approach, due to the Sun's gravitational pull. This requires the use of retrorockets, which use fuel that the probe must carry instead of better instruments.
Physical sciences
Solar System
Astronomy
290749
https://en.wikipedia.org/wiki/Conch
Conch
Conch ( , ) is a common name of a number of different medium-to-large-sized sea snails. Conch shells typically have a high spire and a noticeable siphonal canal (in other words, the shell comes to a noticeable point on both ends). Conches that are sometimes referred to as "true conches" are marine gastropods in the family Strombidae, specifically in the genus Strombus and other closely related genera. For example, Lobatus gigas, the queen conch, is a true conch. True conch are identified by their long spire. Many other species are also often called "conch", but are not at all closely related to the family Strombidae, including Melongena species (family Melongenidae) and the horse conch Triplofusus papillosus (family Fasciolariidae). Species commonly referred to as conches also include the sacred chank or shankha shell (Turbinella pyrum) and other Turbinella species in the family Turbinellidae. The Triton's trumpet (family Charoniidae) may also be fashioned into a horn and referred to as a conch. Etymology The English word "conch" is attested in Middle English, coming from Latin (shellfish, mussel), which in turn comes from Greek (same meaning) ultimately from Proto-Indo-European root , cognate with Sanskrit word . General description Conch are species of sea snail in the phylum Mollusca. Their shells consist of about 95% calcium carbonate and 5% organic matter. Conch are harvested for their edible meat and decorative shell. The shells are often used for decoration and as a musical instrument. Culinary use The meat of conches are often eaten raw in salads or cooked in burgers, chowders, fritters, and gumbos. Conch is indigenous to the Caribbean and West Indies. Conch is particularly popular in the Bahamas, Turks and Caicos, and Jamaica. In Bahamas, conch is often cooked into fritters. In Jamaica conch is eaten in stews and curries. In the Dominican Republic, Grenada, and Haiti, conch is commonly eaten in curries or in a spicy soup. It is locally referred to as lambi. In Puerto Rico, conch is served as a ceviche, often called ensalada de carrucho (conch salad), consisting of raw conch marinated in lime juice, olive oil, vinegar, garlic, green peppers, and onions. It is also used to fill empanadas. In Panama, conch is known as cambombia and is often served as ceviche de cambombia, consisting of raw conch marinated in lime juice, chopped onions, finely chopped habaneros, and often vinegar. Conch is very popular in Italy and among Italian Americans. Called sconcigli, it is eaten in a variety of ways, but most often in salads or cooked in a sauce for pasta. It is often included as one of the dishes prepared for the Feast of the Seven Fishes. In East Asian cuisines, conch is often cut into thin slices and then steamed or stir-fried. Eighty per cent of the queen conch meat in international trade is imported into the United States. The Florida Keys were a major source of queen conches until the 1970s, but the conches are now scarce and all harvesting of them in Florida waters is prohibited and individuals who have harvested them have been punished by law enforcement. Festival In the Turks and Caicos Islands, the Conch Festival is held in November each year at the Three Queens Bar/Restaurant in Blue Hills. Local restaurateurs compete for the best and most original conch dishes, which are then judged by international chefs. Other competitions, events, and music performances occur. Musical instruments Conch shells can be used as wind instruments. They are prepared by cutting a hole in the spire of the shell near the apex and then blowing into the shell as if it were a trumpet, as a blowing horn. Sometimes a mouthpiece is used, but some shell trumpets are blown without one. Pitch is adjusted by moving one's hand in and out of the aperture; the deeper the hand, the lower the note. Various species of large marine gastropod shells can be turned into blowing shells, but some of the best-known species used are the sacred chank or shankha Turbinella pyrum, the Triton's trumpet Charonia tritonis, and the queen conch Strombus gigas. One of the most famous musical instruments was found in the Marsoulas cave in the Pyrenees Mountains, in France, in 1932. CT scans showed how ancient humans adapted the Concho to make it a musical instrument, such as creating a mouthpiece that was held together by an organic matter like clay or wax. Researchers from the Sorbonne, together with a professional horn player, were able to use it again as a musical instrument and play it. Examples of this practice in the Americas can be seen in the form of historical artifacts at the Museo Larco in Lima, Peru, and Museo Nacional de Antropología in Mexico City, Mexico. Pearls Many kinds of molluscs can produce pearls. Pearls from the queen conch, S. gigas, are rare and have been collectors' items since Victorian times. Conch pearls occur in a range of hues, including white, brown, and orange, with many intermediate shades, but pink is the colour most associated with the conch pearl, such that these pearls are sometimes referred to simply as "pink pearls". In some gemological texts, non-nacreous gastropod pearls used to be referred to as "calcareous concretions" because they were porcellaneous (shiny and ceramic-like in appearance), rather than nacreous (with a pearly luster). The Gemological Institute of America and World Jewellery Confederation now use the simple term "pearl"—or, where appropriate, the more-descriptive term "non-nacreous pearl"—for such items, and, under Federal Trade Commission rules, various mollusk pearls may be referred to as "pearls" without qualification. Although not nacreous, the surfaces of fine conch pearls have a unique appearance. The microstructure of conch pearls comprises partly aligned bundles of microcrystalline fibers that create a shimmering, slightly iridescent effect known as flame structure. The effect is a form of chatoyancy, caused by the interaction of light rays with the microcrystals in the pearl's surface, and it somewhat resembles moiré silk. Other uses Conch shells are used as biologically grown calcium carbonate fertilizer. Conch shells are sometimes used as decoration, as decorative planters, and in cameo making. In the Aztec culture, the conch played an important role in rituals, war, art, music, mythology, festivals, and even the calendar. In India, some artisans make souvenirs, deity idols and other crafts by carving natural conch shells by hands. Conch shells have been used as shell money in several cultures. Some American Aboriginals used cylindrical conch columella beads as part of breastplates and other personal adornment. In India, the Bengali bride-to-be is adorned with conch shell and coral bangles called shakha paula. It is a traditional wedding ritual for every Bengali bride. In India and Bangladesh, the conch is blown every day in the evening in Bengali houses as a daily ritual. In some Afro-Caribbean and African-American cemeteries, conch shells are placed on graves. In some Caribbean countries such as Jamaica and the Bahamas, cleaned queen conch shells, or polished fragments, are sold, mainly to tourists, as souvenirs or in jewellery. Responding to a 2003 recommendation from CITES, some countries in the Caribbean have banned the export of queen conch shells. CITES has also asked all countries to ban import of these shells from countries that are not complying with CITES recommendations for managing the fishery. Queen conch fisheries have been closed in several countries. Conch shells or fragments taken home by tourists from noncomplying countries may be confiscated on return to the tourist's home country while clearing customs. In the UK, conch shells are the 9th-most seized import. Conch shells have been used as a building material since ancient times, and new research is being conducted, to replicate their material for practical uses such as bone replacement, and also in architecture, to construct stronger structures. In Grenada, fishermen use the conch shell as a trombone to announce to the community that fish is available for sale. Conchs are used at carnival times in the Jouvert Jump where Diab Diab (Jab Jab) blow conch shells as part of the festivities. Especially in Guadeloupe, hearing conch shells being blown near ports at dawn and during Carnival times, too, is not uncommon. Many bands and trumpeters like Steve Turre use the conch shell in their performances. In the Caribbean, broken or up-turned conch shells are embedded into the tops of outdoor walls in an effort to maintain home security. In Tamil Nadu, India, the conch horn is blown during funerals as an acoustic indication of the funeral and to ward off evil spirits. In Key West, Florida, US, a native-born resident is affectionately called a "conch". In Japan, a conch is horagai (or jinkai). It was used as a trumpet in special ceremonies such as a royal cremation during the Edo period. Conch shells, (pu in Hawaiian) have been historically used as a method of communication, a tradition that is still observed in parts of modern life in Hawai'i. Religion Ancient Peru The Moche people of ancient Peru worshipped the sea and often depicted conch shells in their art. Aztec Quetzalcoatl, the Aztec god of wind and learning, wears around his neck the "wind breastplate" ehecailacocozcatl, "the spirally voluted wind jewel" made of a conch shell. Blowing a conch was considered a religious act. Hinduism A shankha shell (the shell of a T. pyrum, a species in the gastropod family Turbinellidae) is often referred to in the West as a conch shell or a chank shell. This shell is used as an important ritual object in Hinduism. The shell is used as a ceremonial trumpet, as part of religious practices, for example puja. The chank trumpet is sounded during worship at specific points, accompanied by ceremonial bells and singing. As it is an auspicious instrument, it is of purity and brilliance (Om, Devas, Brahman the Almighty Supreme creator, referred to in mantras, the Gayatri mantra explains a meditation on the brilliance of the sun), it is often played in a Lakshmi puja in temple or at home. In the story of Dhruva, the divine conch plays a special part. The warriors of ancient India blew conch shells to announce battle, as is described in the beginning of the war of Kurukshetra, in the Hindu epic the Mahabharata. The god of preservation, Vishnu, is said to hold a special conch, Panchajanya, that represents life, as it has come out of life-giving waters. According to Hindu mythology, Devas (gods) and Asuras (demons) once decided to churn the ocean to get a special divine nectar. This divine nectar, also known as amrit, was known to give immortality to whoever drank it. All the gods were on one side of it and the demons were on the other end. The churning (samudra manthan) produced a number of things from the ocean. One of the first things to come out of it was lethal poison called halahala. Everyone was terrified, as the poison was potent enough to destroy entire creation, so they went to Lord Shiva for protection and he consumed the poison to safeguard the universe. Lord Shiva took the poison in his mouth, but did not swallow it. Shankha also was one of divine objects that was obtained from samudra manthan. Also, the sound of the conch is believed to drive away the evil spirits. The blowing of the conch or "the shankha" needs a tremendous power and respiratory capacity. Hence, blowing it daily helps keep the lungs healthy. A newlywed Bengali bride wears bangles called shakha paula, made from coral and conch-shell powder. They have been a part of Bengali custom and tradition. In an ancient era, the Bengali farming community is thought to have resided near the river. They collected conch shells and powdered them to create bangles. They also used red coral for the bangles. They gave these beautiful bangles to their wives, as they could not afford ivory bangles. They were also known as poor-man's ivory, as they were cheap substitute for ivory bangles. Literature and the oral tradition In the Hindu tradition, the conch shell is used in ceremony as the sound it makes is said to correspond with higher frequency universal sounds associated with music of the spheres.
Biology and health sciences
Mollusks
null
290842
https://en.wikipedia.org/wiki/Obsessive%E2%80%93compulsive%20personality%20disorder
Obsessive–compulsive personality disorder
Obsessive–compulsive personality disorder (OCPD) is a cluster C personality disorder marked by a spectrum of obsessions with rules, lists, schedules, and order, among other things. Symptoms are usually present by the time a person reaches adulthood, and are visible in a variety of situations. The cause of OCPD is thought to involve a combination of genetic and environmental factors, namely problems with attachment. Obsessive–compulsive personality disorder is distinct from obsessive–compulsive disorder (OCD), and the relation between the two is contentious. Some studies have found high comorbidity rates between the two disorders but others have shown little comorbidity. Both disorders may share outside similarities, such as rigid and ritual-like behaviors. OCPD is highly comorbid with other personality disorders, autism spectrum, eating disorders, anxiety, mood disorders, and substance use disorders. The disorder is the most common personality disorder in the United States, and is diagnosed twice as often in males than in females; however, there is evidence to suggest the prevalence between men and women is equal. Signs and symptoms Obsessive–compulsive personality disorder (OCPD) is marked by an excessive obsession with rules, lists, schedules, and order; a need for perfection that interferes with efficiency and the ability to complete tasks; a devotion to productivity that hinders interpersonal relationships and leisure time; rigidity and zealousness on matters of morality and ethics; an inability to delegate responsibilities or work to others; restricted functioning in interpersonal relationships; restricted expression of emotion and affect; and a need for control over one's environment and self. Some of OCPD's symptoms are persistent and stable, whilst others are unstable. The obsession with perfectionism, reluctance to delegate tasks to others, and rigidity and stubbornness are stable symptoms. On the other hand, the symptoms that were most likely to change over time were the miserly spending style and the excessive devotion to productivity. This discrepancy in the stability of symptoms may lead to mixed results in terms of the course of the disorder, with some studies showing a remission rate of 58% after a 12-month period, whilst others suggest that the symptoms are stable and may worsen with age. Attention to order and perfection People with OCPD tend to be obsessed with controlling their environments; to satisfy this need for control, they become preoccupied with trivial details, lists, procedures, rules, and schedules. This preoccupation with details and rules makes the person unable to delegate tasks and responsibilities to other people unless they submit to their exact way of completing a task because they believe that there is only one correct way of doing something. They stubbornly insist that a task or job must be completed their way, and only their way, and may micromanage people when they are assigned a group task. They are frustrated when other people suggest alternative methods. A person with this disorder may reject help even when they desperately need it as they believe that only they can do something correctly. People with OCPD are obsessed with maintaining perfection. The perfectionism and the extremely high standards that they establish are to their detriment and may cause delays and failures to complete objectives and tasks. Mistakes are generally exaggerated. For example, a person may write an essay and, believing that it fell short of perfection, continues rewriting it, missing the deadline or even failing to complete the task. The subject may remain unaware that others become frustrated and annoyed by repeated delay and inconvenience so caused. Work relationships may then become a source of tension. Devotion to productivity Individuals with OCPD devote themselves to work and productivity at the expense of interpersonal relationships and recreation. Economic necessity, such as poverty, cannot account for this behavior. They may believe that they do not have sufficient time to relax because they have to prioritize their work above all. They may refuse to spend time with friends and family because of that. They may find it difficult to go on a vacation, and even if they book a vacation, they may keep postponing it so that it never happens. They may feel uncomfortable when they do go on a vacation and will take something along with them so they can work. They choose hobbies that are organized and structured, and they approach them as a serious task requiring work to perfect. The devotion to productivity in OCPD, however, is distinct from work addiction. OCPD is controlled and egosyntonic, whereas work addiction is uncontrolled and egodystonic, and the affected person may display signs of withdrawal. Rigidity Individuals with OCPD are overconscientious, scrupulous and rigid, and inflexible on matters of morality, ethics and other areas of life. They may force themselves and others to follow rigid moral principles and strict standards of performance. They are self-critical and harsh about their mistakes. These symptoms should not be accounted for or caused by a person's culture or religion. Their view of the world is polarised and dichotomous; there is no grey area between what is right and what is wrong. Whenever this dichotomous view of the world cannot be applied to a situation, this causes internal conflict as the person's perfectionist tendencies are challenged. People with this disorder are so obsessed with doing everything the "right and correct" way that they have a hard time understanding and appreciating the ideas, beliefs, and values of other people, and are reluctant to change their views, especially on matters of morality and politics. Restricted emotions and interpersonal functioning Individuals with this disorder may display little affection and warmth; their relationships and speech tend to have a formal and professional approach, and not much affection is expressed even to loved ones, such as greeting or hugging a significant other at an airport or train station. They are extremely careful in their interpersonal interactions. They have little spontaneity when interacting with others, and ensure that their speech follows rigid and austere standards by excessively scrutinising it. They filter their speech for embarrassing or imperfect articulation, while they maintain a high bar for what they consider to be acceptable. They raise their bar even higher when they are communicating with their superiors or with a person of high status. Communication becomes a time-consuming and exhausting effort, and they start avoiding it altogether. Others regard them as cold and detached as a result. Their need for restricting affection is a defense mechanism used to control their emotions. They may expunge emotions from their memories and organize them as a library of facts and data; the memories are intellectualized and rationalized, not experiences that they can feel. This helps them avoid unexpected emotions and feelings and allows them to remain in control. They can view self-exploration as a waste of time and have a patronising attitude towards emotional people. Interpersonal control Individuals with OCPD are at one extreme of the conscientiousness continuum. While conscientiousness is a desirable trait generally, its extreme presentation for those with OCPD leads to interpersonal problems. OCPD individuals present as over-controlled and this extends to the relationships they have with other people. Individuals with OCPD are reverential to authority and rules. OCPD individuals may therefore punish those who violate their strict standards. The inability to accept differences in belief or behaviors from others often leads to high conflict and controlling relationships with coworkers, spouses, and children. Cause The cause of OCPD is thought to involve a combination of genetic and environmental factors. There is clear evidence to support the theory that OCPD is genetically inherited; however, the relevance and impact of genetic factors vary with studies placing it somewhere between 27% and 78%. A twin study on the influence of genetics on the development of personality disorders over multiple personality disorders found that OCPD had a 0.78 heritability correlation, thus demonstrating that the development of OCPD can be strongly linked to genetics. Other studies have found links between attachment theory and the development of OCPD. According to this hypothesis, those with OCPD have never developed a secure attachment style, had overbearing parents, were shown little care, and/or were unable to develop empathetically and emotionally. Diagnosis DSM-5 The fifth edition of the Diagnostic and Statistical Manual of Mental Disorders, a widely used manual for diagnosing mental disorders, places obsessive–compulsive personality disorder under section II, under the "personality disorders" chapter, and defines it as: "a pervasive pattern of preoccupation with orderliness, perfectionism, and mental and interpersonal control, at the expense of flexibility, openness, and efficiency, beginning by early adulthood and present in a variety of contexts". A diagnosis of OCPD is only received when four out of the eight criteria are met. The eight criteria of OCPD described in the DSM-5 (of which four are required to be present in a patient for a diagnosis) are: Preoccupation with details Perfectionism interfering with task completion Rigidity and stubbornness Reluctance to delegate Excessive conscientiousness and pedantry (excessive concern with minor details and rules) Workaholic behavior Miserliness (excessive desire to save money) Inability to discard worn-out or worthless objects The list of criteria for the ICD-10 is similar, but does not include the last three criteria in the above list, and additionally includes the symptoms "intrusive thoughts" and "excessive doubt and caution" as criteria for diagnosis. Alternative model for diagnosis The DSM-5 also includes an alternative set of diagnostic criteria as per the dimensional model of conceptualizing personality disorders. Under the proposed set of criteria, a person only receives a diagnosis when there is an impairment in two out of four areas of one's personality functioning, and when there are three out of four pathological traits, one of which must be rigid perfectionism. The patient must also meet the general criteria C through G for a personality disorder, which state that the traits and symptoms being displayed by the patient must be stable and unchanging over time with an onset of at least adolescence or early adulthood, visible in a variety of situations, not caused by another mental disorder, not caused by a substance or medical condition, and abnormal in comparison to a person's developmental stage and culture/religion. Differential diagnosis There are several mental disorders in the DSM-5 that are listed as differential diagnoses for OCPD. They are as follows: Obsessive–compulsive disorder. OCD and OCPD have a similar name which may cause confusion; however, OCD can be easily distinguished from OCPD: OCPD is not characterized by true obsessions or compulsions. Hoarding disorder. A diagnosis of hoarding disorder is only considered when the hoarding behavior exhibited is causing severe impairment in the functioning of the person, such as an inability to access rooms in a house due to excessive hoarding. Narcissistic personality disorder. Individuals with a narcissistic personality disorder usually believe that they have achieved perfection (especially compared to other people) and cannot get better, whereas those with OCPD do not believe that they have achieved perfection, and are self-critical. Those with NPD tend to be stingy and lack generosity; however, they are usually generous when spending on themselves, unlike those with OCPD who hoard money and are miserly on themselves and others. Antisocial personality disorder. Similarly, individuals with antisocial personality disorder are not generous, but miserly around others, although they usually over-indulge themselves and are sometimes reckless in spending. Schizoid personality disorder. Schizoid personality disorder and obsessive–compulsive personality disorder may both display restricted affectivity and coldness; however, in OCPD, this is usually due to a controlling attitude, whereas, in SPD, it occurs due to a lack of ability to experience emotion and display affection. Other personality traits. Obsessive–compulsive personality traits may be particularly useful and helpful, especially in productive environments. Only when these traits become extreme and maladaptive and cause clinically significant impairment in several aspects of one's life should a diagnosis of OCPD be considered. Personality change due to another medical condition. Obsessive–compulsive personality disorder must be differentiated from a personality change due to a medical condition, which affects the central nervous system, and may cause changes in behavior and traits. Substance use disorders. Substance use may cause the advent of obsessive–compulsive traits. It is necessary that this is distinguished from underlying and persistent behavior, which must occur when a person is not under influence of a substance. ICD-10 The World Health Organization's ICD-10 uses the term (). At least four of the following must be present: Feelings of doubt Perfectionism Excessive conscientiousness Checking and preoccupation with details Stubbornness Caution Rigidity Insistent and unwelcome thoughts or impulses that do not attain the severity of an obsessive–compulsive disorder. Millon's subtypes In his book, Personality Disorders in Modern Life, Theodore Millon describes five types of obsessive–compulsive personality disorder, which he shortened to compulsive personality disorder. Comorbidity Obsessive–compulsive disorder OCPD is often confused with obsessive–compulsive disorder (OCD). Despite the similar names, they are two distinct disorders. Some OCPD individuals do have OCD, and the two can be found in the same family, sometimes along with eating disorders. The rate of comorbidity of OCPD in patients with OCD is estimated to be around 15–28%. However, due to the addition of the hoarding disorder diagnosis in the DSM-5, and studies showing that hoarding may not be a symptom of OCPD, the true rate of comorbidity may be much lower. There is significant similarity in the symptoms of OCD and OCPD, which can lead to complexity in distinguishing them clinically. For example, perfectionism is an OCPD criterion and a symptom of OCD if it involves the need for tidiness, symmetry, and organization. Hoarding is also considered both a compulsion found in OCD and a criterion for OCPD in the DSM-5. Even though OCD and OCPD are seemingly separate disorders there are obvious redundancies between the two concerning several symptoms. Regardless of similarities between the OCPD criteria and the obsessions and compulsions found in OCD, there are discrete qualitative dissimilarities between these disorders, predominantly in the functional part of symptoms. Unlike OCPD, OCD is described as invasive, and stressful. Time-consuming obsessions and habits are aimed at reducing obsession-related stress. OCD symptoms are at times regarded as egodystonic because they are experienced as alien and repulsive to the person. Therefore, there is a greater mental anxiety associated with OCD. In contrast, the symptoms seen in OCPD, although repetitive, are not linked with repulsive thoughts, images, or urges. OCPD characteristics and behaviors are known as egosyntonic, as people with this disorder view them as suitable and correct. On the other hand, the main features of perfectionism and inflexibility can result in considerable suffering in an individual with OCPD as a result of the associated need for control. The presence of OCPD in patients with OCD has been linked to a worse prognosis of OCD, especially when cognitive behavioral therapy was used. This may be due to the egosyntonic nature of OCPD which may lead to the obsessions becoming aligned with one's personal values. In contrast, the trait of perfectionism may improve the outcome of treatment as patients are likely to complete homework assigned to them with determination. The findings with regards to pharmacological treatment has also been mixed, with some studies showing a lower reception to SRIs in OCD patients with comorbid OCPD, with others showing no relationship. Comorbidity between OCD and OCPD has been linked to a more severe presentation of symptoms, a younger age of onset, more significant impairment in functioning, poorer insight, and higher comorbidity of depression and anxiety. Autism spectrum There are considerable similarities and overlap between autism spectrum disorder (ASD) and OCPD, such as list-making, inflexible adherence to rules, and obsessive aspects of ASD, although the latter may be distinguished from OCPD especially regarding affective behaviors, worse social skills, difficulties with Theory of Mind and intense intellectual interests, e.g. an ability to recall every aspect of a hobby. A 2009 study involving adult autistic people found that 32% of those diagnosed with ASD met the diagnostic requirements for a comorbid OCPD diagnosis. Eating disorders Perfectionism has been linked with anorexia nervosa in research for decades. A researcher in 1949 described the behavior of the average "anorexic girl" as being "rigid" and "hyperconscious", observing a tendency to "[n]eatness, meticulosity, and a mulish stubbornness not amenable to reason [which] make her a rank perfectionist." So common are such traits as perfectionism and rigidity among anorectics, that they have been referred to in clinical literature as "classical childhood features of patients with anorexia nervosa" or "classical premorbid personality descriptors of anorexia nervosa". Regardless of the prevalence of the full-fledged OCPD among eating disordered samples, the presence of this personality disorder or its traits, such as perfectionism, has been found to be positively correlated with a range of complications in eating disorders and a negative outcome, as opposed to impulsive features—those linked with histrionic personality disorder, for example—which predict a better outcome from treatment. OCPD predicts more severe symptoms of anorexia nervosa, and worse remission rates, however, OCPD and perfectionistic traits predicted a higher acceptance of treatment, which was defined as undergoing 5 weeks of treatment. People with anorexia nervosa who exercise excessively display a higher prevalence of several OCPD traits when compared to their counterparts who did not exercise excessively. The traits included self-imposed perfectionism, and the childhood OCPD traits of being rule-bound and cautious. It may be that people with OCPD traits are more likely to use exercise alongside restricting food intake in order to mitigate fears of increased weight, reduce anxiety, or reduce obsessions related to weight gain. Samples that had the childhood traits of rigidity, extreme cautiousness, and perfectionism endured more severe food restriction and higher levels of exercise and underwent longer periods of underweight status. It may be that OCPD traits are an indicator of a more severe manifestation of AN which is harder to treat. Gambling disorder A majority of those with lifelong gambling disorder have some sort of personality disorder, and the most common personality disorder amongst them is obsessive compulsive personality disorder. OCPD has a strong comorbidity with individuals who have gambling disorder. A study of data collected in the 2001-2002 National Epidemiologic Survey on Alcohol and Related Conditions looked at pathological gambling and psychiatric conditions as defined by the DSM-IV. Of the surveyed population consistent with gambling disorder, 60.8% also had a personality disorder, with OCPD appearing most frequently at 30%. About 300,000 U.S citizens have both a gambling disorder and obsessive compulsive personality disorder; and yet, there is little research on the comorbidity of the two disorders. Those with gambling disorders and OCPD do, indeed, exhibit different behavioral patterns than those with gambling disorders alone. More research on the relationship between the disorders is thought to help uncover causes and develop treatments for patients. Mental fatigue Recently, in 2020, the connection between mental fatigue and OCPD was published for the first time, even though mental fatigue has been previously associated with identified characteristics of OCPD such as workaholic behavior and perfectionism. Other disorders and conditions A diagnosis of OCPD is common with anxiety disorders, substance use disorders, and mood disorders. OCPD is also highly comorbid with Cluster A personality disorders,[4] especially paranoid and schizotypal personality disorders. OCPD is also linked to hypochondriasis, with some studies estimating a rate of co-occurrence as high as 55.7%. Moreover, OCPD has been found to be very common among some medical conditions, including Parkinson's disease and the hypermobile subtype of Ehler-Danlos syndrome. The latter may be explained by the need for control that arises from musculoskeletal problems and the associated features that arise early in life, whilst the former can be explained by dysfunctions in the fronto-basal ganglia circuitry. Treatment The best-validated treatment for OCPD is cognitive therapy (CT) or cognitive behavioral therapy (CBT), with studies showing an improvement in areas of personality impairment, and reduced levels of anxiety and depression. Group CBT is also associated with an increase in extraversion and agreeableness and reduced neuroticism. Interpersonal psychotherapy has been linked to even better results when it came to reducing depressive symptoms. Epidemiology Estimates for the prevalence of OCPD in the general population are 3%, making it the most common personality disorder. Current evidence is inconclusive as to whether OCPD is more common in men than women, or in equal rates among sexes. It is estimated to occur in 8.7% of psychiatric outpatient settings. A study of data collected in the 2001-2002 National Epidemiologic Survey on Alcohol and Related Conditions looked specifically for seven personality disorders as defined by the DSM-IV. The study concluded the most prevalent personality disorder of the survey's population to be OCPD, at 7.88%. This study also concluded there were no gender differences in prevalence and that OCPD was not a predictor of disability. History In 1908, Sigmund Freud named what is now known as obsessive–compulsive or anankastic personality disorder "anal retentive character". He identified the main strands of the personality type as a preoccupation with orderliness, parsimony (frugality), and obstinacy (rigidity and stubbornness). The concept fits his theory of psychosexual development. Freud believed that the anal retentive character faced difficulties regulating the control of defecation, leading to repercussions by the parents, and it is the latter that would cause the anal retentive character. Aubrey Lewis, in his 1936 book Problems of Obsessional Illness, suggests that anal-erotic characteristics are found in patients without obsessive thoughts, and proposed two types of obsessional personality, one melancholy and stubborn, the other uncertain and indecisive. In the book Contributions to the theory of the anal character, Karl Abraham noted that the core feature of the anal character is being perfectionistic, and he believed that these traits will help an individual in becoming industrious and productive, whilst hindering their social and interpersonal functioning, such as working with others. OCPD was included in the first edition of the Diagnostic and Statistical Manual of Mental Disorders in 1952 by the American Psychiatric Association under the name "compulsive personality". It was defined as a chronic and excessive preoccupation with adherence to rules and standards of conscience. Other symptoms included rigidity, over-conscientiousness, and a reduced ability to relax. The DSM-II (1968) changed the name to "obsessive–compulsive personality", and also suggested the term "anankastic personality" in order to reduce confusion between OCPD and OCD, but the proposed name was removed from later editions. The symptoms described in the DSM-II closely resembled those in the original DSM. In 1980, the DSM-III was released, and it renamed the disorder back to "compulsive personality disorder", and also included new symptoms of the disorder: a restricted expression of affect, and an inability to delegate tasks. Devotion to productivity, perfectionism, and indecisiveness were the other symptoms included. The DSM-III-R (1987) renamed the disorder again to "obsessive–compulsive personality disorder" and the name has remained since then. A diagnosis of OCPD was given when 5 of the 9 symptoms were met, and the 9 symptoms included perfectionism, preoccupation with details, an insistence that others submit to one's way, indecisiveness, devotion to work, restricted expression of affect, excessive conscientiousness, lack of generosity, and hoarding. With DSM-IV, OCPD was classified as a 'Cluster C' personality disorder. There was a dispute about the categorization of OCPD as an Axis II anxiety disorder. Although the DSM-IV attempted to distinguish between OCPD and OCD by focusing on the absence of obsessions and compulsions in OCPD, OC personality traits are easily mistaken for abnormal cognitions or values considered to underpin OCD. The disorder is a neglected and understudied area of research.
Biology and health sciences
Mental disorders
Health
291114
https://en.wikipedia.org/wiki/Parathyroid%20gland
Parathyroid gland
Parathyroid glands are small endocrine glands in the neck of humans and other tetrapods. Humans usually have four parathyroid glands, located on the back of the thyroid gland in variable locations. The parathyroid gland produces and secretes parathyroid hormone in response to low blood calcium, which plays a key role in regulating the amount of calcium in the blood and within the bones. Parathyroid glands share a similar blood supply, venous drainage, and lymphatic drainage to the thyroid glands. Parathyroid glands are derived from the epithelial lining of the third and fourth pharyngeal pouches, with the superior glands arising from the fourth pouch and the inferior glands arising from the higher third pouch. The relative position of the inferior and superior glands, which are named according to their final location, changes because of the migration of embryological tissues. Hyperparathyroidism and hypoparathyroidism, characterized by alterations in the blood calcium levels and bone metabolism, are states of either surplus or deficient parathyroid function. Structure The parathyroid glands are two pairs of glands usually positioned behind the left and right lobes of the thyroid. Each gland is a yellowish-brown flat ovoid that resembles a lentil seed, usually about 6 mm long and 3 to 4 mm wide, and 1 to 2 mm anteroposteriorly. There are typically four parathyroid glands. The two parathyroid glands on each side which are positioned higher are called the superior parathyroid glands, while the lower two are called the inferior parathyroid glands. Healthy parathyroid glands generally weigh about 30 mg in men and 35 mg in women. These glands are not visible or able to be felt during examination of the neck. Each parathyroid vein drains into the superior, middle and inferior thyroid veins. The superior and middle thyroid veins drain into the internal jugular vein, and the inferior thyroid vein drains into the brachiocephalic vein. Lymphatic drainage Lymphatic vessels from the parathyroid glands drain into deep cervical lymph nodes and paratracheal lymph nodes. Variation The parathyroid glands are variable in number: three or more small glands, and can usually be located on the posterior surface of the thyroid gland. Occasionally, some individuals may have six, eight, or even more parathyroid glands. Rarely, the parathyroid glands may be within the thyroid gland itself, the chest, or even the thymus. Microanatomy The parathyroid glands are named for their proximity to the thyroid—and serve a completely different role than the thyroid gland. The parathyroid glands are quite easily recognizable from the thyroid as they have densely packed cells, in contrast with the follicular structure of the thyroid. Two unique types of cells are present in the parathyroid gland: Chief cells, which synthesize and release parathyroid hormone. These cells are small, and appear dark when loaded with parathyroid hormone, and clear when the hormone has been secreted, or in their resting state. Oxyphil cells, which are lighter in appearance and increase in number with age, have an unknown function. Development In the early development of the human embryo, a series of five pharyngeal arches and four pharyngeal pouches form that give rise to the face, neck, and surrounding structures. The pouches are numbered such that the first pouch is the closest to the top of the embryo's head and the fourth is the farthest from it. The parathyroid glands originate from the interaction of the endoderm of the third and fourth pouch and neural crest mesenchyme. The position of the glands reverses during embryological development. The pair of glands which is ultimately inferior develops from the third pouch with the thymus, whereas the pair of glands which is ultimately superior develops from the fourth pouch. During embryological development, the thymus migrates downward, dragging the inferior glands with it. The superior pair are not dragged downward by the fourth pouch to the same degree. The glands are named after their final, not embryological, positions. Since the thymus's ultimate destination is in the mediastinum of the chest, it is occasionally possible to have ectopic parathyroids derived from the third pouch within the chest cavity if they fail to detach in the neck. Parathyroid development is regulated by a number of genes, including those coding for several transcription factors. Function The major function of the parathyroid glands is to maintain the body's calcium and phosphate levels within a very narrow range, so that the nervous and muscular systems can function properly. The parathyroid glands do this by secreting parathyroid hormone (PTH). Parathyroid hormone (also known as parathormone) is a small protein that takes part in the control of calcium and phosphate homeostasis, as well as bone physiology. Parathyroid hormone has effects antagonistic to those of calcitonin. Calcium. PTH increases blood calcium levels by directly stimulating osteoblasts and thereby indirectly stimulating osteoclasts (through RANK/RANKL mechanism) to break down bone and release calcium. PTH increases gastrointestinal calcium absorption by activating vitamin D, and promotes calcium conservation (reabsorption) by the kidneys. Phosphate. PTH is the major regulator of serum phosphate concentrations via actions on the kidney. It is an inhibitor of proximal tubular reabsorption of phosphorus. Through activation of vitamin D the absorption (intestinal) of phosphate is increased. Disorders Parathyroid disease is conventionally divided into states where the parathyroid is overactive (hyperparathyroidism), and states where the parathyroid is under- or hypoactive (hypoparathyroidism). Both states are characterised by their symptoms, which relate to the excess or deficiency of parathyroid hormone in the blood. Hyperparathyroidism Primary Hyperparathyroidism is the state in which there is excess parathyroid hormone circulating in the blood. This may cause bone pain and tenderness, due to increased bone resorption. With increased circulating calcium, there may be other symptoms associated with hypercalcemia, most commonly dehydration. Hyperparathyroidism is most commonly caused by a benign proliferation of chief cells in one parathyroid gland, and rarely in MEN syndrome. This is known as primary hyperparathyroidism, which is generally managed by surgical removal of the abnormal parathyroid gland. Secondary Renal disease may lead to hyperparathyroidism. When too much calcium is lost from the blood via urination, there is a compensation by the parathyroid, and parathyroid hormone is released. The glands enlarge (hypertrophy) to synthesize more parathyroid hormone. This is known as secondary hyperparathyroidism. Tertiary If secondary hyperparathyroidism persists over months, the parathyroid tissue may become unresponsive to the blood calcium levels, and begin to autonomously release parathyroid hormone. This is known as tertiary hyperparathyroidism. Hypoparathyroidism The state of decreased parathyroid activity is known as hypoparathyroidism. This is most commonly associated with damage to the glands or their blood supply during thyroid surgeryit may be associated with rarer genetic syndromes such as DiGeorge syndrome, which is inherited as an autosomal dominant syndrome. Hypoparathyroidism will occur after surgical removal of the parathyroid glands. Occasionally, an individual's tissues are resistant to the effects of parathyroid hormone. This is known as pseudohypoparathyroidism. In this case the parathyroid glands are fully functional, and the hormone itself is not able to function, resulting in a decrease in blood calcium levels. Pseudohypoparathyroidism is often associated with the genetic condition Albright's hereditary osteodystrophy. Pseudo-pseudohypoparathyroidism, one of the longest words in the English language, is used to describe an individual with Albright's hereditary osteodystrophy with normal parathyroid hormone and serum calcium levels. Hypoparathyroidism may present with symptoms associated with decreased calcium, and is generally treated with Vitamin D analogues. History The parathyroid glands were first discovered in the Indian rhinoceros by Richard Owen in 1852. In his description of the neck anatomy, Owen referred to the glands as "a small compact yellow glandular body attached to the thyroid at the point where the veins emerged". The glands were first discovered in humans by Ivar Viktor Sandström (1852–1889), a Swedish medical student, in 1880 at Uppsala University. Unaware of Owen's description, he described the glands in his monograph "On a New Gland in Man and Fellow Animals" as the "glandulae parathyroidae", noting its existence in dogs, cats, rabbits, oxen, horses and humans. For several years, Sandström's description received little attention. Eugene Gley, Giulio Vassale, and others documented the putative function of the glands in 1891, noting the connection between their removal and the development of muscular tetany. William G. MacCallum in 1908, investigating tumours of the parathyroid, proposed their role in calcium metabolism. He noted that "Tetany occurs spontaneously in many forms and may be produced by the destruction of the parathyroid glands". The first successful removal of the parathyroid may have been carried out in 1928 by medical doctor Isaac Y Olch, whose intern had noticed elevated calcium levels in an elderly patient with muscle weakness. Prior to this surgery, patients with removed parathyroid glands typically died from muscular tetany. Parathyroid hormone was isolated in 1923 by Adolph M. Hanson and 1925 by James B. Collip. Studies of parathyroid hormone levels by Roger Guillemin, Andrew Schally and Rosalyn Sussman Yalow led to the development of immunoassays capable of measuring body substances and a Nobel Prize in 1977. Other animals Parathyroid glands are found in all adult tetrapods; they vary in their number and position. Mammals typically have four parathyroid glands, while other types of animals typically have six. The removal of parathyroid glands in animals produces a condition resembling acute poisoning with irregular muscle contractions. Fish do not possess parathyroid glands; several species have been found to express parathyroid hormone. Developmental genes and calcium-sensing receptors in fish gills are similar to those within the parathyroid glands of birds and mammals. It has been suggested that the tetrapod glands may have been evolutionarily derived from these fish gills. Additional images
Biology and health sciences
Endocrine system
Biology
291122
https://en.wikipedia.org/wiki/Hazelnut
Hazelnut
The hazelnut is the fruit of the hazel tree and therefore includes any of the nuts deriving from species of the genus Corylus, especially the nuts of the species Corylus avellana. They are also known as cobnuts or filberts according to species. Hazelnuts are used as a snack food, in baking and desserts, and in breakfast cereals such as muesli. In confectionery, they are used to make praline, and also used in combination with chocolate for chocolate truffles and products such as chocolate bars and hazelnut cocoa spreads such as Nutella. They are also used in Frangelico liqueur. Hazelnut oil, pressed from hazelnuts, is strongly flavored and high in monounsaturated fat. It is used as a cooking oil and as a salad or vegetable dressing. Turkey is the world's largest producer of hazelnuts, accounting for 64% of total production in 2021. Description A hazelnut cob is roughly spherical to oval, about long and in diameter, with an outer fibrous husk surrounding a smooth shell, while a filbert is more elongated, being about twice as long as its diameter. The nut falls out of the husk when ripe, about seven to eight months after pollination. The seed is edible and consumed raw, roasted or ground into a paste. The seed has a thin, dark brown skin, which is sometimes removed before cooking. Cultivation History In 1995, evidence of large-scale Mesolithic nut processing, some 8,000 years old, was found in a midden pit on the island of Colonsay in Scotland. The evidence consists of a large, shallow pit full of the remains of hundreds of thousands of burned hazelnut shells. Hazelnuts have been found on other Mesolithic sites, but rarely in such quantities or concentrated in one pit. The nuts were radiocarbon dated to 7720±110 BP, which calibrates to . Similar sites in Britain are known only at Farnham in Surrey and Cass ny Hawin on the Isle of Man. This discovery gives an insight into communal activity and planning in the period. The nuts were harvested in a single year, and pollen analysis suggests that all of the hazel trees were cut down at the same time. The scale of the activity and the lack of large game on the island suggest that Colonsay may have contained a community with a largely vegetarian diet for the time they spent on the island. Originally, the pit was on a beach close to the shore and was associated with two smaller, stone-lined pits whose function remains obscure, a hearth and a second cluster of pits. The traditional method to increase nut production is called brutting, which involves prompting more of the tree's energy to go into flower bud production by snapping, but not breaking off, the tips of the new year shoots six or seven leaf groups from where they join with the trunk or branch, at the end of the growing season. The traditional term for an area of cultivated hazelnuts is a plat. Cultivars The many cultivars of the hazel include 'Atababa', 'Barcelona', 'Butler', 'Casina', 'Clark', 'Cosford', 'Daviana', 'Delle Langhe', 'England', 'Ennis', 'Halls Giant', 'Jemtegaard', 'Kent Cob', 'Lewis', 'Tokolyi', 'Tonda Gentile', 'Tonda di Giffoni', 'Tonda Romana', 'Wanliss Pride', and 'Willamette'. Some of these are grown for specific qualities of the nut, including large nut size or early or late fruiting, whereas others are grown as pollinators. The majority of commercial hazelnuts are propagated from root sprouts. Some cultivars are of hybrid origin between common hazel and filbert. In Ireland and the United Kingdom, hazelnuts are sometimes referred to as cobnuts, for which a specific cultivated variety – Kentish cobnuts – is the main variety cultivated in fields known as plats, hand-picked, and eaten green. According to the BBC, a national collection of cobnut varieties exists at Roughway Farm, near Plaxtol in Kent. They are called cobnuts because cob was a word used to refer to the head or "noggin,” and children had a game in which they would tie a string to a hazelnut and use it to try to hit an opponent on the head. Cropping system In Europe hazelnuts are traditionally grown as multi-trunk trees where the rootstock is formed by the variety itself. To enhance the possibility for mechanization and to prevent suckering, a single-trunk tree can be formed by grafting a scion of the desired variety on a Corylus colurna rootstock. There are different approaches in orchard floor management. Strong infestations of certain weeds can lower the yield. Therefore, to prevent yield loss by competition, herbicides are used to create bare soil. Using cover crops protects the soil and suppresses weed establishment without a yield reduction. Silvopastoral systems where animals like pigs are kept in the orchard or silvoarable systems where crops are grown between the rows could introduce new sources of revenue into the hazelnut production and may have beneficial effects on hazelnut yield by fertilizing the soil through nitrogen fixation or animal dung. These systems limit the plant protection measures, potentially impacting the yield quality and level. There are very few studies on these systems, leading to a higher uncertainty for the producer. Harvesting Hazelnuts are harvested annually in mid-autumn. As autumn comes to a close, the trees drop their nuts and leaves. Most commercial growers wait for the nuts to drop rather than using equipment to shake them from the tree. The harvesting of hazelnuts is performed either by hand or by manual or mechanical raking of fallen nuts. Four primary pieces of equipment are used in commercial harvesting: the sweeper, the harvester, the nut cart, and the forklift. The sweeper moves the nuts into the center of the rows, the harvester lifts and separates the nuts from any debris (i.e., twigs and leaves), the nut cart holds the nuts picked up by the harvester, and the forklift brings a tote to offload the nuts from the nut cart and then stacks the totes to be shipped to the processor (nut dryer). The sweeper is a low-to-the-ground machine that makes two passes in each tree row. It has a belt attached to the front that rotates to sweep leaves, nuts, and small twigs from left to right, depositing the material in the center of the row as it drives forward. On the rear of the sweeper is a powerful blower to blow material left into the adjacent row with air speeds up to . Careful grooming during the year and patient blowing at harvest may eliminate the need for hand raking around the tree trunk, where nuts may accumulate. The sweeper prepares a single center row of nuts, narrow enough for the harvesting tractor to drive over without driving on the center row. It is best to sweep only a few rows ahead of the harvesters at any given time, to prevent the tractor that drives the harvester from crushing the nuts that may still be falling from the trees. Hazelnut orchards may be harvested up to three times during the harvest season, depending on the number of nuts in the trees and the rate of nut drop due to the weather. The harvester is a slow-moving machine pushed by a tractor, which lifts the material off the ground and separates the nuts from the leaves, empty husks, and twigs. As the harvester drives over the rows, a rotating cylinder with hundreds of tines rakes the material onto a belt. The belt takes the material over a blower and under a powerful vacuum that sucks any lightweight soil, leaves from the nuts, and discharges them into the orchard. The remaining nuts are conveyed into a cart pulled behind the harvester. Once a tote is filled with nuts, the forklift hauls away the full totes and brings empty ones back to the harvester to maximize the harvester's time. Two different timing strategies are used for collecting the fallen nuts. The first is to harvest early when about half of the nuts have fallen. With less material on the ground, the harvester can work faster with less chance of a breakdown. The second option is to wait for all the nuts to fall before harvesting. Although the first option is considered the better of the two, two or three passes do take more time to complete than one. Production In 2021, world production of hazelnuts (in shells) was 1.1 million tonnes. The hazelnut production in Turkey accounts for 64% of the world total, followed by Italy, the United States, and Azerbaijan as secondary producers. Food Hazelnuts are used in confections to make pralines, chocolate truffles, and hazelnut paste products. The (solid) combination of ground hazelnuts with chocolate is called gianduja. In Austria, hazelnut paste is an ingredient for making tortes, such as Viennese hazelnut torte. In Kyiv cake, hazelnut flour is used to flavor its meringue body, and crushed hazelnuts are sprinkled over its sides. Dacquoise, a French dessert cake, often contains a layer of hazelnut meringue. Hazelnuts are used in Turkish cuisine and Georgian cuisine; the snack churchkhela and sauce satsivi are used, often with walnuts. Hazelnuts are also a common constituent of muesli. The nuts may be eaten fresh or dried, having different flavors. Nutrition Raw hazelnuts are 5% water, 61% fat, 17% carbohydrates, and 15% protein (table). In a reference amount, raw hazelnuts supply of food energy and are a rich source (20% or more of the Daily Value, DV) of numerous essential nutrients (see table). Hazelnuts contain particularly high amounts of protein, dietary fiber, vitamin E, iron, thiamin, phosphorus, manganese, and magnesium, all exceeding 30% DV (table). Several B vitamins have appreciable content. In lesser but still significant amounts (moderate content, 10–19% DV) are vitamin K, calcium, zinc, and potassium (table). Hazelnuts are a rich source of dietary fat, accounting for 93% DV in a 100-gram amount. The fat components are monounsaturated fat as oleic acid (75% of total), polyunsaturated fat mainly as linoleic acid (13% of total), and saturated fat, mainly as palmitic acid and stearic acid (together, 7% of total). In culture The hazelnut is used as a literary device by Julian of Norwich (c. 1343 – after 1416) within her mystical Christian treatise Revelations of Divine Love. The hazelnut shell is imagined as a chariot for the fairy Queen Mab within English playwright and poet William Shakespeare's play Romeo and Juliet. The hazel fruit is also used as a metaphorical device in the poem To Autumn by the English Romantic poet John Keats. Gallery
Biology and health sciences
Fagales
null
291164
https://en.wikipedia.org/wiki/Nymphalidae
Nymphalidae
The Nymphalidae are the largest family of butterflies, with more than 6,000 species distributed throughout most of the world. Belonging to the superfamily Papilionoidea, they are usually medium-sized to large butterflies. Most species have a reduced pair of forelegs and many hold their colourful wings flat when resting. They are also called brush-footed butterflies or four-footed butterflies, because they are known to stand on only four legs while the other two are curled up; in some species, these forelegs have a brush-like set of hairs, which gives this family its other common name. Many species are brightly coloured and include popular species such as the emperors, monarch butterfly, admirals, tortoiseshells, and fritillaries. However, the under wings are, in contrast, often dull and in some species look remarkably like dead leaves, or are much paler, producing a cryptic effect that helps the butterflies blend into their surroundings. Nomenclature Rafinesque introduced the name Nymphalia as a subfamily name in diurnal Lepidoptera. Rafinesque did not include Nymphalis among the listed genera, but Nymphalis was unequivocally implied in the formation of the name (Code Article 11.7.1.1). The attribution of the Nymphalidae to Rafinesque has now been widely adopted. Morphology In the adult butterflies, the first pair of legs is small or reduced, giving the family the other names of four-footed or brush-footed butterflies. The caterpillars are hairy or spiky with projections on the head, and the chrysalids have shiny spots. The forewings have the submedial vein (vein 1) unbranched and in one subfamily forked near the base; the medial vein has three branches, veins 2, 3, and 4; veins 5 and 6 arise from the points of junction of the discocellulars; the subcostal vein and its continuation beyond the apex of cell, vein 7, has never more than four branches, veins 8–11; 8 and 9 always arise from vein 7, 10, and 11 sometimes from vein 7 but more often free, i.e., given off by the subcostal vein before apex of the cell. The hindwings have internal (1a) and precostal veins. The cell in both wings is closed or open, often closed in the fore, open in the hindwing. The dorsal margin of the hindwing is channelled to receive the abdomen in many of the forms. The antennae always have two grooves on the underside; the club is variable in shape. Throughout the family, the front pair of legs in the male, and with three exceptions (Libythea, Pseudergolis, and Calinaga) in the female also, is reduced in size and functionally impotent; in some, the atrophy of the forelegs is considerable, e.g., the Danainae and Satyrinae. In many of the forms of these subfamilies, the forelegs are kept pressed against the underside of the thorax, and are in the male often very inconspicuous. Systematics and phylogeny The phylogeny of the Nymphalidae is complex. Several taxa are of unclear position, reflecting the fact that some subfamilies were formerly well-recognized as distinct families due to insufficient study. The five main clades within the family are: The libytheine clade (basal) Libytheinae (snout butterflies, earlier treated as the distinct family Libytheidae) The danaine clade (basal) Danainae (milkweed butterflies, earlier treated as the distinct family Danaidae) Host plant families include Apocynaceae, Asclepiadoideae (subfamily of Apocynaceae), and Moraceae. Ithomiini (about 300 Neotropical species, sometimes considered a subfamily Ithomiinae) Most species have long wings, and some have transparent wings. Host plants are in the families Apocynaceae, Gesneriaceae, and Solanaceae. Tellervini (about 6–10 species in Australasia, sometimes considered a subfamily Tellervinae) Caterpillars resemble those of the Danainae and feed on Apocynaceae. The satyrine clade Calinaginae (about six species, restricted to the Himalayas) Mimics of the Danainae, they are restricted to host plants in the family Moraceae. Charaxinae Tropical canopy butterflies, the caterpillars often have head spines or projections. Mostly edible species, have some Batesian mimics. Host plants are in the families Annonaceae, Celastraceae, Convolvulaceae, Euphorbiaceae, Fabaceae, Flacourtiaceae, Lauraceae, Myrtaceae, Piperaceae, Poaceae, Rhamnaceae, Rutaceae, Santalaceae, and Sapindaceae. Morphinae (including Amathusiini, sometimes considered a subfamily Amathusiinae) Include the spectacular neotropical Morpho, its food plants include the Arecaceae, Bignoniaceae, Fabaceae, Menispermaceae, Poaceae, and Sapindaceae. Brassolini (owls, neotropical with 70–80 species, mostly crepuscular, sometimes considered a subfamily Brassolinae) Host plants in the families Arecaceae, Bromeliaceae, Heliconiaceae, Musaceae, and Poaceae. Satyrinae (satyrs and browns, earlier treated as distinct family Satyridae) Host plants are in the families Arecaceae, Araceae, Cyperaceae, Heliconiaceae, Poaceae, and Selaginellaceae. The heliconiine clade (sister group of the nymphaline clade, excludes former tribes Biblidini and Cyrestini, and tribes Pseudergolini and Coeini) Heliconiinae (earlier treated as distinct family Heliconiidae) Colourful tropical butterflies, they are noted for Müllerian mimicry. All species use host plants in the family Passifloraceae. Acraeini (mostly African, but some species in Asia, sometimes considered a family Acraeinae) Host plants are in the families Asteraceae, Passifloraceae, Sterculiaceae, Tiliaceae, and Urticaceae. Limenitidinae The nymphaline clade (sister group of the heliconiine clade, also includes tribes Coeini and Pseudergolini) Apaturinae (mostly tropical) Host plants are in the family Ulmaceae. Caterpillars are smooth with bifid tails and horns on the head. Biblidinae (formerly in Limenitidinae) Cyrestinae (formerly in Limenitidinae) Nymphalinae (a large subfamily that sometimes includes the Limenitidinae and Biblidinae) Some species migrate. Caterpillars are sometimes covered in spines. Host plants include Acanthaceae, Caprifoliaceae, Convolvulaceae, Euphorbiaceae, Fagaceae, Flacourtiaceae, Lamiaceae, Loranthaceae, Moraceae, Plantaginaceae, Poaceae, Rubiaceae, Rutaceae, Salicaceae, Sapindaceae, Scrophulariaceae, Urticaceae, and Verbenaceae. Example species from this family Actinote zikani, genus Actinote Archdukes, genus Lexias California tortoiseshell, Nymphalis californica Comma, Polygonia c-album Common buckeye, Junonia coenia Common snout butterfly, Libytheana carinenta Cracker butterflies, genus Hamadryas Crimson patch, Chlosyne janais Edith's checkerspot, Euphydryas editha Grayling (butterfly), Hipparchia semele Hackberry emperor, Asterocampa celtis Lorquin's admiral, Limenitis lorquini Marsh fritillary, Euphydryas aurinia Meadow brown, Maniola jurtina Mourning cloak, Nymphalis antiopa Monarch butterfly, Danaus plexippus Blue morpho, Morpho menelaus Painted lady, Vanessa cardui Peacock, Aglais io Plain tiger, Danaus chrysippus Question mark, Polygonia interrogationis Red admiral, Vanessa atalanta Small heath, Coenonympha pamphilus Small tortoiseshell, Nymphalis urticae Gatekeeper, Pyronia tithonus Small pearl-bordered fritillary, Boloria selene Andromeda satyr, Cithaerias andromeda Texan crescentspot butterfly, Anthanassa texana texana Zerene fritillary, Speyeria zerene (includes several subspecies such as Oregon silverspot, Speyeria zerene hippolyta) Morphology The trait for which these butterflies are most known is the use of only four legs; the reason their forelegs have become vestigial is not yet completely clear. Some suggest the forelegs are used to amplify the sense of smell, because some species possess a brush-like set of soft hair called setae, which has led researchers to believe the forelegs are used to improve signaling and communication between the species, while standing in the other four. This ability proves useful in terms of reproduction and the overall health of the species, and it is the leading theory so far.
Biology and health sciences
Lepidoptera
Animals
291169
https://en.wikipedia.org/wiki/Open-chain%20compound
Open-chain compound
In chemistry, an open-chain compound (or open chain compound) or acyclic compound (Greek prefix α 'without' and κύκλος 'cycle') is a compound with a linear structure, rather than a cyclic one. An open-chain compound having no side groups is called a straight-chain compound (also spelled as straight chain compound). Many of the simple molecules of organic chemistry, such as the alkanes and alkenes, have both linear and ring isomers, that is, both acyclic and cyclic. For those with 4 or more carbons, the linear forms can have straight-chain or branched-chain isomers. The lowercase prefix n- denotes the straight-chain isomer; for example, n-butane is straight-chain butane, whereas i-butane is isobutane. Cycloalkanes are isomers of alkenes, not of alkanes, because the ring's closure involves a C-C bond. Having no rings (aromatic or otherwise), all open-chain compounds are aliphatic. Typically in biochemistry, some isomers are more prevalent than others. For example, in living organisms, the open-chain isomer of glucose usually exists only transiently, in small amounts; D-glucose is the usual isomer; and L-glucose is rare. Straight-chain molecules are often not literally straight, in the sense that their bond angles are often not 180°, but the name reflects that they are schematically straight. For example, the straight-chain alkanes are wavy or "puckered", as the models below show.
Physical sciences
Substance
Chemistry
291328
https://en.wikipedia.org/wiki/Beijing%20Capital%20International%20Airport
Beijing Capital International Airport
Beijing Capital International Airport is the busiest of the two international airports serving Beijing, the capital city of China (the other one being Beijing Daxing International Airport). The airport is located northeast of downtown Beijing, in an exclave of Chaoyang and the surroundings of that exclave in suburban Shunyi. The airport is owned and operated by the Beijing Capital International Airport Company Limited, a state-controlled company. The airport's IATA Airport code, PEK, is based on the city's former romanized name, Peking. The facility covers an area of 3,657 acres (14.8 sq. km) of airport property. History Beijing Airport was opened on 1 March 1958. The airport then consisted of one small terminal building, which still stands to this day, apparently for the use of VIPs and charter flights, along with a single runway on its eastern side, which was extended to in 1966 and in 1982 respectively. Another runway on the west was completed in October 1978. On 1 January 1980, a newer, larger Terminal 1 – green in color – opened, with docks for 10 to 12 aircraft. The terminal was larger than the one built in the 1950s but, by the mid-1990s, its size had become inadequate. The first international flight to China and Beijing Capital International Airport was of Pakistan International Airlines from Islamabad. In late 1999, to mark the 50th anniversary of the founding of the PRC, the airport underwent a new round of expansion. Terminal 2 opened on 1 November of that year and Terminal 1 was temporarily closed for renovation. 20 September 2004 saw the opening of the renovated Terminal 1 which, at that time, only handled China Southern Airlines' domestic and international flights from Beijing. Other airlines' domestic and international flights still operated in Terminal 2. More expansion began in 2007. A third runway opened on 29 October 2007, to relieve congestion on the other two runways. Terminal 3 (T3) was completed in February 2008, in time for the 2008 Beijing Summer Olympics. The significant expansion included a rail link to the city center. At its opening, the new Terminal 3 was the largest man-made structure in the world in terms of area covered, and a major landmark in the developing Chinese capital. The expansion was largely funded by a 30 billion yen loan from Japan and a 500-million-euro (US$625 million) loan from the European Investment Bank (EIB). The loan was the largest ever granted by the EIB in Asia, and the agreement was signed during the eighth China-EU Summit held in September 2005. Following the 2008 Summer Olympics and the addition of adding the new terminal building, Beijing Capital overtook Tokyo Haneda as the busiest airport in Asia based on scheduled seat capacity. Due to limited capacity of Beijing Capital International Airport, plans were announced for the construction of a new airport at Daxing. The project was given final approval on 13 January 2013. Construction began in late 2014 and was completed in 2019. The new Daxing Airport became the Beijing home of China Eastern Airlines, China Southern Airlines, and China United Airlines, while Air China and Hainan Airlines remained at Capital. Terminals Shuttle buses connect the airport's three terminals. Terminal 2 serves Hainan Airlines and SkyTeam with the exception of China Airlines, Oneworld member SriLankan Airlines, and also other domestic and international flights. Terminal 3, the newest terminal, serves Air China, Star Alliance, Oneworld members with the exception of SriLankan Airlines, plus SkyTeam member China Airlines, and some other domestic and international flights that do not operate from Terminals 2. Terminal 2 Terminal 2 opened on 1 November 1999, with a floor area of . This terminal was used to replace Terminal 1 while the latter was undergoing renovation, cramping all airlines despite being far bigger than Terminal 1. It can handle twenty aircraft at docks connecting directly to the terminal building. Prior to the opening of Terminal 3, all international flights (and the majority of domestic flights) operated from this terminal. This terminal now houses Hainan Airlines (all international, Hong Kong, Macau, and Taiwan flights), SkyTeam with the exception of China Airlines, which uses Terminal 3, Oneworld member SriLankan Airlines, Air Koryo, and other domestic and international flights other than those operated by Shanghai Airlines, Star Alliance members and Oneworld members. A gate capable of handling the A380 (gate 21) was also built at the terminal. Star Alliance member Air China also uses Terminal 2 for some of its domestic flights. Terminals 1 and 2 are linked by a public walkway that takes about 10–15 minutes to traverse. Westwing Satellite Terminal (formerly Terminal 1) Terminal 1, with of space, opened on 1 January 1980, and replaced the smaller existing terminal, which had been in operation since 1958. Terminal 1 was closed for renovation from 1 November 1999 to 20 September 2004, during which all airlines operated from Terminal 2. Featuring 16 gates, it was the operational base for the domestic routes of China Southern Airlines and a few other airlines such as XiamenAir and Chongqing Airlines, and was originally planned to handle domestic traffic excluding those to Hong Kong and Macau. With the opening of Terminal 3, the terminal was closed for light refurbishment, and its airlines were moved to Terminal 2 on 20 May 2008. Terminal 1 reopened for a second time on 27 June 2008, and became the operational base for all domestic flights operated by the HNA Group including those of Hainan Airlines, Grand China Air and Tianjin Airlines, while all HNA Group's international, Hong Kong, Macau, and Taiwan flights remain in Terminal 2. On May 3, 2020, Terminal 1 temporary closed for reproposing, after Hainan Airlines moved its all domestic routes to Terminal 2. The Terminal 1 was reopened on August 1, 2023 as Westwing Satellite Terminal as part of Terminal 2. Terminal 3 Construction of Terminal 3 started on 28 March 2004, and the terminal opened in two stages. Trial operations commenced on 29 February 2008, when seven airlines including El Al, Qantas, Qatar Airways, Shandong Airlines and Sichuan Airlines moved into the terminal. Twenty other airlines followed when the terminal became fully operational on 26 March 2008. Currently, it mainly houses Air China, Star Alliance, Oneworld with the exception of SriLankan Airlines, which uses Terminal 2, SkyTeam member China Airlines, and other domestic and international flights that are not operated from Terminal 2. Star Alliance members LOT Polish Airlines, Scandinavian Airlines, Lufthansa, Austrian Airlines, United Airlines, Air Canada, Turkish Airlines, Thai Airways International, Singapore Airlines, All Nippon Airways, Asiana Airlines, and Air China use Terminal 3-E as part of the Move Under One Roof program to co-locate alliance members. Terminal 3 was designed by a consortium of Netherlands Airport Consultants (NACO), Foster and Partners, Arup and the Beijing Institute of Architectural Design (BIAD). Lighting was designed by UK lighting architects Speirs and Major Associates. The budget for the expansion is US$3.5 billion. Much larger in size and scale than the other two terminals, Terminal 3 was the largest airport terminal-building complex in the world to be built in a single phase, with in total floor area at its opening. It features a main passenger terminal (Terminal 3C) and two satellite concourses (Terminal 3D and Terminal 3E), all of the five floors above ground and two underground, with the letters "A and B" omitted to avoid confusion with the existing Terminals 1 and 2. Only two concourses were initially opened, namely, Terminal 3C dedicated for domestic flights and Terminal 3E for international flights. Terminal 3D officially opened on 18 April 2013. The newly opened concourse is temporarily used solely by Air China for some of its domestic flights. At the time of its opening, Terminal 3 was the largest airport passenger terminal building in the world. Its title as the world's largest passenger terminal was surrendered on 14 October 2008 to Dubai International Airport's Terminal 3, which has of floor space. On 20 July 2013, a man in a wheelchair detonated small homemade explosives in Terminal 3 of the Beijing International Airport. The bomber, reported to be Ji Zhongxing, was injured and taken to a hospital for his injuries. No other people were hurt. System, security and luggage Terminal 3 has a transportation hub with a 7,000-car garage. The transportation center has designated traffic lanes for airport buses, taxis, and private vehicles. Travelers bound for T3 can exit their vehicles and enter T3 within five minutes. There is also a station for the Capital Airport Express of the Beijing Subway. Terminal 3 has 243 elevators, escalators or moving walkways. One of Terminal 3's highlights is the US$240 million luggage-transfer systems. The luggage system is equipped with yellow carts, each of which has a code that matches the bar code on every piece of luggage loaded and allows easy and accurate tracking. More than 200 cameras are used to monitor activities in the luggage area. The luggage system can handle 19,200 pieces of luggage per hour. After luggage is checked in at any of the 292 counters in Terminal 3C, it can be transferred at a speed of ten meters per second. Hence, a suitcase can travel from T3C to T3E in five minutes. Arriving passengers should be able to begin retrieving their luggage within 4.5 minutes after airplanes are unloaded. Besides X-ray scanners, additional equipment is used to conduct baggage screening. Passengers will be able to check-in their luggage at the airport from several hours to even a day before their flights. The airport will store the luggage in its luggage system and then load it on the correct aircraft. Appearance The highest building at the airport, a monitoring tower, stands at the southern end of T3. The roof of T3 is red, the Chinese color for good luck. The terminal's ceilings use white strips for decoration and to indicate directions. Under the white strips, the basic color of the ceiling is orange with light to dark tones indicating where a passenger is inside the building. The roof is light orange in the center. The color deepens as the roof extends to the sides in T3E and goes the other way round in T3C. The roof of T3 has dozens of triangular windows to let in the daylight. Light angles can be adjusted to ensure adequate interior lighting. Many traditional Chinese elements will be employed in the terminal's interior decoration, including a "Menhai", a big copper vat used to store water for fighting fires in the Forbidden City, and the carvings imitating the famous Nine-Dragon Wall. An indoor garden in the T3E waiting area is built in the style of imperial gardens such as the Summer Palace. In T3C, a tunnel landscape of an underground garden has been finished with plants on each side so that passengers can appreciate them inside the mini-train. Facilities The T3 food-service area is called a "global kitchen", where 72 stores provide food ranging from formal dishes to fast food, from Chinese to Western, and from bakery goods to ice cream. Airport officials have promised that people who buy products at the airport will find the same prices in central Beijing. In addition to food and beverage areas, there is a domestic retail area, a duty-free-store area and a nearly convenience-service area, which includes banks, business centers, Internet services and more. At , the commercial area is twice the size of Beijing's Lufthansa Shopping Center. The terminal provides 72 aerobridges or jetways and is further complemented with remote parking bays that bring the total number of gates to 150. Terminal 3 comes with an additional runway. It increases BCIA's total capacity by 72 million passengers per year to approximately 90 million. Airbus A380 The terminal has gates and a nearby runway that can handle the Airbus A380. This capability was proven when Singapore Airlines briefly offered A380 flights to Beijing in August 2008 during the Summer Olympics. Emirates started its scheduled daily operation to Dubai on 1 August 2010. Singapore Airlines has been using an A380 since June 2014 and increased to two A380s in 2015. China Southern Airlines operated two flights to Guangzhou Baiyun Airport, Chongqing Jiangbei Airport, and Amsterdam Schiphol Airport starting from 2011, 2013, and 2015 before retiring them in 2023. Lufthansa has been using these facilities since October 2010 to handle up to five A380 flights per week. Airlines and destinations Passenger Cargo Ground transportation Intraterminal transportation Terminal 3 consists of three sub-concourses. Both domestic and international travellers check in at concourse T3C. Gates for domestic flights are in T3C, and gates for domestic flights operated by Air China are also located in concourse T3D. All international, Hong Kong, and Macau, and Taiwan flights are handled in concourse T3E. In conjunction with the construction of the new terminal, Bombardier Transportation installed a automated people mover which connects T3C and T3E via T3D in a 2–5-minute one-way trip. The line uses Innovia APM 100 vehicles running at 6-minute intervals at a maximum speed of . New Innovia APM 300 vehicles is being delivered to Beijing Capital International Airport Terminal 3 People Mover in July 2021. Interterminal transportation The airport provides a free interterminal shuttle bus between Terminals 1/2 and 3. They operate every 10 minutes from 6 am to 11 pm, and every 30 minutes from 11 pm until 6 am. Terminals 1 and 2 are connected by a lengthy corridor. Rail Beijing Capital International Airport is served by the Capital Airport Express, a dedicated rail link operated as part of the Beijing Subway system. The line runs from Terminal 3 to Terminal 2 and then to the city with stops at Sanyuanqiao and Dongzhimen before ending at Beixinqiao. The line opened on 19 July 2008, in time for the 2008 Summer Olympics, while a one-stop extension to Beixinqiao station was opened on 31 December 2021. A one-way trip takes approximately 16–20 minutes. Bus There are 18 bus routes to and from points throughout the city including Xidan, Beijing railway station, Beijing South railway station, Beijing West railway station, Zhongguancun, Fangzhuang and Shangdi. The airport buses run to each of the three terminals and cost up to ¥30 per ride depending on the route. The airport buses accept only paper tickets that are sold at each terminal and certain bus stops in the city. The airport also offers intercity bus services to and from neighboring cities including Tianjin, Qinhuangdao, Baoding, Langfang and Tangshan. Car The airport is accessible by four expresses tollways, two of which run directly from northeastern Beijing to the airport. The other two connect to the airport from nearby highways. The Airport Expressway is a toll road that runs from the northeastern 3rd Ring Road at Sanyuanqiao directly to Terminals 1 and 2. It was built in the 1990s and has served as the primary road connection to the city. The 2nd Airport Expressway, opened in 2008, is a toll road that runs east from Yaojiayuan Lu at the eastern 5th Ring Road and then north to Terminal 3. The Northern Airport Line, opened in 2006, is an toll road that runs east from the Jingcheng Expressway to Terminals 1 and 2. The Southern Airport Line, opened in 2008, is a toll road that runs parallel and to the south of the Northern Airport Line from the Jingcheng Expressway to the eastern Sixth Ring Road at the Litian Bridge. This highway crosses the Airport Expressway and 2nd Airport Expressway, and enables drivers on the former to reach Terminal 3 and the latter to head to Terminals 1 and 2. Accolades 2009 – first on the ranking of the World's Best Airport by Condé Nast Traveler magazine, based on its satisfaction survey. 2011 – third Best Airport Worldwide of the Airport Service Quality Awards by Airports Council International. 2011–2022 – ACI Director General’s Roll of Excellence by Airports Council International 2020, 2021 – best airport in the Asia-Pacific serving over 40 million passengers per year by Airports Council International 2021 – best hygiene measures in the Asia-Pacific by Airports Council International 2021 – Voice of the Customer by Airports Council International Statistics Climate Other facilities Beijing Capital Airlines has its headquarters in the Capital Airlines Building () at the airport. Accidents and incidents at or near PEK On December 5, 1968, the airport was the site of two fatal accidents in less than 24 hours; A Civil Aviation Administration of China (CAAC) Ilyushin Il-14 on approach plunged towards the ground for undetermined reasons (possibly wind shear), broke apart and caught fire, and killed 13 out of the 14 occupants on board. Another CAAC Ilyushin Il-14 crashed 1.2 km (0.8mls) from PEK during a nighttime approach in poor visibility because of an incorrect altimeter setting and the absence of the flight instructor in the cockpit during the approach. Both occupants died. On August 27, 2019, an Air China Airbus A330 caught fire while parked at the gate during boarding. All 161 passengers and crew evacuated safely, but the aircraft was substantially damaged and written off. Sister airports O'Hare International Airport Helsinki Airport Hong Kong International Airport Los Angeles International Airport Manchester Airport Munich Airport Suvarnabhumi Airport Sydney Airport Stockholm Arlanda Airport Zayed International Airport
Technology
Asia
null
291453
https://en.wikipedia.org/wiki/Renormalization
Renormalization
Renormalization is a collection of techniques in quantum field theory, statistical field theory, and the theory of self-similar geometric structures, that are used to treat infinities arising in calculated quantities by altering values of these quantities to compensate for effects of their self-interactions. But even if no infinities arose in loop diagrams in quantum field theory, it could be shown that it would be necessary to renormalize the mass and fields appearing in the original Lagrangian. For example, an electron theory may begin by postulating an electron with an initial mass and charge. In quantum field theory a cloud of virtual particles, such as photons, positrons, and others surrounds and interacts with the initial electron. Accounting for the interactions of the surrounding particles (e.g. collisions at different energies) shows that the electron-system behaves as if it had a different mass and charge than initially postulated. Renormalization, in this example, mathematically replaces the initially postulated mass and charge of an electron with the experimentally observed mass and charge. Mathematics and experiments prove that positrons and more massive particles such as protons exhibit precisely the same observed charge as the electron – even in the presence of much stronger interactions and more intense clouds of virtual particles. Renormalization specifies relationships between parameters in the theory when parameters describing large distance scales differ from parameters describing small distance scales. Physically, the pileup of contributions from an infinity of scales involved in a problem may then result in further infinities. When describing spacetime as a continuum, certain statistical and quantum mechanical constructions are not well-defined. To define them, or make them unambiguous, a continuum limit must carefully remove "construction scaffolding" of lattices at various scales. Renormalization procedures are based on the requirement that certain physical quantities (such as the mass and charge of an electron) equal observed (experimental) values. That is, the experimental value of the physical quantity yields practical applications, but due to their empirical nature the observed measurement represents areas of quantum field theory that require deeper derivation from theoretical bases. Renormalization was first developed in quantum electrodynamics (QED) to make sense of infinite integrals in perturbation theory. Initially viewed as a suspect provisional procedure even by some of its originators, renormalization eventually was embraced as an important and self-consistent actual mechanism of scale physics in several fields of physics and mathematics. Despite his later skepticism, it was Paul Dirac who pioneered renormalization. Today, the point of view has shifted: on the basis of the breakthrough renormalization group insights of Nikolay Bogolyubov and Kenneth Wilson, the focus is on variation of physical quantities across contiguous scales, while distant scales are related to each other through "effective" descriptions. All scales are linked in a broadly systematic way, and the actual physics pertinent to each is extracted with the suitable specific computational techniques appropriate for each. Wilson clarified which variables of a system are crucial and which are redundant. Renormalization is distinct from regularization, another technique to control infinities by assuming the existence of new unknown physics at new scales. Self-interactions in classical physics The problem of infinities first arose in the classical electrodynamics of point particles in the 19th and early 20th century. The mass of a charged particle should include the mass–energy in its electrostatic field (electromagnetic mass). Assume that the particle is a charged spherical shell of radius . The mass–energy in the field is which becomes infinite as . This implies that the point particle would have infinite inertia and thus cannot be accelerated. Incidentally, the value of that makes equal to the electron mass is called the classical electron radius, which (setting and restoring factors of and ) turns out to be where is the fine-structure constant, and is the reduced Compton wavelength of the electron. Renormalization: The total effective mass of a spherical charged particle includes the actual bare mass of the spherical shell (in addition to the mass mentioned above associated with its electric field). If the shell's bare mass is allowed to be negative, it might be possible to take a consistent point limit. This was called renormalization, and Lorentz and Abraham attempted to develop a classical theory of the electron this way. This early work was the inspiration for later attempts at regularization and renormalization in quantum field theory. (
Physical sciences
Physics basics: General
Physics
291692
https://en.wikipedia.org/wiki/Truffle
Truffle
A truffle is the fruiting body of a subterranean ascomycete fungus, one of the species of the genus Tuber. More than one hundred other genera of fungi are classified as truffles including Geopora, Peziza, Choiromyces, and Leucangium. These genera belong to the class Pezizomycetes and the Pezizales order. Several truffle-like basidiomycetes are excluded from Pezizales, including Rhizopogon and Glomus. Truffles are ectomycorrhizal fungi, so they are found in close association with tree roots. Spore dispersal is accomplished through fungivores, animals that eat fungi. These fungi have ecological roles in nutrient cycling and drought tolerance. Some truffle species are prized as food. Edible truffles are used in Italian, French and other national . Truffles are cultivated and harvested from natural environments. Taxonomic details Species Black The black truffle or black Périgord truffle (Tuber melanosporum), the second-most commercially valuable species, is named after the Périgord region in France. Black truffles associate with oaks, hazelnut, cherry, and other deciduous trees and are harvested in late autumn and winter. The genome sequence of the black truffle was published in March 2010. Summer or Burgundy The black summer truffle (Tuber aestivum) is found across Europe and is prized for its culinary value. Burgundy truffles (designated Tuber uncinatum, but the same species) are harvested in autumn until December and have aromatic flesh of a darker colour. These are associated with various trees and shrubs. White Tuber magnatum, the high-value white truffle () is found mainly in the Langhe and Montferrat areas of the Piedmont region in northern Italy, and most famously, in the countryside around the cities of Alba and Asti. A large percentage of Italy's white truffles also come from Molise. Whitish The "whitish truffle" (Tuber borchii) is a similar species native to Tuscany, Abruzzo, Romagna, Umbria, the Marche, and Molise. It is reportedly not as aromatic as those from Piedmont, although those from Città di Castello are said to come quite close. Other Tuber A less common truffle is "garlic truffle" (Tuber macrosporum). In the U.S. Pacific Northwest, several species of truffle are harvested both recreationally and commercially, most notably, the Leucangium carthusianum, Oregon black truffle; Tuber gibbosum, Oregon spring white truffle; and Tuber oregonense, Oregon winter white truffle. Kalapuya brunnea, the Oregon brown truffle, has also been commercially harvested and is of culinary note. The Oregon white truffle is increasingly harvested due to its high quality and also exported to other countries. Oregon celebrates its traditional truffle harvesting with a 'truffle festival', combined with culinary shows and wine tastings. The pecan truffle (Tuber lyonii) syn. texense is found in the Southern United States, usually associated with pecan trees. Chefs who have experimented with them agree "they are very good and have potential as a food commodity". Although pecan farmers used to find them along with pecans and discard them, considering them a nuisance, they sell for about $160 a pound and have been used in some gourmet restaurants. Beyond Tuber The term "truffle" has been applied to several other genera of similar underground fungi. The genera Terfezia and Tirmania of the family Terfeziaceae are known as the "desert truffles" of Africa and the Middle East. Pisolithus tinctorius, which was historically eaten in parts of Germany, is sometimes called "Bohemian truffle". Geopora spp. are important ectomycorrhizal partners of trees in woodlands and forests throughout the world. Pinus edulis, a widespread pine species of the Southwest US, is dependent on Geopora for nutrient and water acquisition in arid environments. Like other truffle fungi, Geopora produces subterranean sporocarps as a means of sexual reproduction. Geopora cooperi, also known as pine truffle or fuzzy truffle, is an edible species of this genus. Rhizopogon spp. are ectomycorrhizal members of the Basidiomycota and the order Boletales, a group of fungi that typically form mushrooms. Like their ascomycete counterparts, these fungi can create truffle-like fruiting bodies. Rhizopogon spp. are ecologically important in coniferous forests where they associate with various pines, firs, and Douglas fir. In addition to their ecological importance, these fungi hold economic value, as well. Rhizopogon spp. are commonly used to inoculate coniferous seedlings in nurseries and during reforestation. Hysterangium spp. are ectomycorrhizal members of the Basidiomycota and the order Hysterangiales that form sporocarps similar to true truffles. These fungi form mycelial mats of vegetative hyphae that may cover 25–40% of the forest floor in Douglas fir forests, thereby contributing to a significant portion of the biomass present in soils. Like other ectomycorrhizal fungi, Hysterangium spp. play a role in nutrient exchange in the nitrogen cycle by accessing nitrogen unavailable to host plants and acting as nitrogen sinks in forests. Glomus spp. are arbuscular mycorrhizae of the phylum Glomeromycota within the order Glomerales. Members of this genus have low host specificity, associating with a variety of plants including hardwoods, forbs, shrubs, and grasses. These fungi commonly occur throughout the Northern Hemisphere. Members of the genus Elaphomyces are commonly mistaken for truffles. Phylogeny Phylogenetic analysis has demonstrated the convergent evolution of the ectomycorrhizal trophic mode in diverse fungi. The subphylum Pezizomycotina, containing the order Pezizales, is approximately 400 million years old. Within the order Pezizales, subterranean fungi evolved independently at least fifteen times. Contained within Pezizales are the families Tuberaceae, Pezizaceae, Pyronematacae, and Morchellaceae. All of these families contain lineages of subterranean or truffle fungi. The oldest ectomycorrhizal fossils are from the Eocene about 50 million years ago. The specimens are preserved permineralized in-situ in the Eocene Okanagan Highlands Princeton chert site. This indicates that the soft bodies of ectomycorrhizal fungi do not easily fossilize. Molecular clockwork has suggested the evolution of ectomycorrhizal fungi occurred approximately 130 million years ago. The evolution of subterranean fruiting bodies has occurred numerous times within the Ascomycota, Basidiomycota, and Glomeromycota. For example, the genera Rhizopogon and Hysterangium of Basidiomycota both form subterranean fruiting bodies and play similar ecological roles as truffle forming ascomycetes. The ancestors of the Ascomycota genera Geopora, Tuber, and Leucangium originated in Laurasia during the Paleozoic era. Phylogenetic evidence suggests that most subterranean fruiting bodies evolved from above-ground mushrooms. Over time mushroom stipes and caps were reduced, and caps began to enclose reproductive tissue. The dispersal of sexual spores then shifted from wind and rain to utilising animals. The phylogeny and biogeography of the genus Tuber was investigated in 2008 using internal transcribed spacers (ITS) of nuclear DNA and revealed five major clades (Aestivum, Excavatum, Rufum, Melanosporum and Puberulum); this was later improved and expanded in 2010 to nine major clades using 28S large subunits (LSU) rRNA of mitochondrial DNA. The Magnatum and Macrosporum clades were distinguished as distinct from the Aestivum clade. The Gibbosum clade was resolved as distinct from all other clades, and the Spinoreticulatum clade was separated from the Rufum clade. The truffle habit has evolved independently among several basidiomycete genera. Phylogenetic analysis has revealed that basidiomycete subterranean fruiting bodies, like their ascomycete counterparts, evolved from above ground mushrooms. For example, Rhizopogon species likely arose from an ancestor shared with Suillus, a mushroom-forming genus. Studies have suggested that selection for subterranean fruiting bodies among ascomycetes and basidiomycetes occurred in water-limited environments. Etymology Most sources agree that the term "truffle" is derived from the Latin term by way of the Vulgar Latin , meaning "swelling" or "lump". This then entered other languages through Old French dialects. Ecology The mycelia of truffles form symbiotic, mycorrhizal relationships with the roots of several tree species, including beech, birch, hazel, hornbeam, oak, pine, and poplar. Mutualistic ectomycorrhizal fungi such as truffles provide valuable nutrients to plants in exchange for carbohydrates. Ectomycorrhizal fungi cannot survive in the soil without their plant hosts. In fact, many of these fungi have lost the enzymes necessary for obtaining carbon through other means. For example, truffle fungi have lost their ability to degrade the cell walls of plants, limiting their capacity to decompose plant litter. Plant hosts can also depend on their associated truffle fungi. Geopora, Peziza, and Tuber spp. are vital in the establishment of oak communities. Tuber species prefer argillaceous or calcareous soils that are well drained and neutral or alkaline. Tuber truffles fruit throughout the year, depending on the species, and can be found buried between the leaf litter and the soil. Most fungal biomass is found in the humus and litter layers of soil. Most truffle fungi produce both asexual spores (mitospores or conidia) and sexual spores (meiospores or ascospores/basidiospores). Conidia can be produced more readily and with less energy than ascospores and can disperse during disturbance events. Production of ascospores is energy intensive because the fungus must allocate resources to the production of large sporocarps. Ascospores are borne within sac-like structures called asci, which are contained within the sporocarp. Because truffle fungi produce their sexual fruiting bodies underground, spores cannot be spread by wind and water. Therefore, nearly all truffles depend on mycophagous animal vectors for spore dispersal. This is analogous to the dispersal of seeds in fruit of angiosperms. When the ascospores are fully developed, the truffle exudes volatile compounds that attract animal vectors. For successful dispersal, these spores must survive passage through the digestive tracts of animals. Ascospores have thick walls composed of chitin to help them endure the environment of animal guts. Animal vectors include birds, deer, and rodents such as voles, squirrels, and chipmunks. Many species of trees, such as Quercus garryana, are dependent on the dispersal of sporocarps to inoculate isolated individuals. For example, the acorns of Q. garryana may be carried to new territory that lacks the necessary mycorrhizal fungi for establishment. Some mycophagous animals depend on truffles as their dominant food source. Flying squirrels, Glaucomys sabrinus, of North America play a three-way symbiosis with truffles and their associated plants. G. sabrinus is particularly adapted to finding truffles using its refined sense of smell, visual clues, and long-term memory of prosperous populations of truffles. This intimacy between animals and truffles indirectly influences the success of mycorrhizal plant species. After ascospores are dispersed, they remain dormant until germination is initiated by exudates excreted from host plant roots. Following germination, hyphae form and seek out the roots of host plants. Arriving at roots, hyphae begin to form a mantle or sheath on the outer surface of root tips. Hyphae then enter the root cortex intercellularly to form the Hartig net for nutrient exchange. Hyphae can spread to other root tips colonising the entire root system of the host. Over time, the truffle fungus accumulates sufficient resources to form fruiting bodies. Rate of growth is correlated with increasing photosynthetic rates in the spring as trees leaf out. Nutrient exchange Truffle fungi receive carbohydrates from their host plants, providing them with valuable micro- and macronutrients. Plant macronutrients include potassium, phosphorus, nitrogen, and sulfur. In contrast, micronutrients include iron, copper, zinc, and chloride. In truffle fungi, as in all ectomycorrhizae, the majority of nutrient exchange occurs in the Hartig net, the intercellular hyphal network between plant root cells. A unique feature of ectomycorrhizal fungi is the formation of the mantle on the outer surface of fine roots. Truffles have been suggested to co-locate with the orchid species Epipactis helleborine and Cephalanthera damasonium, though this is not always the case. Nutrient cycling Truffle fungi are ecologically important in nutrient cycling. Plants obtain nutrients via their fine roots. Mycorrhizal fungi are much smaller than fine roots, so they have a higher surface area and a greater ability to explore soils for nutrients. Acquisition of nutrients includes the uptake of phosphorus, nitrate or ammonium, iron, magnesium, and other ions. Many ectomycorrhizal fungi form fungal mats in the upper layers of soils surrounding host plants. These mats have significantly higher carbon and fixed nitrogen concentrations than surrounding soils. Because these mats are nitrogen sinks, leaching of nutrients is reduced. Mycelial mats can also help maintain the structure of soils by holding organic matter in place and preventing erosion. Often, these networks of mycelium provide support for smaller organisms in the soil, such as bacteria and microscopic arthropods. Bacteria feed on the exudates released by mycelium and colonise the soil surrounding them. Microscopic arthropods such as mites feed directly on mycelium and release valuable nutrients for the uptake of other organisms. Thus, truffle fungi and other ectomycorrhizal fungi facilitate a complex system of nutrient exchange between plants, animals, and microbes. Importance in arid-land ecosystems Plant community structure is often affected by the availability of compatible mycorrhizal fungi. In arid-land ecosystems, these fungi become essential for the survival of their host plants by enhancing the ability to withstand drought. A foundation species in arid-land ecosystems of the Southwest United States is Pinus edulis, commonly known as pinyon pine. P. edulis associates with the subterranean fungi Geopora and Rhizopogon. As global temperatures rise, so does the occurrence of severe droughts, detrimentally affecting the survival of aridland plants. This variability in climate has increased the mortality of P. edulis. Therefore, the availability of compatible mycorrhizal inoculum can greatly affect the successful establishment of P. edulis seedlings. Associated ectomycorrhizal fungi will likely play a significant role in the survival of P. edulis with continuing global climate change. Extraction Because truffles are subterranean, they are often located with the help of an animal (sometimes called a truffler) possessing a refined sense of smell. Traditionally, pigs have been used to extract truffles. Both the female pig's natural truffle-seeking and her intent to eat the truffle were thought to be due to a compound within the truffle similar to androstenol, the sex pheromone of boar saliva, to which the sow is keenly attracted. Studies in 1990 demonstrated that the compound actively recognised by both truffle pigs and dogs is dimethyl sulfide. In Italy, the use of pigs to hunt truffles has been prohibited since 1985 because of damage caused by animals to truffle mycelia during the digging that dropped the production rate of the area for some years. An alternative to truffle pigs are dogs. Dogs offer an advantage because they do not have a strong desire to eat truffles, so they can be trained to locate sporocarps without digging them up. Pigs attempt to dig up truffles. Fly species of the genus Suillia can also detect the volatile compounds associated with subterranean fruiting bodies. These flies lay their eggs above truffles to provide food for their young. At ground level, Suilla flies can be seen flying above truffles. Volatile constituents The mycelia or fruiting bodies release the volatile constituents responsible for the natural aroma of truffles or derive from truffle-associated microbes. The chemical ecology of truffle volatiles is complex, interacting with plants, insects, and mammals, which contribute to spore dispersal. Depending on the truffle species, lifecycle, or location, they include: Sulfur volatiles, which occur in all truffle species, such as dimethyl mono- (DMS), di- (DMDS) and tri- (DMTS) sulfides, as well as 2-methyl-4,5-dihydrothiophene, characteristic of the white truffle T. borchii and 2,4-Dithiapentane occurring in all species but mostly characteristic of the white truffle T. magnatum. Some very aromatic white truffles are notably pungent, even irritating the eye when cut or sliced. Metabolites of nonsulfur amino acid constituents (simple and branched-chain hydrocarbons) such as ethylene (produced by mycelia of white truffles affecting root architecture of host tree), as well as 2-methylbutanal, 2-methylpropanal, and 2-phenylethanol (also common in baker's yeast). Fatty acid-derived volatiles (C8-alcohols and aldehydes with a characteristic fungal odor, such as 1-octen-3-ol and 2-octenal). The former is derived from linoleic acid and produced by mature white truffle T. borchii. Thiophene derivatives appear to be produced by bacterial symbionts living in the truffle body. The most abundant of these, 3-methyl, 4-5 dihydrothiophene, contributes to the white truffle's aroma. Several truffle species and varieties are differentiated based on their relative contents or absence of sulfides, ethers or alcohols, respectively. The sweaty-musky aroma of truffles is similar to that of the pheromone androstenol that also occurs in humans. , the volatile profiles of seven black and six white truffle species have been studied. Cultivation Truffles long eluded techniques of cultivation, as Jean-Anthelme Brillat-Savarin (1825) noted: The most learned men have sought to ascertain the secret and fancied they discovered the seed. Their promises, however, were vain, and no planting was ever followed by a harvest. This perhaps is all right, for as one of the great values of truffles is their dearness, perhaps they would be less highly esteemed if they were cheaper. Truffles can be cultivated. As early as 1808, attempts to cultivate truffles, known in French as , were successful. People had long observed that truffles were growing among the roots of certain trees, and in 1808, Joseph Talon, from Apt ( of Vaucluse) in southern France, had the idea of transplanting some seedlings that he had collected at the foot of oak trees known to host truffles in their root system. For discovering how to cultivate truffles, some sources now give priority to Pierre II Mauléon (1744–1831) of Loudun (in western France), who began to cultivate truffles around 1790. Mauléon saw an "obvious symbiosis" between the oak tree, the rocky soil, and the truffle and attempted to reproduce such an environment by taking acorns from trees known to have produced truffles and sowing them in chalky soil. His experiment was successful, with truffles found in the soil around the newly grown oak trees years later. In 1847, Auguste Rousseau of Carpentras (in Vaucluse) planted of oak trees (again from acorns found on the soil around truffle-producing oak trees), and he subsequently obtained large harvests of truffles. He received a prize at the 1855 World's Fair in Paris. Others imitated these successful attempts in France and Italy. In the late 19th century, an epidemic of phylloxera destroyed many of the vineyards in southern France. Another epidemic killed most of the silkworms there, too, making the fields of mulberry trees useless. Trufficulture became an important source of income for those affected. The calcareous and exposed vineyard soils were well-suited to the cultivation of truffles. By 1890, truffières (truffle plantations) covered 750 km2 of land in France, and 2,000 tonnes of truffles were produced in that year. From the 19th century to the present, truffle production fell by 97–99% to 20–50 tonnes annually. Reasons given for this decline include the Industrial Revolution, the subsequent rural flight and the multiple European wars of the 20th century, which reduced the rural population. For example, World War I resulted in the mobilisation of 65% of the agricultural workers from the region of Lot alone. Knowledge of truffle cultivation, the soil and the seasons was lost along with the people. Another consequence was no more grazing sheep or shepherds who pruned trees for feed and fuelwood, so former truffle plantations turned into closed forests that no longer produced truffles. Truffles were once sold at weekly markets (bi-weekly in the case of a market in Martel, Lot) and in quantities of two to six tonnes in good weeks, but only Lalbenque and Limogne today have weekly truffle markets. Prices have increased so that truffles, once seen as a food of the middle class, have become a luxury. The situation changed in the late 1960s and early 1970s, with researchers in France and Italy establishing mycorrhizas with truffle spores. Beginning from the 1980s, truffle plantations are compensating for some of the decline in wild truffles, and exist in various countries including France, Italy, Spain and Australia. About 80% of the truffles now produced in France come from specially planted truffle groves. Investments in cultivated plantations are underway in many parts of the world using controlled irrigation for regular and resilient production. A critical phase of the cultivation is the quality control of the mycorrhizal plants. Between 7 and 10 years are needed for the truffles to develop their mycorrhizal network, and only after that do the host plants come into production. Complete soil analysis to avoid contamination by other dominant fungi and very strict control of the formation of mycorrhizae are necessary to ensure the success of a plantation. Total investment per hectare for an irrigated and barrier-sealed plantation (against wild boars) can cost up to €10,000. Considering the level of initial investment and the maturity delay, farmers who have not taken care of both soil conditions and seedling conditions are at high risk of failure. New Zealand and Australia The first black truffles (Tuber melanosporum) to be produced in the Southern Hemisphere were harvested in Gisborne, New Zealand in 1993. New Zealand's first burgundy truffle was found in July 2012 at a Waipara truffle farm. It weighed 330 g and was found by the farm owner's beagle. In 1999, the first Australian truffles were harvested in Tasmania, the result of eight years of work. Trees were inoculated with the truffle fungus to create a local truffle industry. Their success and the value of the resulting truffles has encouraged a small industry to develop. Truffle production has expanded into the colder regions of Victoria, New South Wales and Western Australia. In 2014, over of truffles were harvested by Truffle Hill, Manjimup, Western Australia. In June 2014, a grower harvested Australia's largest truffle from their property at Robertson, in the Southern Highlands of New South Wales. It was a French black périgord fungus weighing in at and was valued at over $2,000 per kilogram. United States Périgord truffles were first farmed in Tennessee in 2007. At its peak in the 2008–2009 season, one farm produced about 200 pounds of truffles, but Eastern filbert blight almost entirely wiped out the hazel trees by 2013 and production dropped, essentially ending the business. Eastern filbert blight similarly destroyed the orchards of other once-promising commercial farms in East Tennessee, while newer farms in California, North Carolina, Oregon, and Arkansas were started. , the Appalachian truffle (Tuber canaliculatum) was being developed as a potential market. Uses Because of their high price and their strong aroma, truffles are used sparingly. Supplies can be found commercially as unadulterated fresh produce or preserved, typically in a light brine. Their chemical compounds infuse well with fats such as butter, cream, cheeses, avocados, and coconut cream. As the volatile aromas dissipate quicker when heated, truffles are generally served raw and shaved over warm, simple foods which highlight their flavour, such as buttered pasta or eggs. Thin truffle slices may be inserted into meats, under the skins of roasted fowl, in foie gras preparations, in pâtés, or in stuffings. Some speciality cheeses contain truffles, as well. Truffles are also used for producing truffle salt and truffle honey. While chefs once peeled truffles, in modern times, most restaurants brush the truffle carefully and shave it or dice it with the skin on to make the most of the valuable ingredient. Some restaurants stamp out circular discs of truffle flesh and use the skins for sauces. Oil Truffle oil is used as a lower-cost and convenient substitute for truffles, to provide flavouring, or to enhance the flavour and aroma of truffles in cooking. Some products called "truffle oils" contain no truffles or include pieces of inexpensive, unprized truffle varietals, which have no culinary value, simply for show. The vast majority is oil that has been artificially flavoured using a synthetic agent such as 2,4-dithiapentane. The scientific name is included on the ingredient list of truffle oils infused with natural truffles. Vodka Because more aromatic molecules in truffles are soluble in alcohol, they can carry a more complex and accurate truffle flavour than oil without synthetic flavourings. Many commercial producers use 2,4-dithiapentane regardless, as it has become the dominant flavour most consumers, unexposed to fresh truffles but familiar with oils, associate with them. Because most Western nations do not have ingredient labelling requirements for spirits, consumers often do not know if artificial flavourings have been used. It is used as a spirit in its own right, a cocktail mix or a food flavouring. Cultural history Antiquity The first mention of truffles appears in the inscriptions of the neo-Sumerians regarding their Amorite enemy's eating habits (Third Dynasty of Ur, 2nd century BCE) and later in writings of Theophrastus in the 4th century BCE. In classical times, their origins were a mystery that challenged many; Plutarch and others thought them to be the result of lightning, warmth, and water in the soil, while Juvenal thought thunder and rain to be instrumental in their origin. Cicero deemed them children of the earth, while Dioscorides thought they were tuberous roots. Rome and Thracia in the Classical period identified three kinds of truffles: Tuber melanosporum, T. magnificus, and T. magnatum. The Romans instead used a variety of fungus called terfez, also sometimes called a "desert truffle". Terfez used in Rome came from Lesbos, Carthage, and especially Libya, where the coastal climate was less dry in ancient times. Their substance is pale, tinged with rose. Unlike truffles, terfez have little inherent flavour. The Romans used the terfez as a flavour carrier because the terfez tends to absorb surrounding flavours. Because Ancient Roman cuisine used many spices and flavourings, the terfez may have been appropriate in that context. Middle Ages Truffles were rarely used during the Middle Ages. Truffle hunting is mentioned by Bartolomeo Platina, the papal historian, in 1481, when he recorded that the sows of Notza were without equal in hunting truffles, but they should be muzzled to prevent them from eating the prize. Renaissance and modernity During the Renaissance, truffles regained popularity in Europe and were honoured at the court of King Francis I of France. They were popular in Parisian markets in the 1780s, imported seasonally from truffle grounds, where peasants had long enjoyed them. Brillat-Savarin (1825) noted that they were so expensive they appeared only at the dinner tables of great nobles and kept women. They were sometimes served with turkey.
Biology and health sciences
Edible fungi
null
291838
https://en.wikipedia.org/wiki/Catheter
Catheter
In medicine, a catheter ( ) is a thin tube made from medical grade materials serving a broad range of functions. Catheters are medical devices that can be inserted in the body to treat diseases or perform a surgical procedure. Catheters are manufactured for specific applications, such as cardiovascular, urological, gastrointestinal, neurovascular and ophthalmic procedures. The process of inserting a catheter is called catheterization. In most uses, a catheter is a thin, flexible tube (soft catheter) though catheters are available in varying levels of stiffness depending on the application. A catheter left inside the body, either temporarily or permanently, may be referred to as an "indwelling catheter" (for example, a peripherally inserted central catheter). A permanently inserted catheter may be referred to as a "permcath" (originally a trademark). Catheters can be inserted into a body cavity, duct, or vessel, brain, skin or adipose tissue. Functionally, they allow drainage, administration of fluids or gases, access by surgical instruments, and also perform a wide variety of other tasks depending on the type of catheter. Special types of catheters, also called probes, are used in preclinical or clinical research for sampling of lipophilic and hydrophilic compounds, protein-bound and unbound drugs, neurotransmitters, peptides and proteins, antibodies, nanoparticles and nanocarriers, enzymes and vesicles. Etymology "Catheter" (from Greek kathetḗr) comes from the Greek verb kathíemai, meaning "to thrust into" or "to send down" because the catheter allowed fluid to be "sent down" from the body. Uses Placement of a catheter into a particular part of the body may allow: Draining urine from the urinary bladder as in urinary catheterization, using intermittent catheters or Foley catheter inserted through urethra. When the urethra is damaged, suprapubic catheterisation is used instead. The suprapubic catheter is inserted through the lower part of the abdomen directly into the urinary bladder. drainage of urine from the kidney by percutaneous (through the skin) nephrostomy drainage of fluid collections, e.g. an abdominal abscess pigtail catheter: used to drain air from around the lung (pneumothorax) administration of intravenous fluids, medication or parenteral nutrition with a peripheral venous catheter or central venous catheter angioplasty, angiography, balloon septostomy, balloon sinuplasty, cardiac electrophysiology testing, catheter ablation. Often the Seldinger technique is used. direct measurement of blood pressure in an artery or vein direct measurement of intracranial pressure administration of anaesthetic medication into the epidural space, the subarachnoid space, or around a major nerve bundle such as the brachial plexus transfer of fertilized embryos, from in vitro fertilization, or sperm, during artificial insemination, into the uterus administration of oxygen, volatile anesthetic agents, and other breathing gases into the lungs using a tracheal tube subcutaneous administration of insulin or other medications, with the use of an infusion set and insulin pump History Ancient inventors Ancient Chinese used onion stalks, the Romans, Hindus, and Greeks used tubes of wood or precious metals. The ancient Egyptians created catheters from reeds. Modern The earliest invention of the flexible catheter was during the 18th century. Extending his inventiveness to his family's medical problems, Benjamin Franklin invented the flexible catheter in 1752 when his brother John suffered from bladder stones. Franklin's catheter was made of metal with segments hinged together with a wire enclosed to provide rigidity during insertion. According to a footnote in his letter in Volume 4 of the Papers of Benjamin Franklin (1959), Franklin credits Francesco Roncelli-Pardino from 1720 as the inventor of a flexible catheter. In fact, Franklin claims the flexible catheter may have been designed even earlier. An early modern application of the catheter was employed by Claude Bernard for the purpose of cardiac catheterization in 1844. The procedure involved entering a horse's ventricles via the jugular vein and carotid artery. In 1929, Werner Forssman first performed central venous catheterization, work which led to the development of cardiac catherization as a treatment, for which he, André F. Cournand and Dickinson W. Richards would win the Nobel Prize for Medicine in 1959. Central venous catheterization allows for continuous administration of medications, fluids and blood products to a large vein, particularly in critically ill patients. Cardiac catheterization is the insertion of a catheter into one of the chambers of the heart, which is used for imaging, diagnosis, and the placement of devices such as stents. David S. Sheridan invented the modern disposable catheter in the 1940s. Prior to this, some reusable catheters consisted of braided cotton tubes, which were varnished, heat-treated and polished. As these were primarily produced in France, the advent of World War II threatened the supply chain. Sheridan was dubbed the "Catheter King" by Forbes magazine in 1988. He also invented the modern "disposable" plastic endotracheal tube now used routinely in surgery. Other reusable catheters consisted of red rubber tubes. Although sterilized prior to reuse, they still posed a high risk of infection and often led to the spread of disease. To prevent clotting, catheters that are not in use may be filled with catheter lock solution. Materials Urinary catheters A range of polymers are used for the construction of catheters, including silicone rubber, nylon, polyurethane, polyethylene terephthalate (PET), latex, and thermoplastic elastomers. Silicone is one of the most common implantable choice because it is inert and unreactive to body fluids and a range of medical fluids with which it might come into contact. On the other hand, the polymer is weak mechanically, and a number of serious fractures have occurred in catheters. For example, silicone is used in Foley catheters where fractures have been reported, often requiring surgery to remove the tip left in the bladder. Catheters used in interventional procedures Depending on the mechanical characteristics required, assorted polymers and polymer-metal composites can be used to build catheters used for interventional purposes. Common materials include polyamide (nylon), polyether block amide, polyuerathane, polyethylene terephthalate, and polyimides. These materials are often used in combination with each other and are frequently layered on top of stainless steel braiding, laser-cut stainless steel tubing, or other scaffold-like structures to impart desirable handling characteristics to the catheter, all dependent on the intended application. For example, the materials and the architectures used to manufacture vascular catheters for neurological applications might differ significantly from catheters destined for cardiovascular use. Guiding catheters (catheters that guides angioplasty balloons and stents) is made up of polytetrafluoroethylene (PTFE) innermost layer which is lubricious, followed by stainless steel braid wire outer layer which helps to provide support for the catheter and prevent kinking while travelling through blood vessels, and Nylon elastomer outermost layer which provides extra support for the catheter and preserve the curvature of the catheter while passing through tortuous vessels. To enhance ease of insertion, some catheters have a lubricious surface coating to lessen friction. A lubricious coating creates a smooth, slippery film making the catheter easier to insert. Interventional procedures Diagnostic catheters There are various catheters used in angiography procedures. Diagnostic catheters direct wires through blood vessels. Radiocontrast agent is then injected through the catheter to visualise the vessels via various imaging methods such as computed tomography (CT), projectional radiography, and fluoroscopy. Pigtail catheter is a non-selective catheter with multiple side holes that can deliver large volumes of contrast into a blood vessel for imaging purposes. Cobra catheter is a selective catheter used to catheterise downgoing vessels in the abdomen. Cobra catheters move forward by pushing and are removed by pulling. Sidewinder catheter is a selective catheter is used to navigate the aorta. Headhunter, Newton, Simmons, Bentson, and Berenstein catheters are used to navigate the into one of the three branches of the arch of aorta. Yashiro Catheter is a selective, hydrophilic catheter designed for optimal entry into celiac trunk. Whereas endothelial cell sampling through endovascular sampling with coils, stents, stent retrievers, or guidewires suffer from poor selectivity and a low or highly variable cell yield, a micro-3D-printed device adapted for endovascular techniques can harvest endothelial cells for transcriptomic analysis. Balloon catheters There are also balloon catheters used in angioplasty procedures such as plain balloon catheters that is useful in passing tight vessel stenosis, drug coated balloons that contains paclitaxel on the surface to prevent smooth muscle cells proliferation of the vessel walls, thus reducing the likehood of vessel blockage in the future, high pressure balloons that can open stubborn vessel stenoses in veins and arteriovenous fistula, and cutting balloon angioplasty that contains 3 to 4 small blades on its surface (endotomes) that helps to control the distribution of balloon dilatation more uniformly and cut through resistant stenosis due to fibrous scar tissue. Dialysis catheters There is no difference in achieving adequacy of blood flow, period of catheter usage, infection, and thromboembolism risk whether the dialysis catheter has step-tip, split-tip, or symmetrical tip. Palidrome catheter is superior to Permcath catheter in terms of maximum blood flow, dialysis adequacy, and annual patency rate. Similar to Permcath, Palidrome catheter has high infection and thromboembolism rate. Adverse effects In interventional procedures, Teflon catheters (which are hydrophobic) have higher risk of thrombus formation when compared to polyurethene catheters. The longer the duration of the catheter left inside the body, the higher the risk of thrombus formation. Larger catheters increase the risk of thrombus formation around the catheter, because they can block the flow of blood. "Any foreign object in the body carries an infection risk, and a catheter can serve as a superhighway for bacteria to enter the bloodstream or body", according to Milisa Manojlovich, a professor at the University of Michigan School of Nursing. Catheters can be difficult to clean, and therefore harbor antibiotic resistant or otherwise pathogenic bacteria.
Technology
Devices
null
291912
https://en.wikipedia.org/wiki/Introduction%20to%20gauge%20theory
Introduction to gauge theory
A gauge theory is a type of theory in physics. The word gauge means a measurement, a thickness, an in-between distance (as in railroad tracks), or a resulting number of units per certain parameter (a number of loops in an inch of fabric or a number of lead balls in a pound of ammunition). Modern theories describe physical forces in terms of fields, e.g., the electromagnetic field, the gravitational field, and fields that describe forces between the elementary particles. A general feature of these field theories is that the fundamental fields cannot be directly measured; however, some associated quantities can be measured, such as charges, energies, and velocities. For example, say you cannot measure the diameter of a lead ball, but you can determine how many lead balls, which are equal in every way, are required to make a pound. Using the number of balls, the density of lead, and the formula for calculating the volume of a sphere from its diameter, one could indirectly determine the diameter of a single lead ball. In field theories, different configurations of the unobservable fields can result in identical observable quantities. A transformation from one such field configuration to another is called a gauge transformation; the lack of change in the measurable quantities, despite the field being transformed, is a property called gauge invariance. For example, if you could measure the color of lead balls and discover that when you change the color, you still fit the same number of balls in a pound, the property of "color" would show gauge invariance. Since any kind of invariance under a field transformation is considered a symmetry, gauge invariance is sometimes called gauge symmetry. Generally, any theory that has the property of gauge invariance is considered a gauge theory. For example, in electromagnetism the electric field E and the magnetic field B are observable, while the potentials V ("voltage") and A (the vector potential) are not. Under a gauge transformation in which a constant is added to V, no observable change occurs in E or B. With the advent of quantum mechanics in the 1920s, and with successive advances in quantum field theory, the importance of gauge transformations has steadily grown. Gauge theories constrain the laws of physics, because all the changes induced by a gauge transformation have to cancel each other out when written in terms of observable quantities. Over the course of the 20th century, physicists gradually realized that all forces (fundamental interactions) arise from the constraints imposed by local gauge symmetries, in which case the transformations vary from point to point in space and time. Perturbative quantum field theory (usually employed for scattering theory) describes forces in terms of force-mediating particles called gauge bosons. The nature of these particles is determined by the nature of the gauge transformations. The culmination of these efforts is the Standard Model, a quantum field theory that accurately predicts all of the fundamental interactions except gravity. History and importance The earliest field theory having a gauge symmetry was James Clerk Maxwell's formulation, in 1864–65, of electrodynamics in "A Dynamical Theory of the Electromagnetic Field". The importance of this symmetry remained unnoticed in the earliest formulations. Similarly unnoticed, David Hilbert had derived Einstein's equations of general relativity by postulating a symmetry under any change of coordinates, just as Einstein was completing his work. Later Hermann Weyl, inspired by success in Einstein's general relativity, conjectured (incorrectly, as it turned out) in 1919 that invariance under the change of scale or "gauge" (a term inspired by the various track gauges of railroads) might also be a local symmetry of electromagnetism. Although Weyl's choice of the gauge was incorrect, the name "gauge" stuck to the approach. After the development of quantum mechanics, Weyl, Vladimir Fock and Fritz London modified their gauge choice by replacing the scale factor with a change of wave phase, and applying it successfully to electromagnetism. Gauge symmetry was generalized mathematically in 1954 by Chen Ning Yang and Robert Mills in an attempt to describe the strong nuclear forces. This idea, dubbed Yang–Mills theory, later found application in the quantum field theory of the weak force, and its unification with electromagnetism in the electroweak theory. The importance of gauge theories for physics stems from their tremendous success in providing a unified framework to describe the quantum-mechanical behavior of electromagnetism, the weak force and the strong force. This gauge theory, known as the Standard Model, accurately describes experimental predictions regarding three of the four fundamental forces of nature. In classical physics Electromagnetism Historically, the first example of gauge symmetry to be discovered was classical electromagnetism. A static electric field can be described in terms of an electric potential (voltage, ) that is defined at every point in space, and in practical work it is conventional to take the Earth as a physical reference that defines the zero level of the potential, or ground. But only differences in potential are physically measurable, which is the reason that a voltmeter must have two probes, and can only report the voltage difference between them. Thus one could choose to define all voltage differences relative to some other standard, rather than the Earth, resulting in the addition of a constant offset. If the potential is a solution to Maxwell's equations then, after this gauge transformation, the new potential is also a solution to Maxwell's equations and no experiment can distinguish between these two solutions. In other words, the laws of physics governing electricity and magnetism (that is, Maxwell equations) are invariant under gauge transformation. Maxwell's equations have a gauge symmetry. Generalizing from static electricity to electromagnetism, we have a second potential, the magnetic vector potential A, which can also undergo gauge transformations. These transformations may be local. That is, rather than adding a constant onto V, one can add a function that takes on different values at different points in space and time. If A is also changed in certain corresponding ways, then the same E (electric) and B (magnetic) fields result. The detailed mathematical relationship between the fields E and B and the potentials V and A is given in the article Gauge fixing, along with the precise statement of the nature of the gauge transformation. The relevant point here is that the fields remain the same under the gauge transformation, and therefore Maxwell's equations are still satisfied. Gauge symmetry is closely related to charge conservation. Suppose that there existed some process by which one could briefly violate conservation of charge by creating a charge q at a certain point in space, 1, moving it to some other point 2, and then destroying it. We might imagine that this process was consistent with conservation of energy. We could posit a rule stating that creating the charge required an input of energy E1=qV1 and destroying it released E2=qV2, which would seem natural since qV measures the extra energy stored in the electric field because of the existence of a charge at a certain point. Outside of the interval during which the particle exists, conservation of energy would be satisfied, because the net energy released by creation and destruction of the particle, qV2-qV1, would be equal to the work done in moving the particle from 1 to 2, qV2-qV1. But although this scenario salvages conservation of energy, it violates gauge symmetry. Gauge symmetry requires that the laws of physics be invariant under the transformation , which implies that no experiment should be able to measure the absolute potential, without reference to some external standard such as an electrical ground. But the proposed rules E1=qV1 and E2=qV2 for the energies of creation and destruction would allow an experimenter to determine the absolute potential, simply by comparing the energy input required to create the charge q at a particular point in space in the case where the potential is and respectively. The conclusion is that if gauge symmetry holds, and energy is conserved, then charge must be conserved. General relativity As discussed above, the gauge transformations for classical (i.e., non-quantum mechanical) general relativity are arbitrary coordinate transformations. Technically, the transformations must be invertible, and both the transformation and its inverse must be smooth, in the sense of being differentiable an arbitrary number of times. An example of a symmetry in a physical theory: translation invariance Some global symmetries under changes of coordinate predate both general relativity and the concept of a gauge. For example, Galileo and Newton introduced the notion of translation invariance, an advancement from the Aristotelian concept that different places in space, such as the earth versus the heavens, obeyed different physical rules. Suppose, for example, that one observer examines the properties of a hydrogen atom on Earth, the other—on the Moon (or any other place in the universe), the observer will find that their hydrogen atoms exhibit completely identical properties. Again, if one observer had examined a hydrogen atom today and the other—100 years ago (or any other time in the past or in the future), the two experiments would again produce completely identical results. The invariance of the properties of a hydrogen atom with respect to the time and place where these properties were investigated is called translation invariance. Recalling our two observers from different ages: the time in their experiments is shifted by 100 years. If the time when the older observer did the experiment was t, the time of the modern experiment is t+100 years. Both observers discover the same laws of physics. Because light from hydrogen atoms in distant galaxies may reach the earth after having traveled across space for billions of years, in effect one can do such observations covering periods of time almost all the way back to the Big Bang, and they show that the laws of physics have always been the same. In other words, if in the theory we change the time t to t+100 years (or indeed any other time shift) the theoretical predictions do not change. Another example of a symmetry: the invariance of Einstein's field equation under arbitrary coordinate transformations In Einstein's general relativity, coordinates like x, y, z, and t are not only "relative" in the global sense of translations like , rotations, etc., but become completely arbitrary, so that, for example, one can define an entirely new time-like coordinate according to some arbitrary rule such as , where has dimensions of time, and yet Einstein's equations will have the same form. Invariance of the form of an equation under an arbitrary coordinate transformation is customarily referred to as general covariance, and equations with this property are referred to as written in the covariant form. General covariance is a special case of gauge invariance. Maxwell's equations can also be expressed in a generally covariant form, which is as invariant under general coordinate transformation as Einstein's field equation. In quantum mechanics Quantum electrodynamics Until the advent of quantum mechanics, the only well known example of gauge symmetry was in electromagnetism, and the general significance of the concept was not fully understood. For example, it was not clear whether it was the fields E and B or the potentials V and A that were the fundamental quantities; if the former, then the gauge transformations could be considered as nothing more than a mathematical trick. Aharonov–Bohm experiment In quantum mechanics, a particle such as an electron is also described as a wave. For example, if the double-slit experiment is performed with electrons, then a wave-like interference pattern is observed. The electron has the highest probability of being detected at locations where the parts of the wave passing through the two slits are in phase with one another, resulting in constructive interference. The frequency, f, of the electron wave is related to the kinetic energy of an individual electron particle via the quantum-mechanical relation E = hf. If there are no electric or magnetic fields present in this experiment, then the electron's energy is constant, and, for example, there will be a high probability of detecting the electron along the central axis of the experiment, where by symmetry the two parts of the wave are in phase. But now suppose that the electrons in the experiment are subject to electric or magnetic fields. For example, if an electric field were imposed on one side of the axis but not on the other, the results of the experiment would be affected. The part of the electron wave passing through that side oscillates at a different rate, since its energy has had −eV added to it, where −e is the charge of the electron and V the electrical potential. The results of the experiment will be different, because phase relationships between the two parts of the electron wave have changed, and therefore the locations of constructive and destructive interference will be shifted to one side or the other. It is the electric potential that occurs here, not the electric field, and this is a manifestation of the fact that it is the potentials and not the fields that are of fundamental significance in quantum mechanics. Explanation with potentials It is even possible to have cases in which an experiment's results differ when the potentials are changed, even if no charged particle is ever exposed to a different field. One such example is the Aharonov–Bohm effect, shown in the figure. In this example, turning on the solenoid only causes a magnetic field B to exist within the solenoid. But the solenoid has been positioned so that the electron cannot possibly pass through its interior. If one believed that the fields were the fundamental quantities, then one would expect that the results of the experiment would be unchanged. In reality, the results are different, because turning on the solenoid changed the vector potential A in the region that the electrons do pass through. Now that it has been established that it is the potentials V and A that are fundamental, and not the fields E and B, we can see that the gauge transformations, which change V and A, have real physical significance, rather than being merely mathematical artifacts. Gauge invariance: the results of the experiments are independent of the choice of the gauge for the potentials Note that in these experiments, the only quantity that affects the result is the difference in phase between the two parts of the electron wave. Suppose we imagine the two parts of the electron wave as tiny clocks, each with a single hand that sweeps around in a circle, keeping track of its own phase. Although this cartoon ignores some technical details, it retains the physical phenomena that are important here. If both clocks are sped up by the same amount, the phase relationship between them is unchanged, and the results of experiments are the same. Not only that, but it is not even necessary to change the speed of each clock by a fixed amount. We could change the angle of the hand on each clock by a varying amount θ, where θ could depend on both the position in space and on time. This would have no effect on the result of the experiment, since the final observation of the location of the electron occurs at a single place and time, so that the phase shift in each electron's "clock" would be the same, and the two effects would cancel out. This is another example of a gauge transformation: it is local, and it does not change the results of experiments. Summary In summary, gauge symmetry attains its full importance in the context of quantum mechanics. In the application of quantum mechanics to electromagnetism, i.e., quantum electrodynamics, gauge symmetry applies to both electromagnetic waves and electron waves. These two gauge symmetries are in fact intimately related. If a gauge transformation θ is applied to the electron waves, for example, then one must also apply a corresponding transformation to the potentials that describe the electromagnetic waves. Gauge symmetry is required in order to make quantum electrodynamics a renormalizable theory, i.e., one in which the calculated predictions of all physically measurable quantities are finite. Types of gauge symmetries The description of the electrons in the subsection above as little clocks is in effect a statement of the mathematical rules according to which the phases of electrons are to be added and subtracted: they are to be treated as ordinary numbers, except that in the case where the result of the calculation falls outside the range of 0≤θ<360°, we force it to "wrap around" into the allowed range, which covers a circle. Another way of putting this is that a phase angle of, say, 5° is considered to be completely equivalent to an angle of 365°. Experiments have verified this testable statement about the interference patterns formed by electron waves. Except for the "wrap-around" property, the algebraic properties of this mathematical structure are exactly the same as those of the ordinary real numbers. In mathematical terminology, electron phases form an Abelian group under addition, called the circle group or U(1). "Abelian" means that addition commutes, so that θ + φ = φ + θ. Group means that addition associates and has an identity element, namely "0". Also, for every phase there exists an inverse such that the sum of a phase and its inverse is 0. Other examples of abelian groups are the integers under addition, 0, and negation, and the nonzero fractions under product, 1, and reciprocal. As a way of visualizing the choice of a gauge, consider whether it is possible to tell if a cylinder has been twisted. If the cylinder has no bumps, marks, or scratches on it, we cannot tell. We could, however, draw an arbitrary curve along the cylinder, defined by some function θ(x), where x measures distance along the axis of the cylinder. Once this arbitrary choice (the choice of gauge) has been made, it becomes possible to detect it if someone later twists the cylinder. In 1954, Chen Ning Yang and Robert Mills proposed to generalize these ideas to noncommutative groups. A noncommutative gauge group can describe a field that, unlike the electromagnetic field, interacts with itself. For example, general relativity states that gravitational fields have energy, and special relativity concludes that energy is equivalent to mass. Hence a gravitational field induces a further gravitational field. The nuclear forces also have this self-interacting property. Gauge bosons Surprisingly, gauge symmetry can give a deeper explanation for the existence of interactions, such as the electric and nuclear interactions. This arises from a type of gauge symmetry relating to the fact that all particles of a given type are experimentally indistinguishable from one another. Imagine that Alice and Betty are identical twins, labeled at birth by bracelets reading A and B. Because the girls are identical, nobody would be able to tell if they had been switched at birth; the labels A and B are arbitrary, and can be interchanged. Such a permanent interchanging of their identities is like a global gauge symmetry. There is also a corresponding local gauge symmetry, which describes the fact that from one moment to the next, Alice and Betty could swap roles while nobody was looking, and nobody would be able to tell. If we observe that Mom's favorite vase is broken, we can only infer that the blame belongs to one twin or the other, but we cannot tell whether the blame is 100% Alice's and 0% Betty's, or vice versa. If Alice and Betty are in fact quantum-mechanical particles rather than people, then they also have wave properties, including the property of superposition, which allows waves to be added, subtracted, and mixed arbitrarily. It follows that we are not even restricted to complete swaps of identity. For example, if we observe that a certain amount of energy exists in a certain location in space, there is no experiment that can tell us whether that energy is 100% A's and 0% B's, 0% A's and 100% B's, or 20% A's and 80% B's, or some other mixture. The fact that the symmetry is local means that we cannot even count on these proportions to remain fixed as the particles propagate through space. The details of how this is represented mathematically depend on technical issues relating to the spins of the particles, but for our present purposes we consider a spinless particle, for which it turns out that the mixing can be specified by some arbitrary choice of gauge θ(x), where an angle θ = 0° represents 100% A and 0% B, θ = 90° means 0% A and 100% B, and intermediate angles represent mixtures. According to the principles of quantum mechanics, particles do not actually have trajectories through space. Motion can only be described in terms of waves, and the momentum p of an individual particle is related to its wavelength λ by p = h/λ. In terms of empirical measurements, the wavelength can only be determined by observing a change in the wave between one point in space and another nearby point (mathematically, by differentiation). A wave with a shorter wavelength oscillates more rapidly, and therefore changes more rapidly between nearby points. Now suppose that we arbitrarily fix a gauge at one point in space, by saying that the energy at that location is 20% A's and 80% B's. We then measure the two waves at some other, nearby point, in order to determine their wavelengths. But there are two entirely different reasons that the waves could have changed. They could have changed because they were oscillating with a certain wavelength, or they could have changed because the gauge function changed from a 20–80 mixture to, say, 21–79. If we ignore the second possibility, the resulting theory does not work; strange discrepancies in momentum will show up, violating the principle of conservation of momentum. Something in the theory must be changed. Again there are technical issues relating to spin, but in several important cases, including electrically charged particles and particles interacting via nuclear forces, the solution to the problem is to impute physical reality to the gauge function θ(x). We say that if the function θ oscillates, it represents a new type of quantum-mechanical wave, and this new wave has its own momentum p = h/λ, which turns out to patch up the discrepancies that otherwise would have broken conservation of momentum. In the context of electromagnetism, the particles A and B would be charged particles such as electrons, and the quantum mechanical wave represented by θ would be the electromagnetic field. (Here we ignore the technical issues raised by the fact that electrons actually have spin 1/2, not spin zero. This oversimplification is the reason that the gauge field θ comes out to be a scalar, whereas the electromagnetic field is actually represented by a vector consisting of V and A.) The result is that we have an explanation for the presence of electromagnetic interactions: if we try to construct a gauge-symmetric theory of identical, non-interacting particles, the result is not self-consistent, and can only be repaired by adding electric and magnetic fields that cause the particles to interact. Although the function θ(x) describes a wave, the laws of quantum mechanics require that it also have particle properties. In the case of electromagnetism, the particle corresponding to electromagnetic waves is the photon. In general, such particles are called gauge bosons, where the term "boson" refers to a particle with integer spin. In the simplest versions of the theory gauge bosons are massless, but it is also possible to construct versions in which they have mass. This is the case for the gauge bosons that carry the weak interaction: the force responsible for nuclear decay.
Physical sciences
Physics basics: General
Physics
291913
https://en.wikipedia.org/wiki/European%20polecat
European polecat
The European polecat (Mustela putorius), also known as the common polecat, black polecat and forest polecat, is a mustelid species native to Europe, Western Asia and North Africa. It is of a generally dark brown colour, with a pale underbelly and a dark mask across the face. Occasionally, colour mutations including albinos, leucists, isabellinists, xanthochromists, amelanists, and erythrists occur. It has a shorter, more compact body than other Mustela species, a more powerfully built skull and dentition, is less agile, and is well known for having the characteristic ability to secrete a particularly foul-smelling liquid to mark its territory. It is much less territorial than other mustelids, with animals of the same sex frequently sharing home ranges. Like other mustelids, the European polecat is polygamous, with pregnancy occurring after mating, following induced ovulation. It usually gives birth in early summer to litters consisting of five to ten kits, which become independent at the age of two to three months. The European polecat feeds on small rodents, birds, amphibians and reptiles. It occasionally cripples its prey by piercing its brain with its teeth and stores it, still living, in its burrow for future consumption. The European polecat originated in Western Europe during the Middle Pleistocene, with its closest living relatives being the steppe polecat, the black-footed ferret and the European mink. With the two former species, it can produce fertile offspring, though hybrids between it and the latter species tend to be sterile, and are distinguished from their parent species by their larger size and more valuable pelts. The European polecat is thought to be the sole ancestor of the ferret, which was domesticated more than 2,000 years ago for the purpose of hunting vermin. The species has otherwise been historically viewed negatively by humans. In Britain especially, the polecat was persecuted by gamekeepers, and became synonymous with promiscuity in early English literature. During modern times, the polecat is still scantly represented in popular culture when compared to other rare British mammals, and misunderstandings of its behaviour still persist in some rural areas. Since 2008, it has been classified as Least Concern on the IUCN Red List due to its wide range and large numbers. Etymology and naming The word "polecat" first appeared after the Norman Conquest of England, written (in Middle English) as polcat. While the second syllable is largely self-explanatory, the origin of the first is unclear. It is possibly derived from the French , meaning "chicken", likely in reference to the species' fondness for poultry, or it may be a variant of the Old English , meaning "foul". In Middle English, the species was referred to as , meaning "foul marten", in reference to its strong odour. In Old French, the polecat was called , which was derived from the Low German and Scandinavian verb for "to make a disagreeable smell". This was later corrupted in English as fitchew or fitchet, which itself became the word "fitch", which is used for the polecat's pelt. The word fitchet is the root word for the North American fisher, which was named by Dutch colonists in America who noted similarities between the two species. In some countries such as New Zealand, the term "fitch" has taken on a wider use to refer to related creatures such as ferrets, especially when farmed for their fur. A 2002 article in The Mammal Society's Mammal Review contested the European polecat's status as an animal indigenous to Britain on account of a scarce fossil record and linguistic evidence. Unlike most native British mammals, the polecat's Welsh name (, derived from the Middle English ) is not of Celtic origin, much as the Welsh names of invasive species such as the European rabbit and fallow deer (, derived from the Middle English and , derived from the Old French , respectively) are of Middle English or Old French origin. Polecats are not mentioned in Anglo-Saxon or Welsh literature prior to the Norman conquest of England in 1066, with the first recorded mention of the species in the Welsh language occurring in the 14th century's Llyfr Coch Hergest and in English in Chaucer's The Pardoner's Tale (1383). In contrast, attestations of the Welsh word for pine marten (), date back at least to the 10th century Welsh Laws and possibly much earlier in northern England. Local and indigenous names Dialectal English names Latin name As well as the several indigenous names referring to smell (see above), the scientific name Mustela putorius is also derived from this species' foul smell. The Latin translates to "stench" or "stink" and is the origin of the English word putrid. Evolution The earliest true polecat was Mustela stromeri, which appeared during the late Villafranchian. It was considerably smaller than the present form, thus indicating polecats evolved at a relatively late period. The oldest modern polecat fossils occur in Germany, Britain and France, and date back to the Middle Pleistocene. The European polecat's closest relatives are the steppe polecat and black-footed ferret, with which it is thought to have shared Mustela stromeri as a common ancestor. The European polecat is, however, not as maximally adapted in the direction of carnivory as the steppe polecat, being less specialised in skull structure and dentition. The European polecat likely diverged from the steppe polecat 1.5 million years ago based on IRBP, though cytochrome b transversions indicate a younger date of 430,000 years. It is also closely related to the European mink, with which it can hybridise. Domestication Morphological, cytological and molecular studies confirm the European polecat is the sole ancestor of the ferret, thus disproving any connection with the steppe polecat, which was once thought to have contributed to the ferret's creation. Ferrets were first mentioned by Aristophanes in 425 BC and by Aristotle in 350 BC. Greek and Roman writers in the first century AD were the first to attest on the ferret's use in bolting rabbits from their burrows. The first accurate descriptions of ferrets come from Strabo during 200 AD, when ferrets were released onto the Balearic Islands to control rabbit populations. As the European rabbit is native to the Iberian Peninsula and northwest Africa, the European polecat likely was first domesticated in these regions. The ferret and European polecat are similar in both size and portions, to the point that dark-coloured ferrets are almost indistinguishable from their wild cousins, though the ferret's skull has a smaller cranial volume, and has a narrower postorbital constriction. Compared to the European polecat, the ferret has a much smaller brain, though this comparison has not been made with Mediterranean polecats, from which ferrets likely derive. The theory of a Mediterranean origin is further strengthened because the ferret is less tolerant of cold than northern polecat subspecies. Unlike other subspecies, which are largely solitary, the ferret will readily live in social groups. The ferret is also slower in all its movements than the polecat, and hardly ever makes any use of its anal scent glands. Overall, the ferret represents a neotenous form of polecat. Subspecies , seven subspecies are recognised. Description Build The appearance of the European polecat is typical of members of the genus Mustela, though it is generally more compact in conformation and, although short-legged, has a less elongated body than the European mink or steppe polecat. The tail is short, about one-third its body length. The eyes are small, with dark brown irises. The hind toes are long and partially webbed, with weakly curved 4 mm-long, nonretractable claws. The front claws are strongly curved, partially retractable, and measure 6 mm in length. The feet are moderately long and more robust than in other members of the genus. The polecat's skull is relatively coarse and massive, more so than the mink's, with a strong, but short and broad facial region and strongly developed projections. In comparison to other similarly sized mustelids, the polecat's teeth are very strong, large and massive in relation to skull size. Sexual dimorphism in the skull is apparent in the lighter, narrower skull of the female, which also has weaker projections. The polecat's running gait is not as complex and twisting as that of the mink or stoat, and it is not as fast as the mountain weasel (solongoi), stoat or least weasel, as it can be outrun by a conditioned man. Its sensory organs are well developed, though it is unable to distinguish between colours. The dimensions of the European polecat vary greatly. The species does not conform to Bergmann's rule, with the pattern of size variation seeming to follow a trend of size increase along an east–west axis. Males measure in body length and females are . The tail measures 115–167 mm in males and 84–150 mm in females. Adult males in middle Europe weigh and females . Gigantism is known among polecats, but specimens exhibiting this are likely the products of polecat-mink hybridisation. Fur The winter fur of the European polecat is brownish black or blackish brown, the intensity of which is determined by the colour of the long guard hairs. On the back and flanks, the dark tone is brightened by bright whitish-yellowish, sometimes yellowish-greyish underfur which shows through. The lightly coloured underfur is not equally visible on different parts of the body. On the back and hindquarters, the underfur is almost completely covered by the dark guard hairs. On the flanks, though, the lightening is well defined, and contrasts sharply with the general tone of the back. The throat, lower neck, chest and abdomen are black or blackish brown. The limbs are pure black or black with brown tints, while the tail is black or blackish brown, completely lacking light underfur. The area around and between the eyes is black-brown, with a longitudinal stripe of similar colour along the top of the nose. The ears are dark brown and edged with white. The summer fur is short, sparse and coarse. It is greyer, duller and lacking in the lustre of the winter fur. The underfur is more weakly developed in the summer fur, and has a brownish-grey or rusty-grey colour. The polecat is a good swimmer, but its fur is not as well insulated against cold water as the American mink's; while a mink will take 118 minutes to cool in a water temperature of , the polecat cools down much faster at 26–28 minutes. Polecats were found in two major phenotypes a typic one and a dark fur one with no black mask. Colour mutations include albinos, leucists, isabellinists, xanthochromists, amelanists, and erythrists. In typical erythristic individuals, the underfur is usually bright reddish. The guard hairs on the trunk are bright reddish or reddish brown. Black guard hairs are absent on the lower body and head. In some rare cases, the guard hairs are so light, they are almost indistinguishable from the pale-yellow underfur. These individuals are called "amelanistic". In these cases, the whole animal is a very light golden-yellow colour. These individuals are called "isabelline" or "xanthochromistic". Behaviour Social and territorial behaviour The European polecat has definite home ranges that vary according to season, habitat, sex and social status. Breeding females settle in discrete areas, whereas breeding males and dispersing juveniles have more fluid home ranges, and are more mobile. Both males and females share home ranges with other members of the same sex; males have larger home ranges than females, but evidence of territorial marking is sparse. European polecats use several den sites distributed throughout their home ranges and are often most active around rabbit warrens. Some European polecats use farm buildings or haystacks as daytime resting sites in winter. Occasionally, European polecats use abandoned European badger or red fox burrows. Like other mustelids, the polecat is usually a silent animal, though it will growl fiercely when angered, and squeak when distressed. It also emits a low, mewling cry to its mate or offspring. Reproduction and development The European polecat is a seasonal breeder, with no courtship rituals. During the mating season, the male grabs the female by the neck and drags her about to stimulate ovulation, then copulates for up to an hour. The species is polygamous, with each male polecat mating with several females. The gestation period lasts 40–43 days, with litters usually being born in May-early June. Each litter typically consists of five to ten kits. At birth, the kits weigh and measure in body length; they are blind and deaf. At the age of one week, the kits are covered in silky, white fur, which is replaced with a cinnamon brown-greyish woolly coat at the age of 3–4 wk. Weaning begins at three weeks of age, while the permanent dentition erupts after 7–8 wk. The kits become independent after two to three months. Females are very protective of their young, and have even been known to confront humans approaching too closely to their litters. Ecology Diet The European polecat's diet consists of mouse-like rodents, followed by amphibians and birds. Its most frequent prey item in the former Soviet Union is the common vole and rarely the red-backed vole. In large river floodlands, water vole are common prey. In spring and winter, amphibians (especially grass frogs and green toads) become important food items. Selective predation on male frogs by the polecat decreases the occurrence of polyandry in frog populations. However, because amphibians have little calorific value, the polecat never grows fat on them, no matter how many it consumes. In Central Europe, the diet in winter months is dominated by birds including quail, grey partridges, grouse, chickens, pigeons and passerines. Seasonal changes in the activity rhythm is synchronised with the activity of the main prey. Some species only rarely preyed upon by the polecat include European hedgehogs, asp vipers, grass snakes, lizards and insects. In Britain, it commonly kills brown rats and European rabbits, and is capable of killing larger prey, such as geese and hares. One polecat was reported to frequently wait at a riverbank and catch eels, which it took back to its burrow. The European polecat hunts its prey by stalking it and seizing it with its canine teeth, killing the animal with a bite to the neck. This killing method is instinctive, but perfected with practice. The polecat sometimes caches its food, particularly during seasonal gluts of frogs and toads. Sometimes, the polecat does not kill these, but bites them at the base of the skull, thus paralyzing them and keeping them fresh for later consumption. Although they are normally shy around humans, naturalist Alfred Brehm in his Brehms Tierleben mentions an exceptional case in which three polecats attacked a baby in Hesse. During winter, some European polecats raid beehives and feed on honey. Enemies and competitors Although the European polecat can coexist with the European mink, it suffers in areas where the invasive American mink also occurs, as the latter species feeds on the same mammals as the polecat much more frequently than the European mink, and has been known to drive the polecat out of wetland habitats. In areas where the European polecat is sympatric with the steppe polecat, the two species overlap greatly in choice of food, though the former tends to consume more household foods and birds, while the latter preys on mammals more frequently. There is at least one record of a beech marten killing a polecat. The European polecat may prey on the much smaller least weasel. Hybridisation In some parts of Britain, the abandonment of domestic ferrets has led to ferret-polecat crossbreeds living in the wild. Ferrets were likely first brought to Britain after the Norman Conquest of England, or as late as the fourteenth century. Crossbreeds between the two animals typically have a distinct white throat patch, white feet and white hairs interspersed among the fur. Typically, first generation crossbreeds between polecats and ferrets develop their wild parents' fear of humans if left with their mothers during the critical socialisation period between 7½ and 8½ weeks of age. The European polecat can hybridise with the European mink, producing offspring termed khor'-tumak by Russian furriers and khonorik by fanciers. Such hybridisation is very rare in the wild, and typically only occurs where the European mink population is in decline. A polecat-mink hybrid has a poorly defined facial mask, yellow fur on the ears, grey-yellow underfur and long, dark brown guard hairs. It is fairly large, with a male attaining the peak sizes known for European polecats weighing and measuring in length; a female is much larger than a female European minks weighing and measuring in length. The majority of polecat-mink hybrids have skulls bearing greater similarities to those of polecats than to minks; they can swim well and burrow for food like polecats, but are difficult to tame and breed, as males are sterile. The first captive polecat-mink hybrid was created in 1978 for fur, but breeding of these hybrids declined as European mink populations decreased. Studies on the behavioural ecology of free-ranging polecat-mink hybrids in the upper reaches of the Lovat River indicate that hybrids stray from aquatic habitats more readily than minks, and tolerate both parent species entering their territories, though the hybrid's larger size, especially the male's deters intrusion. During summer, the diets of wild polecat-mink hybrids are more similar to those of minks than to the polecats, as they feed predominantly on frogs. During winter, their diets overlap more with those of polecats, and will eat a larger proportion of rodents than in the summer, though they still rely heavily on frogs and rarely scavenge ungulate carcasses as polecats do. The European polecat can also hybridise with the Asian steppe polecat or the domestic ferret to produce fertile offspring. European-steppe polecat hybrids are very rare, despite their sympatry in several areas. Nevertheless, hybrids have been recorded in southern Ukraine, the Kursk and Voronezh Oblasts, the Trans-Carpathians and several other localities. Range, history and conservation The European polecat is widespread in the western Palaearctic to the Urals in the Russian Federation, though it is absent from Ireland, northern Scandinavia, and much of the Balkans and eastern Adriatic coast. It occurs only marginally in northern Greece. It is found in Morocco in the Rif Mountains, from sea level to . Its domesticated form, the ferret, was introduced in Britain, and some Mediterranean islands and New Zealand. Britain In Britain, the European polecat was regarded as a serious poultry predator prior to the introduction of wire netting, therefore eliminating it was considered the only option to protect stock. This extreme enmity does not appear to have been universal however. Speaking of Merioneth (Gwynedd) Peter Hope Jones reported that "for a county supposedly well-placed within the known past distribution of this species, Merioneth has relatively very few records of Polecats amongst its parish bounty payments. Perhaps this animal was not generally considered to be an important pest, but whatever the real reason, in only two parishes are direct references made to this species by the name by which we know it today. In the years from 1729 to 1732 about twenty were killed in Towyn parish, where 2/6 was paid for a full-grown polecat and half this sum for a young 'kittin'. Records for Llanfor.... show that only 42 were killed in the 39-year period from 1720 to 1758, the payment being exactly half the going rate for a fox, i.e. 2/6 for a full-grown polecat, and 1/3 for a young animal". In Kent, for example, at least 42 parishes paid bounties for polecats, of which three extended into the 19th century, though by this time only single individuals were recorded, and usually after gaps of many years. In the Kingdom of Scotland, during the reign of David II, an export duty of 4d. was imposed on each polecat fur trimmer, which was raised to 8d. in 1424. The species held an important place in Scotland's fur markets; the annual Dumfries Fur Fair (1816–1874) sold 400 polecat pelts in 1829 and 600 in 1831. The following year, a contemporary account described polecat skins as "a drug on the market". In 1856, the number of sold pelts decreased to 240, 168 in 1860, 12 in 1866 and none in 1869. The decline was halted with the decrease in the intensity of gamekeeping during the 20 year interval between the First and Second World Wars. The European polecat is afforded both national and European protection; it is listed on Schedule 6 of the Wildlife and Countryside Act 1981 and Regulation 41 of the Conservation (Natural Habitats, &c.) Regulations 1994 and is listed on Annex V of the Habitats Directive. A survey carried out by the Vincent Wildlife Trust in 2015 found that the polecat had spread into areas (such as East Anglia and South Yorkshire) where they had not been seen for 100 years. Naturalist Chris Packham termed the spread "...one of the great natural recoveries." The New Forest in Hampshire also now has a small polecat population, a fact discovered after scientists set up cameras to film pine martens. France The European polecat is present in all of France's territories, excepting Corsica, and has been in a state of decline for several decades. Nevertheless, it is listed as Least Concern on France's Red Data Book. The European polecat is rare in numerous regions or départements. In the Rhone-Alps region, its population has undergone a sizeable decline since the 1990s, largely as a consequence of poisoning campaigns against muskrats. A 1999 study on the decline of polecats in this region indicated the species has little chance of surviving there. Elsewhere, it is considered either rare or sporadic in 22 districts and absent or extirpated in 22 others. In Drôme, for example, polecat populations have been decreasing since 1975, and have disappeared in 27 communes in Isère. Its numbers are declining in Morvan and Ariège, and is thinly distributed in Brittany. Although present in Aquitaine, its numbers have been dropping since the 1950s, and is very rare in the mountain regions. In Normandy, the speed of the polecat's decline has somewhat decreased. In the alpine départements, its range is limited by altitude, as the species relies on more Mediterranean climates to thrive. It is, however, especially abundant in the irrigated Crau, but is absent on the eastern part of the area, apparently being restricted by the valleys of the Durance and Rhone Rivers. The largest populations occur in Northern France: Pas de Calais, Central France; Alsace, Lorraine and the areas of the Loire with the Vendée, which holds the largest record of polecat observations. It is common in all the départements of Champagne-Ardenne. Former Soviet Union The western border of the European polecat's range in the former Soviet Union begins from the mouth of the Danube in the south approximately to northwest of Suoyarvi, on the Finnish border in the north. In Karelia, its northern border extends from the former point towards the southeast to the Spassk Bay of Lake Onega, thereby passing around the West Karelian uplands from the south and then, passing around these uplands from the east, it suddenly ascends directly to the north passing in particular, near the western shore of Segozer and reaches Rugozer. From there, the border line turns northeast, crossing the Lakhta and reaching Kem on the White Sea. From Archangelsk, the border reaches Mezen, thus attaining the species' most northerly range. From the Mezen River's mouth, the border abruptly returns south, approaching closer to the upper Mezen near 64° lat. From there, the polecat's northern border goes on to the upper Vychegda River, and descends further on southwards and in the Urals. Its eastern range apparently extends along the Urals, embracing Sverdlovsk from the west. It is probably absent in the southern Urals, where the steppe polecat occurs. The southern border of the polecat's range starts in the west of the Danube's mouth and extends eastward along the coast of the Black Sea reaching the mouth of the Dnepr, from which it moves back from the shore of the Azov Sea and, along it, goes to the mouth of the Don. From the mouth and lower course of the Don, its range passes into the steppe region of western and middle Ciscaucasia. The European polecat is absent from the Saratov steppes of Transvolga, instead being encountered only in the extreme lower Bolshoy and Maly Irgiz Rivers. Further on, the border goes to the north along the Volga River. It steeply returns east somewhat south at the Samara bend, passing around Obshchy Syrt, reaching the Urals at the latitude of Magnitogorsk. The range of the polecat within the former Soviet Union has expanded northwards. From 1930 to 1952 for example, the polecat colonised northwestern Karelia and southern Finland. Prior to the First World War, the Russian Empire produced more than 50% of global polecat skins. The harvesting of polecats in Russia increased substantially after the October Revolution, which coincided with Western Europe's decline in polecat numbers. The Russian population of polecats decreased somewhat after the Second World War, and their hunting was subsequently discouraged, as polecats were acknowledged to limit harmful rodent populations. Diseases and parasites The European polecat may suffer from distemper, influenza, the common cold and pneumonia. Occasionally, it is affected by malignant tumours and hydrocephaly. It commonly has broken teeth and, on rarer occasions, fatal abscesses on the jaw, head and neck. In mainland Europe, it is a carrier of trichinosis, leptospirosis, toxoplasmosis and adiaspiromycosis. Incidences of polecats carrying rabies are high in some localized areas. Ectoparasites known to infest polecats include flea species such as Ctenocephalides felis, Archaeospylla erinacei, Nosopsyllus fasciatus and Paraceras melis. The tick Ixodes hexagonus is the polecat's most common ectoparasite, which is sometimes found in large numbers on the neck and behind the ears. Another, less common species to infest polecats is I. canisuga. The biting louse Trichodectes jacobi is also known to infest polecats. Endoparasites carried by polecats include the cestodes Taenia tenuicollis and T. martis and the nematodes Molineus patens, Strongyloides papillosus, Capilliaria putorii, Filaroides martis and Skjrabingylus nasicola. Relationships with humans Hunting and fur use European polecat hunting was once a favourite sport of the Westmorland dalesmen and the Scots, who hunted them at night in midwinter. However, the majority of polecat deaths caused by humans have been accidental, having mostly been caused by steel traps set for rabbits. Hunting polecats by moonlight was also a popular diversion among midland schoolboys. Until the mid-19th century, polecats in Britain were hunted from early February to late April with mixed packs of hunting dogs on the Welsh hills and Lakeland fells, though otterhounds were used on the fells, the Border country and the Scottish Lowlands. John Tucker Edwardes, the creator of the Sealyham terrier, used captured wild male polecats to test the gameness of yearling terriers. In the former Soviet Union, polecats are hunted chiefly in late autumn and early winter with guns and hunting dogs, as well as foothold traps and wooden snares. However, even in season, hunters rarely catch more than 10–15 polecats. The species does not constitute an important element in former Soviet commercial hunting, and is usually only caught incidentally. The European polecat is a valuable fur bearer, whose pelt (fitch) is more valuable than the steppe polecat's. Its skin is used primarily in the production of jackets, capes and coats. It is particularly well suited for trimmings for women's clothing. The tail is sometimes used for the making of paintbrushes. One disadvantage of polecat skin, however, is its unpleasant odour, which is difficult to remove. The European polecat was first commercially farmed for its fur in Great Britain during the 1920s, but was only elevated to economic importance in Finland in 1979. It never became popular in the United States and Canada, due to import laws regarding non-native species. It did gain economic importance in the USSR, though. Tameability Unlike the stoat and least weasel, the European polecat is easy to breed in captivity. According to Aubyn Trevor-Battye, the European polecat is difficult to tame, but is superior to its domesticated form, the ferret, in bolting rats from their holes due to its greater agility. It is prone to attempting escape once finished bolting rats, but can be easily outrun. Polecat kits can be successfully raised and suckled by mother cats. According to Owen's Welsh Dictionary, the Gwythelians (early Irish settlers in northern Wales) kept polecats as pets. Attempts to tame the European polecat are generally hampered by the adult's nervous and unsociable disposition. First generation hybrids between polecats and ferrets, conceived to improve the latter's bloodlines, produce animals with personalities similar to their wild parent's. In culture In Britain, the polecat historically has had a negative reputation.
Biology and health sciences
Mustelidae
Animals
291928
https://en.wikipedia.org/wiki/Operator%20%28physics%29
Operator (physics)
An operator is a function over a space of physical states onto another space of states. The simplest example of the utility of operators is the study of symmetry (which makes the concept of a group useful in this context). Because of this, they are useful tools in classical mechanics. Operators are even more important in quantum mechanics, where they form an intrinsic part of the formulation of the theory.They play a central role in describing observables (measurable quantities like energy, momentum, etc.). Operators in classical mechanics In classical mechanics, the movement of a particle (or system of particles) is completely determined by the Lagrangian or equivalently the Hamiltonian , a function of the generalized coordinates q, generalized velocities and its conjugate momenta: If either L or H is independent of a generalized coordinate q, meaning the L and H do not change when q is changed, which in turn means the dynamics of the particle are still the same even when q changes, the corresponding momenta conjugate to those coordinates will be conserved (this is part of Noether's theorem, and the invariance of motion with respect to the coordinate q is a symmetry). Operators in classical mechanics are related to these symmetries. More technically, when H is invariant under the action of a certain group of transformations G: . The elements of G are physical operators, which map physical states among themselves. Table of classical mechanics operators where is the rotation matrix about an axis defined by the unit vector and angle θ. Generators If the transformation is infinitesimal, the operator action should be of the form where is the identity operator, is a parameter with a small value, and will depend on the transformation at hand, and is called a generator of the group. Again, as a simple example, we will derive the generator of the space translations on 1D functions. As it was stated, . If is infinitesimal, then we may write This formula may be rewritten as where is the generator of the translation group, which in this case happens to be the derivative operator. Thus, it is said that the generator of translations is the derivative. The exponential map The whole group may be recovered, under normal circumstances, from the generators, via the exponential map. In the case of the translations the idea works like this. The translation for a finite value of may be obtained by repeated application of the infinitesimal translation: with the standing for the application times. If is large, each of the factors may be considered to be infinitesimal: But this limit may be rewritten as an exponential: To be convinced of the validity of this formal expression, we may expand the exponential in a power series: The right-hand side may be rewritten as which is just the Taylor expansion of , which was our original value for . The mathematical properties of physical operators are a topic of great importance in itself. For further information, see C*-algebra and Gelfand–Naimark theorem. Operators in quantum mechanics The mathematical formulation of quantum mechanics (QM) is built upon the concept of an operator. Physical pure states in quantum mechanics are represented as unit-norm vectors (probabilities are normalized to one) in a special complex Hilbert space. Time evolution in this vector space is given by the application of the evolution operator. Any observable, i.e., any quantity which can be measured in a physical experiment, should be associated with a self-adjoint linear operator. The operators must yield real eigenvalues, since they are values which may come up as the result of the experiment. Mathematically this means the operators must be Hermitian. The probability of each eigenvalue is related to the projection of the physical state on the subspace related to that eigenvalue. See below for mathematical details about Hermitian operators. In the wave mechanics formulation of QM, the wavefunction varies with space and time, or equivalently momentum and time (see position and momentum space for details), so observables are differential operators. In the matrix mechanics formulation, the norm of the physical state should stay fixed, so the evolution operator should be unitary, and the operators can be represented as matrices. Any other symmetry, mapping a physical state into another, should keep this restriction. Wavefunction The wavefunction must be square-integrable (see Lp spaces), meaning: and normalizable, so that: Two cases of eigenstates (and eigenvalues) are: for discrete eigenstates forming a discrete basis, so any state is a sum where ci are complex numbers such that ci2 = ci*ci is the probability of measuring the state , and the corresponding set of eigenvalues ai is also discrete - either finite or countably infinite. In this case, the inner product of two eigenstates is given by , where denotes the Kronecker Delta. However, for a continuum of eigenstates forming a continuous basis, any state is an integral where c(φ) is a complex function such that c(φ)2 = c(φ)*c(φ) is the probability of measuring the state , and there is an uncountably infinite set of eigenvalues a. In this case, the inner product of two eigenstates is defined as , where here denotes the Dirac Delta. Linear operators in wave mechanics Let be the wavefunction for a quantum system, and be any linear operator for some observable (such as position, momentum, energy, angular momentum etc.). If is an eigenfunction of the operator , then where is the eigenvalue of the operator, corresponding to the measured value of the observable, i.e. observable has a measured value . If is an eigenfunction of a given operator , then a definite quantity (the eigenvalue ) will be observed if a measurement of the observable is made on the state . Conversely, if is not an eigenfunction of , then it has no eigenvalue for , and the observable does not have a single definite value in that case. Instead, measurements of the observable will yield each eigenvalue with a certain probability (related to the decomposition of relative to the orthonormal eigenbasis of ). In bra–ket notation the above can be written; that are equal if is an eigenvector, or eigenket of the observable . Due to linearity, vectors can be defined in any number of dimensions, as each component of the vector acts on the function separately. One mathematical example is the del operator, which is itself a vector (useful in momentum-related quantum operators, in the table below). An operator in n-dimensional space can be written: where ej are basis vectors corresponding to each component operator Aj. Each component will yield a corresponding eigenvalue . Acting this on the wave function : in which we have used In bra–ket notation: Commutation of operators on Ψ If two observables A and B have linear operators and , the commutator is defined by, The commutator is itself a (composite) operator. Acting the commutator on ψ gives: If ψ is an eigenfunction with eigenvalues a and b for observables A and B respectively, and if the operators commute: then the observables A and B can be measured simultaneously with infinite precision, i.e., uncertainties , simultaneously. ψ is then said to be the simultaneous eigenfunction of A and B. To illustrate this: It shows that measurement of A and B does not cause any shift of state, i.e., initial and final states are same (no disturbance due to measurement). Suppose we measure A to get value a. We then measure B to get the value b. We measure A again. We still get the same value a. Clearly the state (ψ) of the system is not destroyed and so we are able to measure A and B simultaneously with infinite precision. If the operators do not commute: they cannot be prepared simultaneously to arbitrary precision, and there is an uncertainty relation between the observables even if ψ is an eigenfunction the above relation holds. Notable pairs are position-and-momentum and energy-and-time uncertainty relations, and the angular momenta (spin, orbital and total) about any two orthogonal axes (such as Lx and Ly, or sy and sz, etc.). Expectation values of operators on Ψ The expectation value (equivalently the average or mean value) is the average measurement of an observable, for particle in region R. The expectation value of the operator is calculated from: This can be generalized to any function F of an operator: An example of F is the 2-fold action of A on ψ, i.e. squaring an operator or doing it twice: Hermitian operators The definition of a Hermitian operator is: Following from this, in bra–ket notation: Important properties of Hermitian operators include: real eigenvalues, eigenvectors with different eigenvalues are orthogonal, eigenvectors can be chosen to be a complete orthonormal basis, Operators in matrix mechanics An operator can be written in matrix form to map one basis vector to another. Since the operators are linear, the matrix is a linear transformation (aka transition matrix) between bases. Each basis element can be connected to another, by the expression: which is a matrix element: A further property of a Hermitian operator is that eigenfunctions corresponding to different eigenvalues are orthogonal. In matrix form, operators allow real eigenvalues to be found, corresponding to measurements. Orthogonality allows a suitable basis set of vectors to represent the state of the quantum system. The eigenvalues of the operator are also evaluated in the same way as for the square matrix, by solving the characteristic polynomial: where I is the n × n identity matrix, as an operator it corresponds to the identity operator. For a discrete basis: while for a continuous basis: Inverse of an operator A non-singular operator has an inverse defined by: If an operator has no inverse, it is a singular operator. In a finite-dimensional space, an operator is non-singular if and only if its determinant is nonzero: and hence the determinant is zero for a singular operator. Table of Quantum Mechanics operators The operators used in quantum mechanics are collected in the table below (see for example). The bold-face vectors with circumflexes are not unit vectors, they are 3-vector operators; all three spatial components taken together. {| class="wikitable" |- style="vertical-align:top;" ! scope="col" | Operator (common name/s) ! scope="col" | Cartesian component ! scope="col" | General definition ! scope="col" | SI unit ! scope="col" | Dimension |- style="vertical-align:top;" ! Position | | | m | [L] |- style="vertical-align:top;" !rowspan="2"| Momentum | General | General | J s m−1 = N s | [M] [L] [T]−1 |- style="vertical-align:top;" | Electromagnetic field | Electromagnetic field (uses kinetic momentum; A, vector potential) | J s m−1 = N s | [M] [L] [T]−1 |- style="vertical-align:top;" !rowspan="3"| Kinetic energy | Translation | | J | [M] [L]2 [T]−2 |- style="vertical-align:top;" | Electromagnetic field | Electromagnetic field (A, vector potential) | J | [M] [L]2 [T]−2 |- style="vertical-align:top;" | Rotation (I, moment of inertia) | Rotation | J | [M] [L]2 [T]−2 |- style="vertical-align:top;" ! Potential energy | N/A | | J | [M] [L]2 [T]−2 |- style="vertical-align:top;" ! Total energy | N/A | Time-dependent potential: Time-independent: | J | [M] [L]2 [T]−2 |- style="vertical-align:top;" ! Hamiltonian | | | J | [M] [L]2 [T]−2 |- style="vertical-align:top;" ! Angular momentum operator | | | J s = N s m | [M] [L]2 [T]−1 |- style="vertical-align:top;" ! Spin angular momentum | where are the Pauli matrices for spin-1/2 particles. | where σ is the vector whose components are the Pauli matrices. | J s = N s m | [M] [L]2 [T]−1 |- style="vertical-align:top;" ! Total angular momentum | | | J s = N s m | [M] [L]2 [T]−1 |- style="vertical-align:top;" ! Transition dipole moment (electric) | | | C m | [I] [T] [L] |} Examples of applying quantum operators The procedure for extracting information from a wave function is as follows. Consider the momentum p of a particle as an example. The momentum operator in position basis in one dimension is: Letting this act on ψ we obtain: if ψ is an eigenfunction of , then the momentum eigenvalue p is the value of the particle's momentum, found by: For three dimensions the momentum operator uses the nabla operator to become: In Cartesian coordinates (using the standard Cartesian basis vectors ex, ey, ez) this can be written; that is: The process of finding eigenvalues is the same. Since this is a vector and operator equation, if ψ is an eigenfunction, then each component of the momentum operator will have an eigenvalue corresponding to that component of momentum. Acting on ψ obtains:
Physical sciences
Quantum mechanics
Physics
292032
https://en.wikipedia.org/wiki/Food%20storage
Food storage
Food storage is a way of decreasing the variability of the food supply in the face of natural, inevitable variability. It allows food to be eaten for some time (typically weeks to months) after harvest rather than solely immediately. It is both a traditional domestic skill (mainly as root cellaring) and, in the form of food logistics, an important industrial and commercial activity. Food preservation, storage, and transport, including timely delivery to consumers, are important to food security, especially for the majority of people throughout the world who rely on others to produce their food. Significant losses of food are caused by inadequate storage conditions as well as decisions made at earlier stages of the supply chain, which predispose products to a shorter shelf life. Adequate cold storage, in particular, can be crucial to prevent quantitative and qualitative food losses. Food is stored by almost every human society and by many animals. Storing of food has several main purposes: Preventing foodborne illness from consuming decomposing food Reducing food waste by preserving unused or uneaten food for later use Storage of harvested and processed plant and animal food products for distribution to consumers Enabling a better balanced diet throughout the year Preserving pantry food, such as spices or dry ingredients like rice and flour, for eventual use in cooking Preparedness for catastrophes, emergencies and periods of food scarcity or famine, whether as basic emergency preparedness (for most people) or in its more extreme form of survivalism (prepping) Religious reasons: for example, leaders in the LDS Church (Church of Jesus Christ of Latter Day Saints) instruct church members to store food. Protection from animals or theft Domestic food storage The safe storage of food for home use should strictly adhere to guidelines set out by reliable sources, such as the United States Department of Agriculture. These guidelines have been thoroughly researched by scientists to determine the best methods for reducing the real threat of food poisoning from unsafe food storage. It is also important to maintain proper kitchen hygiene, to reduce risks of bacteria or virus growth and food poisoning. The common food poisoning illnesses include Listeriosis, Mycotoxicosis, Salmonellosis, E. coli, Staphylococcal food poisoning and Botulism. There are many other organisms that can also cause food poisoning. There are also safety guidelines available for the correct methods of home canning of food. For example, there are specific boiling times that apply depending upon whether pressure canning or waterbath canning is being used in the process. These safety guidelines are intended to reduce the growth of mold and bacteria and the threat of potentially-fatal food poisoning. Food storage safety Freezing food To preserve food over long periods the temperature should be maintained below . Careful thawing and cooking immediately after thawing are necessary to maintain the safety of food Food frozen at 0 °F and below may be preserved almost indefinitely although the quality of the food is likely to deteriorate over time. The United States Department of Agriculture, Food Safety and Inspection Service publishes a chart showing the suggested freezer storage time for common foods. Refrigeration Food storage in refrigerators may not be safe unless there is close adherence to temperature guidelines. In general the temperature should be maintained at or below but never below . Safe storage times vary from food to food and may depend on how the food has been treated prior to being placed in the refrigerator. Storing oils and fats Oils and fats can begin to go rancid quickly when not stored safely. Rancid cooking oils and fats do not often smell rancid until well after they have spoiled. Oxygen, light and heat all contribute to cooking oils becoming rancid. The higher the level of polyunsaturated fat that an oil contains, the faster it spoils. The percentage of polyunsaturated fat in some common cooking oils is: safflower (74%); sunflower (66%); corn (60%); soybean (37%); peanut (32%); canola (29%); olive (8%); coconut (5%). To help preserve oils from rancidification, they should be stored in a dark place, stored in oxygen-safe, light-reducing containers (e.g. dark glass or metal). Once opened, oils should be refrigerated and used within a few weeks, when some types begin to go rancid. Unopened oils can have a storage life of up to one year, but some types have a shorter shelf-life even when unopened (such as sesame and flaxseed). Dry storage of foods Vegetables The guidelines vary for safe storage of vegetables under dry conditions. This is because different vegetables have different characteristics, for example, tomatoes contain a lot of water, while root vegetables such as carrots and potatoes contain less. These factors, and many others, affect the amount of time that a vegetable can be kept in dry storage, as well as the temperature needed to preserve its usefulness. The following guideline shows the required dry storage conditions: Cool and dry: onion Cool and moist: root vegetable, potato, cabbage Warm and dry: winter squash, pumpkin, sweet potatoes, dried hot peppers Grain Grain, which includes dry kitchen ingredients such as flour, rice, millet, couscous, cornmeal, and so on, can be stored in rigid sealed containers to prevent moisture contamination or insect or rodent infestation. For kitchen use, glass containers are the most traditional method. During the 20th century plastic containers were introduced for kitchen use. They are now sold in a vast variety of sizes and designs. Metal cans are used (in the United States the smallest practical grain storage uses closed-top #10 metal cans measuring about 3 to 3.5 liters). Storage in grain sacks is ineffective; mold and pests destroy a 25 kg cloth sack of grain in a year, even if stored off the ground in a dry area. On the ground or damp concrete, grain can spoil in as little as three days, and the grain might have to be dried before it can be milled. Food stored under unsuitable conditions should not be purchased or used because of risk of spoilage. To test whether grain is still good, it can be sprouted. If it sprouts, it is still good, but if not, it should not be eaten. It may take up to a week for grains to sprout. When in doubt about the safety of the food, throw it out as quickly as possible. Spices and herbs Spices and herbs are today often sold prepackaged in a way that is convenient for pantry storage. The packaging has dual purposes of both storing and dispensing the spices or herbs. They are sold in small glass or plastic containers or resealable plastic packaging. When spices or herbs are homegrown or bought in bulk, they can be stored at home in glass or plastic containers. They can be stored for extended periods, in some cases for years. However, after 6 months to a year, spices and herbs will gradually lose their flavour as oils they contain will slowly evaporate during storage. Spices and herbs can be preserved in vinegar for short periods of up to a month without losing flavor, creating a flavoured vinegar. Alternative methods for preserving herbs include freezing in water or unsalted butter. Herbs can be chopped and added to water in an ice cube tray. After freezing, the ice cubes are emptied into a plastic freezer bag for storing in the freezer. Herbs also can be stirred into a bowl with unsalted butter, then spread on wax paper and rolled into a cylinder shape. The wax paper roll containing the butter and herbs is then stored in a freezer, and can be cut off in the desired amount for cooking. Using either of these techniques, the herbs should be used within a year. Meat Unpreserved meat has only a relatively short life in storage. Perishable meats should be refrigerated, frozen, dried promptly or cured. Storage of fresh meats is a complex discipline that affects the costs, storage life and eating quality of the meat, and the appropriate techniques vary with the kind of meat and the particular requirements. For example, dry ageing techniques are sometimes used to tenderize gourmet meats by hanging them in carefully controlled environments for up to 21 days, while game animals of various kinds may be hung after shooting. Details depend on personal tastes and local traditions. Modern techniques of preparing meat for storage vary with the type of meat and special requirements of tenderness, flavour, hygiene, and economy. Semi-dried meats like salamis and country style hams are processed first with salt, smoke, sugar, acid, or other "cures" then hung in cool dry storage for extended periods, sometimes exceeding a year. Some of the materials added during the curing of meats serve to reduce the risks of food poisoning from anaerobic bacteria such as species of Clostridium that release botulinum toxin that can cause botulism. Typical ingredients of curing agents that inhibit anaerobic bacteria include nitrates. Such salts are dangerously poisonous in their own right and must be added in carefully controlled quantities and according to proper techniques. Their proper use has however saved many lives and much food spoilage. Like the semi-dried meats, most salted, smoked, and simply-dried meats of different kinds that once were staples in particular regions, now are largely luxury snacks or garnishes; examples include jerky, biltong, and varieties of pemmican, but ham and bacon for instance, still are staples in many communities. Food rotation Food rotation is important to preserve freshness. When food is rotated, the food that has been in storage the longest is used first. As food is used, new food is added to the pantry to replace it; the essential rationale is to use the oldest food as soon as possible so that nothing is in storage too long and becomes unsafe to eat. Labelling food with paper labels on the storage container, marking the date that the container is placed in storage, can make this practice simpler. For emergency preparation Guides for surviving emergency conditions in many parts of the world recommend maintaining a store of essential foods; typically water, cereals, oil, dried milk, and protein rich foods such as beans, lentils, tinned meat and fish. A food storage calculator can be used to help determine how much of these staple foods a person would need to store in order to sustain life for one full year. In addition to storing the basic food items many people choose to supplement their food storage with frozen or preserved garden-grown fruits and vegetables and freeze-dried or canned produce. An unvarying diet of staple foods prepared in the same manner can cause appetite exhaustion, leading to less caloric intake. Another benefit to having a basic supply of food storage in the home is for the potential cost savings. Costs of dry bulk foods (before preparation) are often considerably less than convenience and fresh foods purchased at local markets or supermarkets. There is a significant market in convenience foods for campers, such as dehydrated food products. Commercial food logistics Grain and beans are stored in tall grain elevators, almost always at a rail head near the point of production. The grain is shipped to a final user in hopper cars. In the former Soviet Union, where harvest was poorly controlled, grain was often irradiated at the point of production to suppress mold and insects. In the U.S., threshing and drying is performed in the field, and transport is nearly sterile and in large containers that effectively suppresses pest access, which eliminates the need for irradiation. At any given time, the U.S. usually has about two weeks worth of stored grains for the population. Fresh fruits and vegetables are sometimes packed in plastic packages and cups for fresh premium markets, or placed in large plastic tubs for sauce and soup processors. Fruits and vegetables are usually refrigerated at the earliest possible moment, and even so have a shelf life of two weeks or less. In the United States, livestock is usually transported live, slaughtered at a major distribution point, hung and transported for two days to a week in refrigerated rail cars, and then butchered and sold locally. Before refrigerated rail cars, meat had to be transported live, and this placed its cost so high that only farmers and the wealthy could afford it every day. In Europe much meat is transported live and slaughtered close to the point of sale. In much of Africa and Asia most meat is for local populations is raised, slaughtered and eaten locally, which is believed to be less stressful for the animals involved and minimizes meat storage needs. In Australia and New Zealand, where a large proportion of meat production is for export, meat enters the cold chain early, being stored in large freezer plants before being shipped overseas in freezer ships. Food storage facilities Food storage facilities may include those used for dry goods, or in canning, Food dehydration, pickling, curing and more. They include: Pantry Larder Root cellar Fully dedicated food storage facilities include: Cool store — a large refrigerated room or building Cool warehouse — a very large refrigerated building Silo — used to store grains, like wheat and maize 2800 Polar Way — world's largest food freezer
Technology
Food, water and health
null
292211
https://en.wikipedia.org/wiki/Mud
Mud
Mud (probably , or Middle Dutch) is loam, silt or clay mixed with water. It is usually formed after rainfall or near water sources. Ancient mud deposits hardened over geological time to form sedimentary rock such as shale or mudstone (generally called lutites). When geological deposits of mud are formed in estuaries, the resultant layers are termed bay muds. Mud has also been used for centuries as a construction resource for mostly houses and also used as a binder. An Old English word for it was fen, now in most dialects referring to a type of wetland. Building and construction Adhesive In the construction industry, mud is a semi-fluid material that can be used to coat, seal, or adhere materials. The term "mud" can be used for various semi-fluid materials used in construction including slurry, mortar, plaster, stucco, and concrete. Material Mud, cob, adobe, clay, and many other names are historically used synonymously to mean a mixture of subsoil and water possibly with the addition of stones, gravel, straw, lime, and/or bitumen. This material was used a variety of ways to build walls, floors and even roofs. For thousands of years it was common in most parts of the world to build walls using mudbricks or the wattle and daub, rammed earth or cob techniques and cover the surfaces with earthen plaster. Mudbrick Mud can be made into mud bricks, also called adobe, by mixing mud with water, placing the mixture into moulds and then allowing it to dry in open air. Straw is sometimes used as a binder within the bricks, as it makes them a composite. When the brick would otherwise break, the straw will redistribute the force throughout the brick, decreasing the chance of breakage. Such buildings must be protected from groundwater, usually by building upon a masonry, fired brick, rock or rubble foundation, and also from wind-driven rain in damp climates, usually by deep roof overhangs. In extremely dry climates a well-drained flat roof may be protected with a well-prepared (puddled) and properly maintained dried mud coating, viable as the mud will expand when moistened and so become more water resistant. Adobe mudbricks were commonly used by the Pueblo Indians to build their homes and other necessary structures. In some countries there are entire cities made of mud brick houses. Cow dung and biomass are added to regulate indoor climate. Fired brick Mud that is mostly clay, or a mixture of clay and sand may be used for ceramics, of which one form is the common fired brick. Fired brick are more durable but consume much more energy to produce. Stabilized mud Stabilized mud (earth, soil) is mud which has had a binder such as cement or bitumen added. Examples are mudcrete, landcrete, and soil cement. Pottery Pottery is made by forming a clay body into objects of a required shape and heating them to high temperatures in a kiln which removes all the water from the clay, which induces reactions that lead to permanent changes including increasing their strength and hardening and setting their shape. A clay body can be decorated before or after firing. Prior to some shaping processes, clay must be prepared. Kneading helps to ensure an even moisture content throughout the body. Air trapped within the clay body needs to be removed. This is called de-airing and can be accomplished by a machine called a vacuum pug or manually by wedging. Wedging can also help produce an even moisture content. Once a clay body has been kneaded and de-aired or wedged, it is shaped by a variety of techniques. After shaping it is dried and then fired. In ceramics, the making of liquid mud (called slip) is a stage in the process of refinement of the materials, since larger particles will settle from the liquid. Habitat Land Mud can provide a home for numerous types of animals, including varieties of worms, frogs, snails, clams, and crayfish. Other animals, such as hippopotamuses, pigs, rhinoceroses, water buffalo and elephants, bathe in mud in order to cool off and protect themselves from the sun. Submerged mud can be home to larvae of various insects. Marine life Mud plays an important role in the marine ecosystem. The activities of burrowing animals and fish have a dramatic churning effect on muddy seabeds. This allows the exchange and cycling of oxygen, nutrients, and minerals between water and sediment. Below the surface, the burrows of some species form intricate lattice-like networks and may penetrate a meter or more downwards. This means that the burrowed mud is a productive habitat, providing food and shelter for a wide range of mud-dwellers and other animals that forage in and over the mud. Problems Mud can pose problems for motor traffic when moisture is present, because every vehicle function that changes direction or speed relies on friction between the tires and the road surface, so a layer of mud on the surface of the road or tires can cause the vehicle to hydroplane. People and cars can also become stuck in mud, as in quicksand. Heavy rainfall, snowmelt, or high levels of groundwater may trigger a movement of soil or sediments, possibly causing mudslides, landslides, avalanches, or sinkholes. Mudslides in volcanic terrain (called lahars) occur after eruptions as rain remobilizes loose ash deposits. Mudslides are also common in the western United States during El Niño years due to prolonged rainfall. As food Geophagia is the practice of eating earth or soil-like substances, also known as geophagy, and is practiced by some non-human primates and by humans in some cultures. In other human cultures it is considered an eating disorder and classed as Pica. Foods named "mud" Mississippi mud pie is a chocolate based dessert pie. Mud cookies (baked from literal dirt mixed with oil, water, etc.) are also eaten in the poorest parts of Haiti. Children's recipes for "mud" also exist, which is generally a chocolate or cornstarch-based sludge used more for visual appeal than actual taste. However, it does not contain real mud. "Mud" is also a colloquial slang word for coffee, especially when thick, strong, and/or dark. Recreation Mud bath A mud bath is a bath of mud, commonly from areas where hot spring water can combine with volcanic ash. Mud baths have existed for thousands of years, and can be found now in high-end spas. Mud wallow Mud wallows are a common source of entertainment for children. Mud wallows can be any shape, size, depth and some can have water as well as mud. Usually wallows are shallow dips in the ground that have been flooded and were full of dirt and those two have mixed to make a squishy mud wallow. Mud sports Mud sports are sports that take place in, or heavily incorporate, mud. Examples include: Mud bogging, an off-road motorsport popular in Canada and the United States in which the goal is to drive a vehicle through a pit of mud or a track of a set length. Mud runs, where contestants run and crawl through mud bogs and other obstacles. Dirt biking, biking through muddy tracks and courses. Mud wrestling, a form of wrestling that takes place in mud. Other uses Mud can be used in a dunk tank. Baseball rubbing mud is used to remove the sheen from new baseballs. Children often like to make mud pies, throw mud at each other and play barefoot and cover their bare feet in mud and squish it between their toes. Mud can be smeared across the skin as a repellent from mosquitoes. Many animals cover themselves in mud (wallowing) to cool off.
Physical sciences
Sedimentology
Earth science
292340
https://en.wikipedia.org/wiki/Flavonoid
Flavonoid
Flavonoids (or bioflavonoids; from the Latin word flavus, meaning yellow, their color in nature) are a class of polyphenolic secondary metabolites found in plants, and thus commonly consumed in the diets of humans. Chemically, flavonoids have the general structure of a 15-carbon skeleton, which consists of two phenyl rings (A and B) and a heterocyclic ring (C, the ring containing the embedded oxygen). This carbon structure can be abbreviated C6-C3-C6. According to the IUPAC nomenclature, they can be classified into: flavonoids or bioflavonoids isoflavonoids, derived from 3-phenylchromen-4-one (3-phenyl-1,4-benzopyrone) structure neoflavonoids, derived from 4-phenylcoumarin (4-phenyl-1,2-benzopyrone) structure The three flavonoid classes above are all ketone-containing compounds and as such, anthoxanthins (flavones and flavonols). This class was the first to be termed bioflavonoids. The terms flavonoid and bioflavonoid have also been more loosely used to describe non-ketone polyhydroxy polyphenol compounds, which are more specifically termed flavanoids. The three cycles or heterocycles in the flavonoid backbone are generally called ring A, B, and C. Ring A usually shows a phloroglucinol substitution pattern. History In the 1930s, Albert Szent-Györgyi and other scientists discovered that Vitamin C alone was not as effective at preventing scurvy as the crude yellow extract from oranges, lemons or paprika. They attributed the increased activity of this extract to the other substances in this mixture, which they referred to as "citrin" (referring to citrus) or "Vitamin P" (a reference to its effect on reducing the permeability of capillaries). The substances in question (hesperidin, eriodictyol, hesperidin methyl chalcone and neohesperidin) were however later shown not to fulfil the criteria of a vitamin, so that this term is now obsolete. Biosynthesis Flavonoids are secondary metabolites synthesized mainly by plants. The general structure of flavonoids is a fifteen-carbon skeleton, containing two benzene rings connected by a three-carbon linking chain. Therefore, they are depicted as C6-C3-C6 compounds. Depending on the chemical structure, degree of oxidation, and unsaturation of the linking chain (C3), flavonoids can be classified into different groups, such as anthocyanidins, flavonols, flavanones, flavan-3-ols, flavanonols, flavones, and isoflavones. Chalcones, also called chalconoids, although lacking the heterocyclic ring, are also classified as flavonoids. Furthermore, flavonoids can be found in plants in glycoside-bound and free aglycone forms. The glycoside-bound form is the most common flavone and flavonol form consumed in the diet. Functions of flavonoids in plants Flavonoids are widely distributed in plants, fulfilling many functions. They are the most important plant pigments for flower coloration, producing yellow or red/blue pigmentation in petals designed to attract pollinator animals. In higher plants, they are involved in UV filtration, symbiotic nitrogen fixation, and floral pigmentation. They may also act as chemical messengers, physiological regulators, and cell cycle inhibitors. Flavonoids secreted by the root of their host plant help Rhizobia in the infection stage of their symbiotic relationship with legumes like peas, beans, clover, and soy. Rhizobia living in soil are able to sense the flavonoids and this triggers the secretion of Nod factors, which in turn are recognized by the host plant and can lead to root hair deformation and several cellular responses such as ion fluxes and the formation of a root nodule. In addition, some flavonoids have inhibitory activity against organisms that cause plant diseases, e.g. Fusarium oxysporum. Subgroups Over 5000 naturally occurring flavonoids have been characterized from various plants. They have been classified according to their chemical structure, and are usually subdivided into the following subgroups (for further reading see): Anthocyanidins Anthocyanidins are the aglycones of anthocyanins; they use the flavylium (2-phenylchromenylium) ion skeleton. Examples: cyanidin, delphinidin, malvidin, pelargonidin, peonidin, petunidin Anthoxanthins Anthoxanthins are divided into two groups: {| class="wikitable" !rowspan=3|Group !colspan=4|Skeleton !rowspan=3|Examples |- !rowspan=2|Description !colspan=2|Functional groups !rowspan=2|Structural formula |- !|3-hydroxyl !|2,3-dihydro |- |style="text-align:center"| |style="text-align:center"|- |style="text-align:center; font-size:x-large"|✗ |style="text-align:center; font-size:x-large"|✗ || ||Luteolin, Apigenin, Tangeritin |- |style="text-align:center"|or |style="text-align:center"|-- |style="text-align:center; font-size:x-large"|✓ |style="text-align:center; font-size:x-large"|✗ || ||Quercetin, Kaempferol, Myricetin, Fisetin, Galangin, Isorhamnetin, Pachypodol, Rhamnazin, Pyranoflavonols, Furanoflavonols, |} Flavanones Flavanones Flavanonols Flavanonols Flavans Include flavan-3-ols (flavanols), flavan-4-ols and flavan-3,4-diols. Flavan-3-ols (flavanols) Flavan-3-ols use the 2-phenyl-3,4-dihydro-2H-chromen-3-ol skeleton Examples: Catechin (C), Gallocatechin (GC), Catechin 3-gallate (Cg), Gallocatechin 3-gallate (GCg), Epicatechins (Epicatechin (EC)), Epigallocatechin (EGC), Epicatechin 3-gallate (ECg), Epigallocatechin 3-gallate (EGCg) Theaflavin Examples: Theaflavin-3-gallate, Theaflavin-3'-gallate, Theaflavin-3,3'-digallate Thearubigin Proanthocyanidins are dimers, trimers, oligomers, or polymers of the flavanols Isoflavonoids Isoflavonoids Isoflavones use the 3-phenylchromen-4-one skeleton (with no hydroxyl group substitution on carbon at position 2) Examples: Genistein, Daidzein, Glycitein Isoflavanes Isoflavandiols Isoflavenes Coumestans Pterocarpans Dietary sources Flavonoids (specifically flavanoids such as the catechins) are "the most common group of polyphenolic compounds in the human diet and are found ubiquitously in plants". Flavonols, the original bioflavonoids such as quercetin, are also found ubiquitously, but in lesser quantities. The widespread distribution of flavonoids, their variety and their relatively low toxicity compared to other active plant compounds (for instance alkaloids) mean that many animals, including humans, ingest significant quantities in their diet. Foods with a high flavonoid content include parsley, onions, blueberries and strawberries, black tea, bananas, and citrus fruits. One study found high flavonoid content in buckwheat. Citrus flavonoids include hesperidin (a glycoside of the flavanone hesperetin), quercitrin, rutin (two glycosides of quercetin, and the flavone tangeritin. The flavonoids are less concentrated in the pulp than in the peels (for example, 165 versus 1156 mg/100 g in pulp versus peel of satsuma mandarin, and 164 vis-à-vis 804 mg/100 g in pulp versus peel of clementine). Peanut (red) skin contains significant polyphenol content, including flavonoids. {| class="wikitable" |+Flavonoid content in food (mg/100 g) ! Food source ! Flavones ! Flavonols ! Flavanones |- | Red onion | 0 | 4–100 | 0 |- | Parsley, fresh | 24–634 | 8–10 | 0 |- | Thyme, fresh | 56 | 0 | 0 |- | Lemon juice, fresh | 0 | 0–2 | 2–175 |} Dietary intake Food composition data for flavonoids were provided by the USDA database on flavonoids. In the United States NHANES survey, mean flavonoid intake was 190 mg per day in adults, with flavan-3-ols as the main contributor. In the European Union, based on data from EFSA, mean flavonoid intake was 140 mg/d, although there were considerable differences among individual countries. The main type of flavonoids consumed in the EU and USA were flavan-3-ols (80% for USA adults), mainly from tea or cocoa in chocolate, while intake of other flavonoids was considerably lower. Research Neither the United States Food and Drug Administration (FDA) nor the European Food Safety Authority (EFSA) has approved any flavonoids as prescription drugs. The U.S. FDA has warned numerous dietary supplement and food manufacturers, including Unilever, producer of Lipton tea in the U.S., about illegal advertising and misleading health claims regarding flavonoids, such as that they lower cholesterol or relieve pain. Metabolism and excretion Flavonoids are poorly absorbed in the human body (less than 5%), then are quickly metabolized into smaller fragments with unknown properties, and rapidly excreted. Flavonoids have negligible antioxidant activity in the body, and the increase in antioxidant capacity of blood seen after consumption of flavonoid-rich foods is not caused directly by flavonoids, but by production of uric acid resulting from flavonoid depolymerization and excretion. Microbial metabolism is a major contributor to the overall metabolism of dietary flavonoids. Inflammation Inflammation has been implicated as a possible origin of numerous local and systemic diseases, such as cancer, cardiovascular disorders, diabetes mellitus, and celiac disease. There is no clinical evidence that dietary flavonoids affect any of these diseases. Cancer Clinical studies investigating the relationship between flavonoid consumption and cancer prevention or development are conflicting for most types of cancer, probably because most human studies have weak designs, such as a small sample size. There is little evidence to indicate that dietary flavonoids affect human cancer risk in general. Cardiovascular diseases Although no significant association has been found between flavan-3-ol intake and cardiovascular disease mortality, clinical trials have shown improved endothelial function and reduced blood pressure (with a few studies showing inconsistent results). Reviews of cohort studies in 2013 found that the studies had too many limitations to determine a possible relationship between increased flavonoid intake and decreased risk of cardiovascular disease, although a trend for an inverse relationship existed. In 2013, the EFSA decided to permit health claims that 200 mg/day of cocoa flavanols "help[s] maintain the elasticity of blood vessels." The FDA followed suit in 2023, stating that there is "supportive, but not conclusive" evidence that 200 mg per day of cocoa flavanols can reduce the risk of cardiovascular disease. This is greater than the levels found in typical chocolate bars, which can also contribute to weight gain, potentially harming cardiovascular health. Synthesis, detection, quantification, and semi-synthetic alterations Color spectrum Flavonoid synthesis in plants is induced by light color spectrums at both high and low energy radiations. Low energy radiations are accepted by phytochrome, while high energy radiations are accepted by carotenoids, flavins, cryptochromes in addition to phytochromes. The photomorphogenic process of phytochrome-mediated flavonoid biosynthesis has been observed in Amaranthus, barley, maize, Sorghum and turnip. Red light promotes flavonoid synthesis. Availability through microorganisms Research has shown production of flavonoid molecules from genetically engineered microorganisms. Tests for detection Shinoda test Four pieces of magnesium filings are added to the ethanolic extract followed by few drops of concentrated hydrochloric acid. A pink or red colour indicates the presence of flavonoid. Colours varying from orange to red indicated flavones, red to crimson indicated flavonoids, crimson to magenta indicated flavonones. Sodium hydroxide test About 5 mg of the compound is dissolved in water, warmed, and filtered. 10% aqueous sodium hydroxide is added to 2 ml of this solution. This produces a yellow coloration. A change in color from yellow to colorless on addition of dilute hydrochloric acid is an indication for the presence of flavonoids. p-Dimethylaminocinnamaldehyde test A colorimetric assay based upon the reaction of A-rings with the chromogen p-dimethylaminocinnamaldehyde (DMACA) has been developed for flavanoids in beer that can be compared with the vanillin procedure. Quantification Lamaison and Carnet have designed a test for the determination of the total flavonoid content of a sample (AlCI3 method). After proper mixing of the sample and the reagent, the mixture is incubated for ten minutes at ambient temperature and the absorbance of the solution is read at 440 nm. Flavonoid content is expressed in mg/g of quercetin. Semi-synthetic alterations Immobilized Candida antarctica lipase can be used to catalyze the regioselective acylation of flavonoids.
Physical sciences
Polyphenols
Chemistry
292347
https://en.wikipedia.org/wiki/Anthocyanidin
Anthocyanidin
Anthocyanidins are common plant pigments, the aglycones of anthocyanins. They are based on the flavylium cation, an oxonium ion, with various groups substituted for its hydrogen atoms. They generally change color from red through purple, blue, and bluish green as a function of pH. Anthocyanidins are an important subclass of the polymethine dyes and flavonoids. The flavylium cation is a chromenylium cation with a phenyl group substituted in position 2; and chromenylium (also called benzopyrylium) is a bicyclic version of pyrylium. The positive charge can move around the molecule. At least 31 monomeric anthocyanidins have been properly identified in living organisms, mostly as the core components of anthocyanins. The latter are responsible for the red, purple, blue, or black color of many fruits (like grapes and blueberries), flowers (like roses), leaves (like purple cabbage), and even tubers (like radishes and purple yams). They are also found in some animals. Classification 3-Deoxyanthocyanidins such as luteolinidin are a class of anthocyanidins lacking an hydroxyl group on carbon 3. {| class="wikitable centered" style="text-align:center" |+ Selected anthocyanidins and their substitutions |- class="hintergrundfarbe5" ! Anthocyanidin ! Basic structure (R3 & R4′ = −OH) ! R3′ ! R5′ ! R5 ! R6 ! R7 |- | Aurantinidin | rowspan=12 | | −H | −H | −OH | −OH | −OH |- | Capensinidin | −OCH3 | −OCH3 | −OCH3 | −H | −OH |- | Cyanidin | −OH | −H | −OH | −H | −OH |- | Delphinidin | −OH | −OH | −OH | −H | −OH |- | Europinidin | −OCH3 | −OH | −OCH3 | −H | −OH |- | Hirsutidin | −OCH3 | −OCH3 | −OH | −H | −OCH3 |- | Malvidin | −OCH3 | −OCH3 | −OH | −H | −OH |- | Pelargonidin | −H | −H | −OH | −H | −OH |- | Peonidin | −OCH3 | −H | −OH | −H | −OH |- | Petunidin | −OH | −OCH3 | −OH | −H | −OH |- | Pulchellidin | −OH | −OH | −OCH3 | −H | −OH |- | Rosinidin | −OCH3 | −H | −OH | −H | −OCH3 |} Natural occurrence Most plant anthocyanins are based on cyanidin (30%), delphinidin (22%), and pelargonidin (18%), respectively. Altogether 20% of the anthocyanins are based on the three common anthocyanidins (peonidin, malvidin, and petunidin) that are methylated. Around 3%, 3%, and 2% of the anthocyanins or anthocyanidins are respectively labeled as 3-desoxyanthocyanidins, rare methylated anthocyanidins, and 6-hydroxyanthocyanidins, respectively. In bryophytes, anthocyanins are usually based on 3-desoxyanthocyanidins located in the cell wall. A new anthocyanidin, riccionidin A, has been isolated from the liverwort Ricciocarpos natans. It could be derived from 6,7,2′,4′,6′-pentahydroxyflavylium, having undergone ring closure of the 6’ -hydroxyl at the 3-position. Its visible spectrum in methanolic HCl is at 494 nm. This pigment was accompanied by riccionidin B, which most probably is based on two molecules of riccionidin A linked via the 3′- or 5′-positions. Both pigments were also detected in the liverworts Marchantia polymorpha, Riccia duplex, and Scapania undulata. Effect of pH The stability of anthocyanidins is dependent on pH. At a low pH (acidic conditions), colored anthocyanidins are present, whereas at a higher pH (basic conditions) the colorless chalcones forms are present.
Physical sciences
Polyphenols
Chemistry
292371
https://en.wikipedia.org/wiki/Pupa
Pupa
A pupa (; : pupae) is the life stage of some insects undergoing transformation between immature and mature stages. Insects that go through a pupal stage are holometabolous: they go through four distinct stages in their life cycle, the stages thereof being egg, larva, pupa, and imago. The processes of entering and completing the pupal stage are controlled by the insect's hormones, especially juvenile hormone, prothoracicotropic hormone, and ecdysone. The act of becoming a pupa is called pupation, and the act of emerging from the pupal case is called eclosion or emergence. The pupae of different groups of insects have different names such as chrysalis for the pupae of butterflies and tumbler for those of the mosquito family. Pupae may further be enclosed in other structures such as cocoons, nests, or shells. Position in life cycle The pupal stage follows the larval stage, or in some cases a prepupal stage, and precedes adulthood (imago) in insects with complete metamorphosis. The pupa is a non-feeding, usually sessile stage, or highly active as in mosquitoes. It is during the pupal stage that the adult structures of the insect are formed while the larval structures are broken down. The adult structures grow from imaginal discs. Duration The pupal stage may last weeks, months, or even years, depending on temperature and the species of insect. For example, the pupal stage lasts eight to fifteen days in monarch butterflies. The pupa may enter dormancy or diapause until the appropriate season to emerge as an adult insect. In temperate climates pupae usually stay dormant during winter, while in the tropics pupae usually do so during the dry season. Emergence Insects emerge (eclose) from pupae by splitting the pupal case. Most butterflies emerge in the morning. In mosquitoes, the emergence is in the evening or night. In fleas, the process is triggered by vibrations that indicate the possible presence of a suitable host. Prior to emergence, the adult inside the pupal exoskeleton is termed pharate. Once the pharate adult has eclosed from the pupa, the empty pupal exoskeleton is called an exuvia; in most hymenopterans (ants, bees and wasps) the exuvia is so thin and membranous that it becomes "crumpled" as it is shed. Measuring the timing of this emergence is of interest to chronobiologists because the process is regulated by circadian clocks in many species, necessitating different assays to measure eclosion timing. Pupal mating In a few taxa of the Lepidoptera, especially Heliconius, pupal mating is an extreme form of reproductive strategy in which the adult male mates with a female pupa about to emerge, or with the newly moulted female; this is accompanied by other actions such as capping of the reproductive system of the female with the sphragis, denying access to other males, or by exuding an anti-aphrodisiac pheromone. Defense Pupae are usually immobile and are largely defenseless. To overcome this, pupae often are covered with a cocoon, conceal themselves in the environment, or form underground. Some species of Lycaenid butterflies are protected in their pupal stage by ants. Another means of defense by pupae of other species is the capability of making sounds or vibrations to scare potential predators. A few species use chemical defenses including toxic secretions. The pupae of social hymenopterans are protected by adult members of the hive. Types Based on the presence or absence of articulated mandibles that are employed in emerging from a cocoon or pupal case, the pupae can be classified in to two types: Decticous pupa – a pupa with articulated mandibles. Examples are pupae of the orders Neuroptera, Mecoptera, Trichoptera and few Lepidoptera families. Adecticous pupa – a pupa without articulated mandibles. Examples include the orders Strepsiptera, Coleoptera, Hymenoptera, Diptera and Siphonaptera. Based on whether the pupal appendages are free or attached to the body, the pupae can be classified as one of three types: Exarate pupa – appendages are free and are not usually encapsulated within a cocoon. Decticous pupae are always exarate; some adecticous pupae are as well. (Neuroptera, Trichoptera, Cyclorrhapha of Dipterans, Siphonaptera, most Coleoptera, Hymenoptera, and few Lepidoptera). Obtect pupa – appendages are attached closely to the body and are commonly encapsulated within a cocoon. Some adecticous pupa are obtect forms. (Most Lepidoptera, Nematocera and Brachycera of Dipterans, Staphylinidae and Chrysomelidae Coleopterans, many Chalcidoidea Hymenopterans) Coarctate pupa – enclosed in a hardened cuticle of the penultimate larval instar called a puparium. However, the pupa itself is of the exarate adecticous pupal form. (Cyclorrhapha of Dipterans). Chrysalis A chrysalis (, from , , plural: , also known as an aurelia) or nympha is the pupal stage of butterflies. The term is derived from the metallic–gold coloration found in the pupae of many butterflies, referred to by the Ancient Greek term () for gold. When the caterpillar is fully grown, it makes a button of silk which it uses to fasten its body to a leaf or a twig. Then the caterpillar's skin comes off for the final time. Under this old skin is a hard skin called a chrysalis. Because chrysalises are often showy and are formed in the open, they are the most familiar examples of pupae. Most chrysalides are attached to a surface by a Velcro-like arrangement of a silken pad spun by the caterpillar, usually cemented to the underside of a perch, and the cremastral hook or hooks protruding from the rear of the chrysalis or cremaster at the tip of the pupal abdomen by which the caterpillar fixes itself to the pad of silk. ( 'suspended') Like other types of pupae, the chrysalis stage in most butterflies is one in which there is little movement. However, some butterfly pupae are capable of moving the abdominal segments to produce sounds or to scare away potential predators. Within the chrysalis, growth and differentiation occur. The adult butterfly emerges (ecloses) from this and expands its wings by pumping haemolymph into the wing veins. Although this sudden and rapid change from pupa to imago is often called metamorphosis, metamorphosis is really the whole series of changes that an insect undergoes from egg to adult. When emerging, the butterfly uses a liquid, sometimes called cocoonase, which softens the shell of the chrysalis. Additionally, it uses two sharp claws located on the thick joints at the base of the forewings to help make its way out. Having emerged from the chrysalis, the butterfly will usually sit on the empty shell in order to expand and harden its wings. However, if the chrysalis was near the ground (such as if it fell off from its silk pad), the butterfly would find another vertical surface to rest upon and harden its wings (such as a wall or fence). Moth pupae are usually dark in color and either formed in underground cells, loose in the soil, or their pupa is contained in a protective silk case called a cocoon. The pupa of some species such as the hornet moth develops sharp ridges around the outside called adminicula that allow the pupa to move from its place of concealment inside a tree trunk when it is time for the adult to emerge. Pupa, chrysalis, and cocoon are frequently confused, but are quite distinct from each other. The pupa is the stage between the larva and adult stages. The chrysalis generally refers to a butterfly pupa although the term may be misleading as there are some moths whose pupae resembles a chrysalis, e.g.: the plume winged moths of the family Pterophoridae and some geometrid moths. A cocoon is a silk case that the larvae of moths, and sometimes other insects, spin around the pupa. Cocoon A cocoon is a casing spun of silk by many moths and caterpillars, and numerous other holometabolous insect larvae as a protective covering for the pupa. Cocoons may be tough or soft, opaque or translucent, solid or meshlike, of various colors, or composed of multiple layers, depending on the type of insect larva producing it. Many moth caterpillars shed the larval hairs (setae) and incorporate them into the cocoon; if these are urticating hairs then the cocoon is also irritating to the touch. Some larvae attach small twigs, fecal pellets or pieces of vegetation to the outside of their cocoon in an attempt to disguise it from predators. Others spin their cocoon in a concealed location—on the underside of a leaf, in a crevice, down near the base of a tree trunk, suspended from a twig or concealed in the leaf litter. Contrary to popular belief, larvae do not completely liquify inside the cocoon. The silk in the cocoon of the silk moth can be unraveled to harvest silk fibre which makes this moth the most economically important of all lepidopterans. The silk moth is the only completely domesticated lepidopteran; it does not exist in the wild. Insects that pupate in a cocoon must escape from it, and they do this either by the pupa cutting its way out, or by secreting enzymes, sometimes called cocoonase, that soften the cocoon. Some cocoons are constructed with built-in lines of weakness along which they will tear easily from inside, or with exit holes that only allow a one-way passage out; such features facilitate the escape of the adult insect after it emerges from the pupal skin. Puparium Some pupae remain inside the exoskeleton of the final larval instar and this last larval "shell" is called a puparium (plural, puparia). Flies of the group Muscomorpha have puparia, as do members of the order Strepsiptera, and the Hemipteran family Aleyrodidae. Gallery
Biology and health sciences
Animal ontogeny
null
292381
https://en.wikipedia.org/wiki/Brazil%20nut
Brazil nut
The Brazil nut (Bertholletia excelsa) is a South American tree in the family Lecythidaceae, and it is also the name of the tree's commercially harvested edible seeds. It is one of the largest and longest-lived trees in the Amazon rainforest. The fruit and its nutshell – containing the edible Brazil nut – are relatively large and weigh as much as in total. As food, Brazil nuts are notable for diverse content of micronutrients, especially a high amount of selenium. The wood of the Brazil nut tree is prized for its quality in carpentry, flooring, and heavy construction. Common names In Portuguese-speaking countries, like Brazil, they are variously called "" (meaning "chestnut from Brazil" in Portuguese), "" (meaning "chestnut from Pará" in Portuguese), with other names: castanha-da-amazônia, castanha-do-acre, "" (meaning "Amazonian nut" in Portuguese), noz boliviana, tocari ("probably of Carib origin"), and tururi (from Tupi turu'ri) also used. In various Spanish-speaking countries of South America, Brazil nuts are called , , or . In North America, as early as 1896, Brazil nuts were sometimes known by the slang term "nigger toes", a vulgarity that fell out of use after the racial slur became more socially unacceptable. Description The Brazil nut is a large tree, reaching tall, and with a trunk in diameter, making it among the largest of trees in the Amazon rainforest. It may live for 500 years or more, and can often reach a thousand years of age. The stem is straight and commonly without branches for well over half the tree's height, with a large, emergent crown of long branches above the surrounding canopy of other trees. The bark is grayish and smooth. The leaves are dry-season deciduous, alternate, simple, entire or crenate, oblong, long, and broad. The flowers are small, greenish-white, in panicles long; each flower has a two-parted, deciduous calyx, six unequal cream-colored petals, and numerous stamens united into a broad, hood-shaped mass. Reproduction Brazil nut trees produce fruit almost exclusively in pristine forests, as disturbed forests lack the large-bodied bees of the genera Bombus, Centris, Epicharis, Eulaema, and Xylocopa, which are the only ones capable of pollinating the tree's flowers, with different bee genera being the primary pollinators in different areas, and different times of year. Brazil nuts have been harvested from plantations, but production is low and is currently not economically viable. The fruit takes 14 months to mature after pollination of the flowers. The fruit itself is a large capsule in diameter, resembling a coconut endocarp in size and weighing up to . It has a hard, woody shell thick, which contains eight to 24 wedge-shaped seeds long (the "Brazil nuts") packed like the segments of an orange, but not limited to one whorl of segments. Up to three whorls can be stacked onto each other, with the polar ends of the segments of the middle whorl nestling into the upper and lower whorls (see illustration above). The capsule contains a small hole at one end, which enables large rodents like the agouti to gnaw it open. They then eat some of the seeds inside while burying others for later use; some of these are able to germinate into new Brazil nut trees. Most of the seeds are "planted" by the agoutis in caches during wet season, and the young saplings may have to wait years, in a state of dormancy, for a tree to fall and sunlight to reach it, when it starts growing again. Taxonomy The Brazil nut family, the Lecythidaceae, is in the order Ericales, as are other well-known plants such as blueberries, cranberries, sapote, gutta-percha, tea, phlox, and persimmons. The tree is the only species in the monotypic genus Bertholletia, named after French chemist Claude Louis Berthollet. Distribution and habitat The Brazil nut is native to the Guianas, Venezuela, Brazil, eastern Colombia, eastern Peru, and eastern Bolivia. It occurs as scattered trees in large forests on the banks of the Amazon River, Rio Negro, Tapajós, and the Orinoco. The fruit is heavy and rigid; when the fruits fall, they pose a serious threat to vehicles and potential for traumatic brain injury of people passing under the tree. Ecology Capuchin monkeys have been reported to open Brazil nuts using a stone as an anvil. Production In 2020, global production of Brazil nuts (in shells) was 69,658 tonnes, most of which derive from wild harvests in tropical forests, especially the Amazon regions of Brazil and Bolivia which produced 92% of the world total (table). Environmental effects of harvesting Since most of the production for international trade is harvested in the wild, the business arrangement has been advanced as a model for generating income from a tropical forest without destroying it. The nuts are most often gathered by migrant workers known as castañeros (in Spanish) or castanheiros (in Portuguese). Logging is a significant threat to the sustainability of the Brazil nut-harvesting industry. Analysis of tree ages in areas that are harvested shows that moderate and intense gathering takes so many seeds that not enough are left to replace older trees as they die. Sites with light gathering activities had many young trees, while sites with intense gathering practices had nearly none. European Union import regulation In 2003, the European Union imposed strict regulations on the import of Brazilian-harvested Brazil nuts in their shells, as the shells are considered to contain unsafe levels of aflatoxins, a potential cause of liver cancer. Toxicity Brazil nuts are susceptible to contamination by aflatoxins, produced by fungi, once they fall to the ground. Aflatoxins can cause liver damage, including possible cancer, if consumed. Aflatoxin levels have been found in Brazil nuts during inspections that were far higher than the limits set by the EU. However, mechanical sorting and drying was found to eliminate 98% of aflatoxins; a 2003 EU ban on importation was rescinded after new tolerance levels were set. The nuts may contain traces of radium, a radioactive element, with a kilogram of nuts containing an activity between . This level of radium is small, although higher than in other common foods. According to Oak Ridge Associated Universities, elevated levels of radium in the soil does not directly cause the concentration of radium, but "the very extensive root system of the tree" can concentrate naturally occurring radioactive material, when present in the soil. Radium can be concentrated in nuts only if it is present in the soil. Brazil nuts also contain barium, a metal with a chemical behavior quite similar to radium. While barium, if ingested, can have toxic effects, such as weakness, vomiting, or diarrhea, the amount present in Brazil nuts is orders of magnitude too small to have noticeable health effects. Uses Nutrition Brazil nuts are 3% water, 14% protein, 12% carbohydrates, and 66% fats. The fat components are 16% saturated, 24% monounsaturated, and 24% polyunsaturated. In a reference amount, Brazil nuts supply 659 calories, and are a rich source (20% or more of the Daily Value, DV) of dietary fiber (30% DV), thiamin (54% DV), vitamin E (38% DV), magnesium (106% DV), phosphorus (104% DV), manganese (57% DV), and zinc (43% DV). Calcium, iron, and potassium are present in moderate amounts (10–19% DV). Selenium Brazil nuts are a particularly rich source of selenium, with just supplying 544 micrograms of selenium or 10 times the DV of 55 micrograms. However, the amount of selenium within batches of nuts may vary considerably. The high selenium content is used as a biomarker in studies of selenium intake and deficiency. Consumption of just one Brazil nut per day over 8 weeks was sufficient to restore selenium blood levels and increase HDL cholesterol in obese women. Oil Brazil nut oil contains 48% unsaturated fatty acids composed mainly of oleic and linoleic acids, the phytosterol, beta-sitosterol, and fat-soluble vitamin E. The following table presents the composition of fatty acids in Brazil nut oil: Wood The lumber from Brazil nut trees (not to be confused with Brazilwood) is of excellent quality, having diverse uses from flooring to heavy construction. Logging the trees is prohibited by law in all three producing countries (Brazil, Bolivia, and Peru). Illegal extraction of timber and land clearances present continuing threats. In Brazil, cutting down a Brazil nut tree requires previous authorization from the Brazilian Institute of Environment and Renewable Natural Resources. Other uses Brazil nut oil is used as a lubricant in clocks and in the manufacturing of paint and cosmetics, such as soap and perfume. Because of its hardness, the Brazil nutshell is often pulverized and used as an abrasive to polish materials, such as metals and ceramics, in the same way as jeweler's rouge, while charcoal from the shells can be used to purify water.
Biology and health sciences
Ericales
null
292420
https://en.wikipedia.org/wiki/Emission%20spectrum
Emission spectrum
The emission spectrum of a chemical element or chemical compound is the spectrum of frequencies of electromagnetic radiation emitted due to electrons making a transition from a high energy state to a lower energy state. The photon energy of the emitted photons is equal to the energy difference between the two states. There are many possible electron transitions for each atom, and each transition has a specific energy difference. This collection of different transitions, leading to different radiated wavelengths, make up an emission spectrum. Each element's emission spectrum is unique. Therefore, spectroscopy can be used to identify elements in matter of unknown composition. Similarly, the emission spectra of molecules can be used in chemical analysis of substances. Emission In physics, emission is the process by which a higher energy quantum mechanical state of a particle becomes converted to a lower one through the emission of a photon, resulting in the production of light. The frequency of light emitted is a function of the energy of the transition. Since energy must be conserved, the energy difference between the two states equals the energy carried off by the photon. The energy states of the transitions can lead to emissions over a very large range of frequencies. For example, visible light is emitted by the coupling of electronic states in atoms and molecules (then the phenomenon is called fluorescence or phosphorescence). On the other hand, nuclear shell transitions can emit high energy gamma rays, while nuclear spin transitions emit low energy radio waves. The emittance of an object quantifies how much light is emitted by it. This may be related to other properties of the object through the Stefan–Boltzmann law. For most substances, the amount of emission varies with the temperature and the spectroscopic composition of the object, leading to the appearance of color temperature and emission lines. Precise measurements at many wavelengths allow the identification of a substance via emission spectroscopy. Emission of radiation is typically described using semi-classical quantum mechanics: the particle's energy levels and spacings are determined from quantum mechanics, and light is treated as an oscillating electric field that can drive a transition if it is in resonance with the system's natural frequency. The quantum mechanics problem is treated using time-dependent perturbation theory and leads to the general result known as Fermi's golden rule. The description has been superseded by quantum electrodynamics, although the semi-classical version continues to be more useful in most practical computations. Origins When the electrons in the atom are excited, for example by being heated, the additional energy pushes the electrons to higher energy orbitals. When the electrons fall back down and leave the excited state, energy is re-emitted in the form of a photon. The wavelength (or equivalently, frequency) of the photon is determined by the difference in energy between the two states. These emitted photons form the element's spectrum. The fact that only certain colors appear in an element's atomic emission spectrum means that only certain frequencies of light are emitted. Each of these frequencies are related to energy by the formula: where is the energy of the photon, is its frequency, and is the Planck constant. This concludes that only photons with specific energies are emitted by the atom. The principle of the atomic emission spectrum explains the varied colors in neon signs, as well as chemical flame test results (described below). The frequencies of light that an atom can emit are dependent on states the electrons can be in. When excited, an electron moves to a higher energy level or orbital. When the electron falls back to its ground level the light is emitted. The above picture shows the visible light emission spectrum for hydrogen. If only a single atom of hydrogen were present, then only a single wavelength would be observed at a given instant. Several of the possible emissions are observed because the sample contains many hydrogen atoms that are in different initial energy states and reach different final energy states. These different combinations lead to simultaneous emissions at different wavelengths. Radiation from molecules As well as the electronic transitions discussed above, the energy of a molecule can also change via rotational, vibrational, and vibronic (combined vibrational and electronic) transitions. These energy transitions often lead to closely spaced groups of many different spectral lines, known as spectral bands. Unresolved band spectra may appear as a spectral continuum. Emission spectroscopy Light consists of electromagnetic radiation of different wavelengths. Therefore, when the elements or their compounds are heated either on a flame or by an electric arc they emit energy in the form of light. Analysis of this light, with the help of a spectroscope gives us a discontinuous spectrum. A spectroscope or a spectrometer is an instrument which is used for separating the components of light, which have different wavelengths. The spectrum appears in a series of lines called the line spectrum. This line spectrum is called an atomic spectrum when it originates from an atom in elemental form. Each element has a different atomic spectrum. The production of line spectra by the atoms of an element indicate that an atom can radiate only a certain amount of energy. This leads to the conclusion that bound electrons cannot have just any amount of energy but only a certain amount of energy. The emission spectrum can be used to determine the composition of a material, since it is different for each element of the periodic table. One example is astronomical spectroscopy: identifying the composition of stars by analysing the received light. The emission spectrum characteristics of some elements are plainly visible to the naked eye when these elements are heated. For example, when platinum wire is dipped into a sodium nitrate solution and then inserted into a flame, the sodium atoms emit an amber yellow color. Similarly, when indium is inserted into a flame, the flame becomes blue. These definite characteristics allow elements to be identified by their atomic emission spectrum. Not all emitted lights are perceptible to the naked eye, as the spectrum also includes ultraviolet rays and infrared radiation. An emission spectrum is formed when an excited gas is viewed directly through a spectroscope. Emission spectroscopy is a spectroscopic technique which examines the wavelengths of photons emitted by atoms or molecules during their transition from an excited state to a lower energy state. Each element emits a characteristic set of discrete wavelengths according to its electronic structure, and by observing these wavelengths the elemental composition of the sample can be determined. Emission spectroscopy developed in the late 19th century and efforts in theoretical explanation of atomic emission spectra eventually led to quantum mechanics. There are many ways in which atoms can be brought to an excited state. Interaction with electromagnetic radiation is used in fluorescence spectroscopy, protons or other heavier particles in particle-induced X-ray emission and electrons or X-ray photons in energy-dispersive X-ray spectroscopy or X-ray fluorescence. The simplest method is to heat the sample to a high temperature, after which the excitations are produced by collisions between the sample atoms. This method is used in flame emission spectroscopy, and it was also the method used by Anders Jonas Ångström when he discovered the phenomenon of discrete emission lines in the 1850s. Although the emission lines are caused by a transition between quantized energy states and may at first look very sharp, they do have a finite width, i.e. they are composed of more than one wavelength of light. This spectral line broadening has many different causes. Emission spectroscopy is often referred to as optical emission spectroscopy because of the light nature of what is being emitted. History In 1756 Thomas Melvill observed the emission of distinct patterns of colour when salts were added to alcohol flames. By 1785 James Gregory discovered the principles of diffraction grating and American astronomer David Rittenhouse made the first engineered diffraction grating. In 1821 Joseph von Fraunhofer solidified this significant experimental leap of replacing a prism as the source of wavelength dispersion improving the spectral resolution and allowing for the dispersed wavelengths to be quantified. In 1835, Charles Wheatstone reported that different metals could be distinguished by bright lines in the emission spectra of their sparks, thereby introducing an alternative to flame spectroscopy. In 1849, J. B. L. Foucault experimentally demonstrated that absorption and emission lines at the same wavelength are both due to the same material, with the difference between the two originating from the temperature of the light source. In 1853, the Swedish physicist Anders Jonas Ångström presented observations and theories about gas spectra. Ångström postulated that an incandescent gas emits luminous rays of the same wavelength as those it can absorb. At the same time George Stokes and William Thomson (Kelvin) were discussing similar postulates. Ångström also measured the emission spectrum from hydrogen later labeled the Balmer lines. In 1854 and 1855, David Alter published observations on the spectra of metals and gases, including an independent observation of the Balmer lines of hydrogen. By 1859, Gustav Kirchhoff and Robert Bunsen noticed that several Fraunhofer lines (lines in the solar spectrum) coincide with characteristic emission lines identified in the spectra of heated elements. It was correctly deduced that dark lines in the solar spectrum are caused by absorption by chemical elements in the solar atmosphere. Experimental technique in flame emission spectroscopy The solution containing the relevant substance to be analysed is drawn into the burner and dispersed into the flame as a fine spray. The solvent evaporates first, leaving finely divided solid particles which move to the hottest region of the flame where gaseous atoms and ions are produced through the dissociation of molecules. Here electrons are excited as described above, and the spontaneously emit photon to decay to lower energy states. It is common for a monochromator to be used to allow for easy detection. On a simple level, flame emission spectroscopy can be observed using just a flame and samples of metal salts. This method of qualitative analysis is called a flame test. For example, sodium salts placed in the flame will glow yellow from sodium ions, while strontium (used in road flares) ions color it red. Copper wire will create a blue colored flame, however in the presence of chloride gives green (molecular contribution by CuCl). Emission coefficient Emission coefficient is a coefficient in the power output per unit time of an electromagnetic source, a calculated value in physics. The emission coefficient of a gas varies with the wavelength of the light. It has unit m⋅s−3⋅sr−1. It is also used as a measure of environmental emissions (by mass) per MW⋅h of electricity generated, see: Emission factor. Scattering of light In Thomson scattering a charged particle emits radiation under incident light. The particle may be an ordinary atomic electron, so emission coefficients have practical applications. If is the energy scattered by a volume element into solid angle between wavelengths and per unit time then the emission coefficient is . The values of in Thomson scattering can be predicted from incident flux, the density of the charged particles and their Thomson differential cross section (area/solid angle). Spontaneous emission A warm body emitting photons has a monochromatic emission coefficient relating to its temperature and total power radiation. This is sometimes called the second Einstein coefficient, and can be deduced from quantum mechanical theory.
Physical sciences
Molecular physics
Physics
292444
https://en.wikipedia.org/wiki/Sulfide
Sulfide
Sulfide (also sulphide in British English) is an inorganic anion of sulfur with the chemical formula S2− or a compound containing one or more S2− ions. Solutions of sulfide salts are corrosive. Sulfide also refers to large families of inorganic and organic compounds, e.g. lead sulfide and dimethyl sulfide. Hydrogen sulfide (H2S) and bisulfide (HS−) are the conjugate acids of sulfide. Chemical properties The sulfide ion does not exist in aqueous alkaline solutions of Na2S. Instead sulfide converts to hydrosulfide: S2− + H2O → SH− + OH− Upon treatment with an acid, sulfide salts convert to hydrogen sulfide: S2− + H+ → SH− SH− + H+ → H2S Oxidation of sulfide is a complicated process. Depending on the conditions, the oxidation can produce elemental sulfur, polysulfides, polythionates, sulfite, or sulfate. Metal sulfides react with halogens, forming sulfur and metal salts. 8 MgS + 8 I2 → S8 + 8 MgI2 Metal derivatives Aqueous solutions of transition metals cations react with sulfide sources (H2S, NaHS, Na2S) to precipitate solid sulfides. Such inorganic sulfides typically have very low solubility in water, and many are related to minerals with the same composition (see below). One famous example is the bright yellow species CdS or "cadmium yellow". The black tarnish formed on sterling silver is Ag2S. Such species are sometimes referred to as salts. In fact, the bonding in transition metal sulfides is highly covalent, which gives rise to their semiconductor properties, which in turn is related to the deep colors. Several have practical applications as pigments, in solar cells, and as catalysts. The fungus Aspergillus niger plays a role in the solubilization of heavy metal sulfides. Geology Many important metal ores are sulfides. Significant examples include: argentite (silver sulfide), cinnabar (mercury sulfide), galena (lead sulfide), molybdenite (molybdenum sulfide), pentlandite (nickel sulfide), realgar (arsenic sulfide), and stibnite (antimony sulfide), sphalerite (zinc sulfide), and pyrite (iron disulfide), and chalcopyrite (iron-copper sulfide). This sulfide minerals recorded information (like isotopes) of their surrounding environment during their formation. Scientists use these minerals to study environments in the deep sea or in the Earth's past. Corrosion induced by sulfide Dissolved free sulfides (H2S, HS− and S2−) are very aggressive species for the corrosion of many metals such as steel, stainless steel, and copper. Sulfides present in aqueous solution are responsible for stress corrosion cracking (SCC) of steel, and is also known as sulfide stress cracking. Corrosion is a major concern in many industrial installations processing sulfides: sulfide ore mills, deep oil wells, pipelines transporting soured oil and Kraft paper factories. Microbially-induced corrosion (MIC) or biogenic sulfide corrosion are also caused by sulfate reducing bacteria producing sulfide that is emitted in the air and oxidized in sulfuric acid by sulfur oxidizing bacteria. Biogenic sulfuric acid reacts with sewerage materials and most generally causes mass loss, cracking of the sewer pipes and ultimately, structural collapse. This kind of deterioration is a major process affecting sewer systems worldwide and leading to very high rehabilitation costs. Oxidation of sulfide can also form thiosulfate (), an intermediate species responsible for severe problems of pitting corrosion of steel and stainless steel while the medium is also acidified by the production of sulfuric acid when oxidation is more advanced. Organic chemistry In organic chemistry, "sulfide" usually refers to the linkage C–S–C, although the term thioether is less ambiguous. For example, the thioether dimethyl sulfide is CH3–S–CH3. Polyphenylene sulfide (see below) has the empirical formula C6H4S. Occasionally, the term sulfide refers to molecules containing the –SH functional group. For example, methyl sulfide can mean CH3–SH. The preferred descriptor for such SH-containing compounds is thiol or mercaptan, i.e. methanethiol, or methyl mercaptan. Disulfides Confusion arises from the different meanings of the term "disulfide". Molybdenum disulfide (MoS2) consists of separated sulfide centers, in association with molybdenum in the formal +4 oxidation state (that is, Mo4+ and two S2−). Iron disulfide (pyrite, FeS2) on the other hand consists of , or −S–S− dianion, in association with divalent iron in the formal +2 oxidation state (ferrous ion: Fe2+). Dimethyldisulfide has the chemical binding CH3–S–S–CH3, whereas carbon disulfide has no S–S bond, being S=C=S (linear molecule analog to CO2). Most often in sulfur chemistry and in biochemistry, the disulfide term is commonly ascribed to the sulfur analogue of the peroxide –O–O– bond. The disulfide bond (–S–S–) plays a major role in the conformation of proteins and in the catalytic activity of enzymes. Examples Preparation Sulfide compounds can be prepared in several different ways: Direct combination of elements: Example: Fe() + S() → FeS() Reduction of a sulfate: Example: MgSO4() + 4C() → MgS() + 4CO() Precipitation of an insoluble sulfide: Example: M2+ + H2S() → MS() + 2H+() Safety Many metal sulfides are so insoluble in water that they are probably not very toxic. Some metal sulfides, when exposed to a strong mineral acid, including gastric acids, will release toxic hydrogen sulfide. Organic sulfides are highly flammable. When a sulfide burns it produces sulfur dioxide (SO2) gas. Hydrogen sulfide, some of its salts, and almost all organic sulfides have a strong and putrid stench; rotting biomass releases these. Nomenclature The systematic names sulfanediide and sulfide(2−), valid IUPAC names, are determined according to the substitutive and additive nomenclatures, respectively. The name sulfide is also used in compositional IUPAC nomenclature which does not take the nature of bonding involved. Examples of such naming include selenium disulfide and titanium sulfide, which contain no sulfide ions.
Physical sciences
Sulfide salts
Chemistry
292506
https://en.wikipedia.org/wiki/Cube%20root
Cube root
In mathematics, a cube root of a number is a number that has the given number as its third power; that is The number of cube roots of a number depends on the number system that is considered. Every nonzero real number has exactly one real cube root that is denoted and called the real cube root of or simply the cube root of in contexts where complex numbers are not considered. For example, the real cube roots of and are respectively and . The real cube root of an integer or of a rational number is generally not a rational number, neither a constructible number. Every nonzero real of complex number has exactly three cube roots that are complex numbers. If the number is real, one of the cube roots is real and the two other are nonreal complex conjugate numbers. Otherwise, the three cube roots are all nonreal. For example, the real cube root of is and the other cube roots of are and . The three cube roots of are and The number zero has a unique cube root, which is zero itself. The cube root is a multivalued function. The principal cube root is its principal value, that is a unique cube root that has been chosen once for all. The principal cube root is the cube root with the largest real part. In the case of negative real numbers, the largest real part is shared by the two nonreal cube roots, and the principal cube root is the one with positive imaginary part. So, for negative real numbers, the real cube root is not the principal cube root. For positive real numbers, the principal cube root is the real cube root. If is any cube root of the complex number , the other cube roots are and In an algebraically closed field of characteristic different from three, every nonzero element has exactly three cube roots, which can be obtained from any of them by multiplying it by either root of the polynomial In an algebraically closed field of characteristic three, every element has exactly one cube root. In other number systems or other algebraic structures, a number or element may have more than three cube roots. For example, in the quaternions, a real number has infinitely many cube roots. Definition The cube roots of a number are the numbers which satisfy the equation Properties Real numbers For any real number , there is exactly one real number such that . Indeeed, the cube function is increasing, so does not give the same result for two different inputs, and covers all real numbers. In other words, it is a bijection or one-to-one correspondence. That is, one can define the cube root of a real number as its unique cube root that is also real. With this definition, the cube root of a negative number is a negative number. However this definition may be confusing when real numbers are considered as specific complex numbers, since, in this case the cube root is generally defined as the principal cube root, and the principal cube root of a negative real number is not real. If and are allowed to be complex, then there are three solutions (if is non-zero) and so has three cube roots. A real number has one real cube root and two further cube roots which form a complex conjugate pair. For instance, the cube roots of 1 are: The last two of these roots lead to a relationship between all roots of any real or complex number. If a number is one cube root of a particular real or complex number, the other two cube roots can be found by multiplying that cube root by one or the other of the two complex cube roots of 1. Complex numbers For complex numbers, the principal cube root is usually defined as the cube root that has the greatest real part, or, equivalently, the cube root whose argument has the least absolute value. It is related to the principal value of the natural logarithm by the formula If we write as where is a non-negative real number and lies in the range , then the principal complex cube root is This means that in polar coordinates, we are taking the cube root of the radius and dividing the polar angle by three in order to define a cube root. With this definition, the principal cube root of a negative number is a complex number, and for instance will not be −2, but rather This difficulty can also be solved by considering the cube root as a multivalued function: if we write the original complex number in three equivalent forms, namely The principal complex cube roots of these three forms are then respectively Unless , these three complex numbers are distinct, even though the three representations of x were equivalent. For example, may then be calculated to be −2, , or . This is related with the concept of monodromy: if one follows by continuity the function cube root along a closed path around zero, after a turn the value of the cube root is multiplied (or divided) by Impossibility of compass-and-straightedge construction Cube roots arise in the problem of finding an angle whose measure is one third that of a given angle (angle trisection) and in the problem of finding the edge of a cube whose volume is twice that of a cube with a given edge (doubling the cube). In 1837 Pierre Wantzel proved that neither of these can be done with a compass-and-straightedge construction. Numerical methods Newton's method is an iterative method that can be used to calculate the cube root. For real floating-point numbers this method reduces to the following iterative algorithm to produce successively better approximations of the cube root of : The method is simply averaging three factors chosen such that at each iteration. Halley's method improves upon this with an algorithm that converges more quickly with each iteration, albeit with more work per iteration: This converges cubically, so two iterations do as much work as three iterations of Newton's method. Each iteration of Newton's method costs two multiplications, one addition and one division, assuming that is precomputed, so three iterations plus the precomputation require seven multiplications, three additions, and three divisions. Each iteration of Halley's method requires three multiplications, three additions, and one division, so two iterations cost six multiplications, six additions, and two divisions. Thus, Halley's method has the potential to be faster if one division is more expensive than three additions. With either method a poor initial approximation of can give very poor algorithm performance, and coming up with a good initial approximation is somewhat of a black art. Some implementations manipulate the exponent bits of the floating-point number; i.e. they arrive at an initial approximation by dividing the exponent by 3. Also useful is this generalized continued fraction, based on the nth root method: If is a good first approximation to the cube root of and , then: The second equation combines each pair of fractions from the first into a single fraction, thus doubling the speed of convergence. Appearance in solutions of third and fourth degree equations Cubic equations, which are polynomial equations of the third degree (meaning the highest power of the unknown is 3) can always be solved for their three solutions in terms of cube roots and square roots (although simpler expressions only in terms of square roots exist for all three solutions, if at least one of them is a rational number). If two of the solutions are complex numbers, then all three solution expressions involve the real cube root of a real number, while if all three solutions are real numbers then they may be expressed in terms of the complex cube root of a complex number. Quartic equations can also be solved in terms of cube roots and square roots. History The calculation of cube roots can be traced back to Babylonian mathematicians from as early as 1800 BCE. In the fourth century BCE Plato posed the problem of doubling the cube, which required a compass-and-straightedge construction of the edge of a cube with twice the volume of a given cube; this required the construction, now known to be impossible, of the length . A method for extracting cube roots appears in The Nine Chapters on the Mathematical Art, a Chinese mathematical text compiled around the second century BCE and commented on by Liu Hui in the third century CE. The Greek mathematician Hero of Alexandria devised a method for calculating cube roots in the first century CE. His formula is again mentioned by Eutokios in a commentary on Archimedes. In 499 CE Aryabhata, a mathematician-astronomer from the classical age of Indian mathematics and Indian astronomy, gave a method for finding the cube root of numbers having many digits in the Aryabhatiya (section 2.5).
Mathematics
Specific functions
null
292690
https://en.wikipedia.org/wiki/Arabian%20horse
Arabian horse
The Arabian or Arab horse ( , DMG al-ḥiṣān al-ʿarabī) is a breed of horse with historic roots on the Arabian Peninsula. With a distinctive head shape and high tail carriage, the Arabian is one of the most easily recognizable horse breeds in the world. It is also one of the oldest modern breeds. Although modern DNA cannot trace breed purity in the modern population beyond 200 years, there is archaeological evidence of horses in the Middle East with landrace characteristics that resemble modern Arabians dating back 3,500 years. Throughout history, Arabian horses have spread around the world by both war and trade, used to improve other breeds by adding speed, refinement, endurance, and strong bone. Today, Arabian bloodlines are found in almost every modern breed of riding horse. The Arabian developed in a desert climate and was prized by the nomadic Bedouin people, often being brought inside the family tent for shelter and protection from theft. Selective breeding for traits, including an ability to form a cooperative relationship with humans, created a horse breed that is good-natured, quick to learn, and willing to please. The Arabian also developed the high spirit and alertness needed in a horse used for raiding and war. This combination of willingness and sensitivity requires modern Arabian horse owners to handle their horses with competence and respect. The Arabian is a versatile breed. Arabians dominate the discipline of endurance riding and compete today in many other fields of equestrian sport. They are one of the top ten most popular horse breeds in the world. They are now found worldwide, including the United States and Canada, the United Kingdom, Australia, continental Europe, South America (especially Brazil), and their land of origin, the Middle East. Breed characteristics Arabian horses have refined, wedge-shaped heads, a broad forehead, large eyes, large nostrils, and small muzzles. Most display a distinctive concave, or "dished" profile. Many Arabians also have a slight forehead bulge between their eyes, called the jibbah by the Bedouin, that adds additional sinus capacity, believed to have helped the Arabian horse in its native dry desert climate. Another breed characteristic is an arched neck with a large, well-set windpipe set on a refined, clean throatlatch. This structure of the poll and throatlatch was called the mitbah or mitbeh by the Bedouin. In the ideal Arabian, it is long, allowing flexibility in the bridle and room for the windpipe. Other distinctive features are a relatively long, level croup, or top of the hindquarters, and naturally high tail carriage. The USEF breed standard requires Arabians to have solid bone and standard correct equine conformation. Well-bred Arabians have a deep, well-angled hip and well laid-back shoulder. Within the breed, there are variations. Some individuals have wider, more powerfully muscled hindquarters suitable for intense bursts of activity in events such as reining, while others have longer, leaner muscling better suited for long stretches of flatwork such as endurance riding or horse racing. Most have a compact body with a short back. Arabians usually have dense, strong bone, and good hoof walls. They are especially noted for their endurance, and the superiority of the breed in endurance riding competition demonstrates that well-bred Arabians are strong, sound horses with superior stamina. At international FEI-sponsored endurance events, Arabians and half-Arabians are the dominant performers in distance competition. Skeletal analysis Some Arabians, though not all, have 5 lumbar vertebrae instead of the usual 6, and 17 pairs of ribs rather than 18. A quality Arabian has both a relatively horizontal croup and a properly angled pelvis as well as good croup length and depth to the hip (determined by the length of the pelvis), that allows agility and impulsion. A misconception confuses the topline of the croup with the angle of the "hip" (the pelvis or ilium), leading some to assert that Arabians have a flat pelvis angle and cannot use their hindquarters properly. However, the croup is formed by the sacral vertebrae. The hip angle is determined by the attachment of the ilium to the spine, the structure and length of the femur, and other aspects of hindquarter anatomy, which is not correlated to the topline of the sacrum. Thus, the Arabian has conformation typical of other horse breeds built for speed and distance, such as the Thoroughbred, where the angle of the ilium is more oblique than that of the croup. Thus, the hip angle is not necessarily correlated to the topline of the croup. Horses bred to gallop need a good length of croup and good length of hip for proper attachment of muscles, and so unlike angle, length of hip and croup do go together as a rule. Size The breed standard stated by the United States Equestrian Federation, describes Arabians as standing between tall, "with the occasional individual over or under". Thus, all Arabians, regardless of height, are classified as "horses", even though is the traditional cutoff height between a horse and a pony. A common myth is that Arabians are not strong because they are relatively small and refined. However, the Arabian horse is noted for a greater density of bone than other breeds, short cannons, sound feet, and a broad, short back, all of which give the breed physical strength comparable to many taller animals. Thus, even a smaller Arabian can carry a heavy rider. For tasks where the sheer weight of the horse matters, such as farm work done by a draft horse, any lighter-weight horse is at a disadvantage. However, for most purposes, the Arabian is a strong and hardy light horse breed able to carry any type of rider in most equestrian pursuits. Temperament For centuries, Arabian horses lived in the desert in close association with humans. For shelter and protection from theft, prized war mares were sometimes kept in their owner's tent, close to children and everyday family life. Only horses with a naturally good disposition were allowed to reproduce, with the result that Arabians today have a good temperament that, among other examples, makes them one of the few breeds where the United States Equestrian Federation rules allow children to exhibit stallions in nearly all show ring classes, including those limited to riders under 18. On the other hand, the Arabian is also classified as a "hot-blooded" breed, a category that includes other refined, spirited horses bred for speed, such as the Akhal-Teke, the Barb, and the Thoroughbred. Like other hot-bloods, Arabians' sensitivity and intelligence enable quick learning and greater communication with their riders; however, their intelligence also allows them to learn bad habits as quickly as good ones, and they do not tolerate inept or abusive training practices. Some sources claim that it is more difficult to train a "hot-blooded" horse. Though most Arabians have a natural tendency to cooperate with humans, when treated badly, like any horse, they can become excessively nervous or anxious, but seldom become vicious unless seriously spoiled or subjected to extreme abuse. At the other end of the spectrum, romantic myths are sometimes told about Arabian horses that give them near-divine characteristics. Colors The Arabian Horse Association registers purebred horses with the coat colors bay, gray, chestnut, black, and roan. Bay, gray and chestnut are the most common; black is less common. The classic roan gene does not appear to exist in Arabians; rather, Arabians registered by breeders as "roan" are usually expressing rabicano or, sometimes, sabino patterns with roan features. All Arabians, no matter their coat color, have black skin, except under white markings. Black skin provided protection from the intense desert sun. Gray and white Although many Arabians appear to have a "white" hair coat, they are not genetically "white". This color is usually created by the natural action of the gray gene, and virtually all white-looking Arabians are actually grays. A specialized colorization seen in some older gray Arabians is the so-called "bloody-shoulder", which is a particular type of "flea-bitten" gray with localized aggregations of pigment on the shoulder. There are a very few Arabians registered as "white" defined as having a white coat, pink skin and dark eyes from birth. These animals are believed to manifest a form of dominant white, W3, a result of a nonsense mutation in DNA tracing to a single stallion foaled in 1996. It is possible that white mutations have occurred in Arabians in the past and it is likely that mutations other than W3 exist but have not been verified by genetic testing. Sabino One spotting pattern, sabino, does exist in purebred Arabians. Sabino coloring is characterized by white markings such as "high white" above the knees and hocks, irregular spotting on the legs, belly and face, white markings that extend beyond the eyes or under the chin and jaw, and sometimes lacy or roaned edges. The genetic mechanism that produces sabino patterning in Arabians is now thought to be a form of dominant white, and more than one gene may be involved. However, studies at the University of California, Davis indicate that Arabians do not appear to carry the autosomal dominant gene "SB1" or sabino 1, that often produces bold spotting and some completely white horses in other breeds. The inheritance patterns observed in sabino-like Arabians also do not follow the same mode of inheritance as sabino 1. Rabicano or roan There are very few Arabians registered as roan, and according to researcher D. Phillip Sponenberg, roaning in purebred Arabians is actually the action of rabicano genetics. Unlike a genetic roan, rabicano is a partial roan-like pattern; the horse does not have intermingled white and solid hairs over the entire body, only on the midsection and flanks, the head and legs are solid-colored. Some people also confuse a young gray horse with a roan because of the intermixed hair colors common to both. However, a roan does not consistently lighten with age, while a gray does. Colors that do not exist in purebreds There is pictorial evidence from pottery and tombs in Ancient Egypt suggesting that spotting patterns may have existed on ancestral Arabian-type horses in antiquity. Nonetheless, purebred Arabians today do not carry genes for pinto or Leopard complex ("Appaloosa") spotting patterns, except for sabino. Spotting or excess white was believed by many breeders to be a mark of impurity until DNA testing for verification of parentage became standard. For a time, horses with belly spots and other white markings deemed excessive were discouraged from registration and excess white was sometimes penalized in the show ring. Purebred Arabians never carry dilution genes. Therefore, purebreds cannot be colors such as dun, cremello, palomino or buckskin. To produce horses with some Arabian characteristics but coat colors not found in purebreds, they have to be crossbred with other breeds. Though the purebred Arabian produces a limited range of potential colors, they do not appear to carry any color-based lethal disorders such as the frame overo gene ("O") that can produce lethal white syndrome (LWS). Because purebred Arabians cannot produce LWS foals, Arabian mares were used as a non-affected population in some of the studies seeking the gene that caused the condition in other breeds. Nonetheless, partbred Arabian offspring can, in some cases, carry these genes if the non-Arabian parent was a carrier. Genetic disorders There are six known genetic disorders in Arabian horses. Two are inevitably fatal, two are not inherently fatal but are disabling and usually result in euthanasia of the affected animal; the remaining conditions can usually be treated. Three are thought to be autosomal recessive conditions, which means that the flawed gene is not sex-linked and has to come from both parents for an affected foal to be born. One may be an autosomal dominant. Arabians are not the only breed of horse to have problems with inherited diseases; partbred Arabians may inherit deleterious genetics from other breeds. Genetic diseases that can occur in purebred Arabians, or in partbreds with Arabian ancestry in both parents, are the following: Severe Combined Immunodeficiency (SCID). Recessive disorder, fatal when homozygous, carriers (heterozygotes) show no signs. Similar to the "bubble boy" condition in humans, an affected foal is born with a complete lack of an immune system, and thus generally dies of an opportunistic infection, usually before the age of three months. There is a DNA test that can detect healthy horses who are carriers of the gene causing SCID, thus testing and careful, planned matings can now eliminate the possibility of an affected foal ever being born. Lavender Foal Syndrome (LFS), also called Coat Color Dilution Lethal (CCDL). Recessive disorder, fatal when homozygous, carriers show no signs. The condition has its name because most affected foals are born with a coat color dilution that lightens the tips of the coat hairs, or even the entire hair shaft. Foals with LFS are unable to stand at birth, often have seizures, and are usually euthanized within a few days of birth. In November 2009, Cornell University announced that a DNA test has been developed to detect carriers of LFS. Simultaneously, the University of Pretoria also announced that they had also developed a DNA test. Cerebellar abiotrophy (CA or CCA). Recessive disorder, homozygous horses are affected, carriers show no signs. An affected foal is usually born without clinical signs, but at some stage, usually after six weeks of age, develops severe incoordination, a head tremor, wide-legged stance and other symptoms related to the death of the purkinje cells in the cerebellum. Such foals are frequently diagnosed only after they have crashed into a fence or fallen over backwards, and often are misdiagnosed as suffering from a head injury caused by an accident. Severity varies, with some foals having fast onset of severe coordination problems, others showing milder signs. Mildly affected horses can live a full lifespan, but most are euthanized before adulthood because they are so accident-prone as to be dangerous. As of 2008, there is a genetic test that uses DNA markers associated with CA to detect both carriers and affected animals. Clinical signs are distinguishable from other neurological conditions, and a diagnosis of CA can be verified by examining the brain after euthanasia. Occipital Atlanto-Axial Malformation (OAAM). This is a condition where the occiput, atlas and axis vertebrae in the neck and at the base of the skull are fused or malformed. Symptoms range from mild incoordination to the paralysis of both front and rear legs. Some affected foals cannot stand to nurse, in others the symptoms may not be seen for several weeks. This is the only cervical spinal cord disease seen in horses less than 1 month of age, and a radiograph can diagnose the condition. There is now a genetic test for OAAM. Equine juvenile epilepsy, or Juvenile Idiopathic Epilepsy, sometimes referred to as "benign" epilepsy, is not usually fatal. Foals appear normal between epileptic seizures, and seizures usually stop occurring between 12 and 18 months. Affected foals may show signs of epilepsy anywhere from two days to six months from birth. Seizures can be treated with traditional anti-seizure medications, which may reduce their severity. Though the condition has been studied since 1985 at the University of California, Davis, the genetic mode of inheritance is unclear, though the cases studied were all of one general bloodline group. Recent research updates suggest that a dominant mode of inheritance is involved in transmission of this trait. One researcher hypothesized that epilepsy may be linked in some fashion to Lavender Foal Syndrome due to the fact that it occurs in similar bloodlines and some horses have produced foals with both conditions. Guttural Pouch Tympany (GPT) occurs in horses ranging from birth to 1 year of age and is more common in fillies than in colts. It is thought to be genetic in Arabians, possibly polygenic in inheritance, but more study is needed. Foals are born with a defect that causes the pharyngeal opening of the eustachian tube to act like a one-way valve – air can get in, but it cannot get out. The affected guttural pouch is distended with air and forms a characteristic nonpainful swelling. Breathing is noisy in severely affected animals. Diagnosis is based on clinical signs and radiographic examination of the skull. Medical management with NSAID and antimicrobial therapy can treat upper respiratory tract inflammation. Surgical intervention is needed to correct the malformation of the guttural pouch opening, to provide a route for air in the abnormal guttural pouch to pass to the normal side and be expelled into the pharynx. Foals that are successfully treated may grow up to have fully useful lives. The Arabian Horse Association in the United States has created a foundation that supports research efforts to uncover the roots of genetic diseases. The organization F.O.A.L. (Fight Off Arabian Lethals) is a clearinghouse for information on these conditions. Additional information is available from the World Arabian Horse Association (WAHO). Recent trends in halter breeding have given rise to Arabian horses with extremely concave features, raising concerns that the trait is detrimental to the animal's welfare. Comparisons have been made to a similar trend with some dog breeds, where show judging awarding certain features has led to breeders seeking an ever more exaggerated form, with little concern as to the inherent function of the animal. Some veterinarians speculate that an extremely concave face is detrimental to a horse's breathing, but the issue has not been formally studied. Legends Arabian horses are the topic of many myths and legends. One origin story tells how Muhammad chose his foundation mares by a test of their courage and loyalty. While there are several variants on the tale, a common version states that after a long journey through the desert, Muhammad turned his herd of horses loose to race to an oasis for a desperately needed drink of water. Before the herd reached the water, Muhammad called for the horses to return to him. Only five mares responded. Because they faithfully returned to their master, though desperate with thirst, these mares became his favorites and were called Al Khamsa, meaning, the five. These mares became the legendary founders of the five "strains" of the Arabian horse. Although the Al Khamsa are generally considered fictional horses of legend, some breeders today claim the modern Bedouin Arabian actually descended from these mares. Another origin tale claims that King Solomon was given a pure Arabian-type mare named Safanad ("the pure") by the Queen of Sheba. A different version says that Solomon gave a stallion, Zad el-Raheb or Zad-el-Rakib ("Gift to the Rider"), to the Banu Azd people when they came to pay tribute to the king. This legendary stallion was said to be faster than the zebra and the gazelle, and every hunt with him was successful, thus when he was put to stud, he became a founding sire of legend. Yet another creation myth puts the origin of the Arabian in the time of Ishmael, the son of Abraham. In this story, the Angel Jibril (also known as Gabriel) descended from Heaven and awakened Ishmael with a "wind-spout" that whirled toward him. The Angel then commanded the thundercloud to stop scattering dust and rain, and so it gathered itself into a prancing, handsome creature - a horse - that seemed to swallow up the ground. Hence, the Bedouins bestowed the title "Drinker of the Wind" to the first Arabian horse. Finally, a Bedouin story states that Allah created the Arabian horse from the south wind and exclaimed, "I create thee, Oh Arabian. To thy forelock, I bind Victory in battle. On thy back, I set a rich spoil and a Treasure in thy loins. I establish thee as one of the Glories of the Earth... I give thee flight without wings." Other versions of the story claim Allah said to the South Wind: "I want to make a creature out of you. Condense." Then from the material condensed from the wind, he made a kamayt-colored animal (a bay or burnt chestnut) and said: "I call you Horse; I make you Arabian and I give you the chestnut color of the ant; I have hung happiness from the forelock which hangs between your eyes; you shall be the Lord of the other animals. Men shall follow you wherever you go; you shall be as good for flight as for pursuit; you shall fly without wings; riches shall be on your back and fortune shall come through your meditation." Origins Arabians are thought to be one of the oldest human-developed horse breeds in the world. Recent genetic studies of mitochondrial DNA in Arabian horses of Polish and American breeding suggest that the modern breed has heterogeneous origins with ten haplogroups. The modern concept of breed purity in the modern population cannot be traced beyond 200 years. The progenitor stock, the Oriental subtype, had characteristics similar to the modern Arabian. Horses with these features appeared in rock paintings and inscriptions in the Arabian Peninsula dating back 3500 years. In ancient history throughout the Ancient Near East, horses with refined heads and high-carried tails were depicted in artwork, particularly that of Ancient Egypt in the 16th century BC. Some 20th-century scholars of the Arabian horse once theorized that the Arabian came from a separate subspecies of horse, known as equus caballus pumpelli. However, others, including Gladys Brown Edwards, a noted Arabian researcher, stated that the "dry" oriental horses of the desert, from which the modern Arabian developed, were Equus ferus caballus with specific landrace characteristics based on the environments in which they lived, rather than being a separate subspecies. Horses with similar, though not identical, physical characteristics include the Marwari horse of India, the Barb of North Africa, the Akhal-Teke of western Asia and the now-extinct Turkoman Horse. Desert roots There are different theories about where the ancestors of the Arabian originally lived. Most evidence suggests the proto-Arabian came from the area along the northern edge of the Fertile Crescent. Another hypothesis suggests the southwestern corner of the Arabian peninsula, in modern-day Yemen, where three now-dry riverbeds indicate good natural pastures existed long ago, perhaps as far back as the Ice Age. This hypothesis has gained renewed attention following a 2010 discovery of artifacts dated between 6590 and 7250 BCE in Al-Magar, in southwestern Saudi Arabia, that appeared to portray horses. The proto-Arabian horse may have been domesticated by the people of the Arabian Peninsula known today as the Bedouin, some time after they learned to use the camel, approximately 4,000–5,000 years ago. One theory is that this development occurred in the Nejd plateau in central Arabia. Other scholars, noting that horses were common in the Fertile Crescent but rare in the Arabian peninsula prior to the rise of Islam, theorize that the breed as it is known today only developed in large numbers when the conversion of the Persians to Islam in the 7th century brought knowledge of horse breeding and horsemanship to the Bedouin. The oldest depictions in the Arabian Peninsula of horses that are clearly domesticated date no earlier than 1800-2000 BCE. Regardless of origin, climate and culture ultimately created the Arabian. The desert environment required a domesticated horse to cooperate with humans to survive; humans were the only providers of food and water in certain areas, and even hardy Arabian horses needed far more water than camels in order to survive (most horses can only live about 72 hours without water). Where there was no pasture or water, the Bedouin fed their horses dates and camel's milk. The desert horse needed the ability to thrive on very little food, and to have anatomical traits to compensate for life in a dry climate with wide temperature extremes from day to night. Weak individuals were weeded out of the breeding pool, and the animals that remained were also honed by centuries of human warfare. The Bedouin way of life depended on camels and horses: Arabians were bred to be war horses with speed, endurance, soundness, and intelligence. Because many raids required stealth, mares were preferred over stallions as they were quieter, and therefore would not give away the position of the fighters. A good disposition was also critical; prized war mares were often brought inside family tents to prevent theft and for protection from weather and predators. Though appearance was not necessarily a survival factor, the Bedouin bred for refinement and beauty in their horses as well as for more practical features. Strains and pedigrees For centuries, the Bedouin tracked the ancestry of each horse through an oral tradition. Horses of the purest blood were known as Asil and crossbreeding with non-Asil horses was forbidden. Mares were the most valued, both for riding and breeding, and pedigree families were traced through the female line. The Bedouin did not believe in gelding male horses, and considered stallions too intractable to be good war horses, thus they kept very few colts, selling most, and culling those of poor quality. Over time, the Bedouin developed several sub-types or strains of Arabian horse, each with unique characteristics, and traced through the maternal line only. According to the Arabian Horse Association, the five primary strains were known as the Keheilan, Seglawi, Abeyan, Hamdani and Hadban. Carl Raswan, a promoter and writer about Arabian horses from the middle of the 20th century, held the belief that there were only three strains, Kehilan, Seglawi and Muniqi. Raswan felt that these strains represented body "types" of the breed, with the Kehilan being "masculine", the Seglawi being "feminine" and the Muniqi being "speedy". There were also lesser strains, sub-strains, and regional variations in strain names. Therefore, many Arabian horses were not only Asil, of pure blood, but also bred to be pure in strain, with crossbreeding between strains discouraged, though not forbidden, by some tribes. Purity of bloodline was very important to the Bedouin, and they also believed in telegony, believing if a mare was ever bred to a stallion of "impure" blood, the mare herself and all future offspring would be "contaminated" by the stallion and hence no longer Asil. This complex web of bloodline and strain was an integral part of Bedouin culture; they not only knew the pedigrees and history of their best war mares in detail, but also carefully tracked the breeding of their camels, Saluki dogs, and their own family or tribal history. Eventually, written records began to be kept; the first written pedigrees in the Middle East that specifically used the term "Arabian" date to 1330 AD. As important as strain was to the Bedouin, modern studies of mitochondrial DNA suggest that Arabian horses alive today with records stating descent from a given strain may not actually share a common maternal ancestry. Historic development Role in the ancient world Fiery war horses with dished faces and high-carried tails were popular artistic subjects in Ancient Egypt and Mesopotamia, often depicted pulling chariots in war or for hunting. Horses with oriental characteristics appear in later artwork as far north as that of Ancient Greece and the Roman Empire. While this type of horse was not called an "Arabian" in the Ancient Near East until later, these proto-Arabians shared many characteristics with the modern Arabian, including speed, endurance, and refinement. For example, a horse skeleton unearthed in the Sinai peninsula, dated to 1700 BC and probably brought by the Hyksos invaders, is considered the earliest physical evidence of the horse in Ancient Egypt. This horse had a wedge-shaped head, large eye sockets and small muzzle, all characteristics of the Arabian horse. In Islamic history Following the Hijra in AD 622 (also sometimes spelled Hegira), the Arabian horse spread across the known world of the time, and became recognized as a distinct, named breed. It played a significant role in the History of the Middle East and of Islam. By 630, Muslim influence expanded across the Middle East and North Africa, by 711 Muslim warriors had reached Spain, and they controlled most of the Iberian Peninsula by 720. Their war horses were of various oriental types, including both Arabians and the Barb horse of North Africa. More Arabian horses were introduced to North Africa as a result of the migration of Banu Hilal. Arabian horses also spread to the rest of the world via the Ottoman Empire, which rose in 1299. Though it never fully dominated the heart of the Arabian Peninsula, this Turkish empire obtained many Arabian horses through trade, diplomacy and war. The Ottomans encouraged formation of private stud farms in order to ensure a supply of cavalry horses, and Ottoman nobles, such as Muhammad Ali of Egypt also collected pure, desert-bred Arabian horses. El Naseri, or Al-Nasir Muhammad, Sultan of Egypt (1290–1342) imported and bred numerous Arabians in Egypt. A stud farm record was made of his purchases describing many of the horses as well as their abilities, and was deposited in his library, becoming a source for later study. Through the Ottomans, Arabian horses were often sold, traded, or given as diplomatic gifts to Europeans and, later, to Americans. Egypt Historically, Egyptian breeders imported horses bred in the deserts of Palestine and the Arabian peninsula as the source of their foundation bloodstock. By the time that the Ottoman Empire dominated Egypt, the political elites of the region still recognized the need for quality bloodstock for both war and for horse racing, and some continued to return to the deserts to obtain pure-blooded Arabians. One of the most famous was Muhammad Ali of Egypt, also known as Muhammad Ali Pasha, who established an extensive stud farm in the 19th century. After his death, some of his stock was bred on by Abbas I of Egypt, also known as Abbas Pasha. However, after Abbas Pasha was assassinated in 1854, his heir, El Hami Pasha, sold most of his horses, often for crossbreeding, and gave away many others as diplomatic gifts. A remnant of the herd was obtained by Ali Pasha Sherif, who then went back to the desert to bring in new bloodstock. At its peak, the stud of Ali Pasha Sherif had over 400 purebred Arabians. Unfortunately, an epidemic of African horse sickness in the 1870s that killed thousands of horses throughout Egypt decimated much of his herd, wiping out several irreplaceable bloodlines. Late in his life, he sold several horses to Wilfred and Lady Anne Blunt, who exported them to Crabbet Park Stud in England. After his death, Lady Anne was also able to gather many remaining horses at her Sheykh Obeyd stud. Meanwhile, the passion brought by the Blunts to saving the pure horse of the desert helped Egyptian horse breeders to convince their government of the need to preserve the best of their own remaining pure Arabian bloodstock that descended from the horses collected over the previous century by Muhammad Ali Pasha, Abbas Pasha and Ali Pasha Sherif. The government of Egypt formed the Royal Agricultural Society (RAS) in 1908, which is known today as the Egyptian Agricultural Organization (EAO). RAS representatives traveled to England during the 1920s and purchased eighteen descendants of the original Blunt exports from Lady Wentworth at Crabbet Park, and brought them to Egypt in order to restore bloodlines had been lost. Other than several horses purchased by Henry Babson for importation to the United States in the 1930s, and one other small group exported to the US in 1947, relatively few Egyptian-bred Arabian horses were exported until the overthrow of King Farouk I in 1952. Many of the private stud farms of the princes were then confiscated and the animals taken over by the EAO. In the 1960s and 1970s, as oil development brought more foreign investors to Egypt, some of whom were horse fanciers, Arabians were exported to Germany and to the United States, as well as to the former Soviet Union. Today, the designation "Straight Egyptian" or "Egyptian Arabian" is popular with some Arabian breeders, and the modern Egyptian-bred Arabian is an outcross used to add refinement in some breeding programs. Arrival in Europe Probably the earliest horses with Arabian bloodlines to enter Europe came indirectly, through Spain and France. Others would have arrived with returning Crusaders—beginning in 1095, European armies invaded Palestine and many knights returned home with Arabian horses as spoils of war. Later, as knights and the heavy, armored war horses who carried them became obsolete, Arabian horses and their descendants were used to develop faster, agile light cavalry horses that were used in warfare into the 20th century. Another major infusion of Arabian horses into Europe occurred when the Ottoman Turks sent 300,000 horsemen into Hungary in 1522, many of whom were mounted on pure-blooded Arabians, captured during raids into Arabia. By 1529, the Ottomans reached Vienna, where they were stopped by the Polish and Hungarian armies, who captured these horses from the defeated Ottoman cavalry. Some of these animals provided foundation bloodstock for the major studs of eastern Europe. Polish and Russian breeding programs With the rise of light cavalry, the stamina and agility of horses with Arabian blood gave an enormous military advantage to any army who possessed them. As a result, many European monarchs began to support large breeding establishments that crossed Arabians on local stock, one example being Knyszyna, the royal stud of Polish king Zygmunt II August, and another the Imperial Russian Stud of Peter the Great. European horse breeders also obtained Arabian stock directly from the desert or via trade with the Ottomans. In Russia, Count Alexey Orlov obtained many Arabians, including Smetanka, an Arabian stallion who became a foundation sire of the Orlov trotter. Orlov then provided Arabian horses to Catherine the Great, who in 1772 owned 12 pure Arabian stallions and 10 mares. By 1889 two members of the Russian nobility, Count Stroganov and Prince Nikolai Borisovich Shcherbatov, established Arabian stud farms to meet the continued need to breed Arabians as a source of pure bloodstock. In Poland, notable imports from Arabia included those of Prince Hieronymous Sanguszko (1743–1812), who founded the Slawuta stud. Poland's first state-run Arabian stud farm, Janów Podlaski, was established by the decree of Alexander I of Russia in 1817, and by 1850, the great stud farms of Poland were well-established, including Antoniny, owned by the Polish Count Potocki (who had married into the Sanguszko family); later notable as the farm that produced the stallion Skowronek. Central and western Europe The 18th century marked the establishment of most of the great Arabian studs of Europe, dedicated to preserving "pure" Arabian bloodstock. The Prussians set up a royal stud in 1732, originally intended to provide horses for the royal stables, and other studs were established to breed animals for other uses, including mounts for the Prussian army. The foundation of these breeding programs was the crossing of Arabians on native horses; by 1873 some English observers felt that the Prussian cavalry mounts were superior in endurance to those of the British, and credited Arabian bloodlines for this superiority. Other state studs included the Babolna Stud of Hungary, set up in 1789, and the Weil stud in Germany (now Weil-Marbach or the Marbach stud), founded in 1817 by King William I of Württemberg. King James I of England imported the first Arabian stallion, the Markham Arabian, to England in 1616. Arabians were also introduced into European race horse breeding, especially in England via the Darley Arabian, Byerly Turk, and Godolphin Arabian, the three foundation stallions of the modern Thoroughbred breed, who were each brought to England during the 18th century. Other monarchs obtained Arabian horses, often as personal mounts. One of the most famous Arabian stallions in Europe was Marengo, the war horse ridden by Napoleon Bonaparte. During the mid-19th century, the need for Arabian blood to improve the breeding stock for light cavalry horses in Europe resulted in more excursions to the Middle East. Queen Isabel II of Spain sent representatives to the desert to purchase Arabian horses and by 1847 had established a stud book; her successor, King Alfonso XII imported additional bloodstock from other European nations. By 1893, the state military stud farm, Yeguada Militar was established in Córdoba, Spain for breeding both Arabian and Iberian horses. The military remained heavily involved in the importation and breeding of Arabians in Spain well into the early 20th century, and the Yeguada Militar is still in existence today. This period also marked a phase of considerable travel to the Middle East by European civilians and minor nobility, and in the process, some travelers noticed that the Arabian horse as a pure breed of horse was under threat due to modern forms of warfare, inbreeding and other problems that were reducing the horse population of the Bedouin tribes at a rapid rate. By the late 19th century, the most farsighted began in earnest to collect the finest Arabian horses they could find in order to preserve the blood of the pure desert horse for future generations. The most famous example was Lady Anne Blunt, the daughter of Ada Lovelace and granddaughter of Lord Byron. Rise of the Crabbet Park Stud Perhaps the most famous of all Arabian breeding operations founded in Europe was the Crabbet Park Stud of England, founded 1878. Starting in 1877, Wilfrid Scawen Blunt and Lady Anne Blunt made repeated journeys to the Middle East, including visits to the stud of Ali Pasha Sherif in Egypt and to Bedouin tribes in the Nejd, bringing the best Arabians they could find to England. Lady Anne also purchased and maintained the Sheykh Obeyd stud farm in Egypt, near Cairo. Upon Lady Anne's death in 1917, the Blunts' daughter, Judith, Lady Wentworth, inherited the Wentworth title and Lady Anne's portion of the estate, and obtained the remainder of the Crabbet Stud following a protracted legal battle with her father. Lady Wentworth expanded the stud, added new bloodstock, and exported Arabian horses worldwide. Upon her death in 1957, the stud passed to her manager, Cecil Covey, who ran Crabbet until 1971, when a motorway was cut through the property, forcing the sale of the land and dispersal of the horses. Along with Crabbet, the Hanstead Stud of Lady Yule also produced horses of worldwide significance. Early 20th-century Europe In the early 20th century, the military was involved in the breeding of Arabian horses throughout Europe, particularly in Poland, Spain, Germany, and Russia; private breeders also developed a number of breeding programs. Significant among the private breeders in continental Europe was Spain's Cristóbal Colón de Aguilera, XV Duque de Veragua, a direct descendant of Christopher Columbus, who founded the Veragua Stud in the 1920s. Modern warfare and its impact on European studs Between World War I, the Russian Revolution, and the collapse of the Ottoman Empire, many historic European stud farms were lost; in Poland, the Antoniny and Slawuta Studs were wiped out except for five mares. Notable among the survivors was the Janów Podlaski Stud Farm. The Russian Revolution, combined with the effects of World War I, destroyed most of the breeding programs in Russia, but by 1921, the Soviet government reestablished an Arabian program, the Tersk Stud, on the site of the former Stroganov estate, which included Polish bloodstock as well as some importations from the Crabbet Stud in England. The programs that survived the war re-established their breeding operations and some added to their studs with new imports of desert-bred Arabian horses from the Middle East. Not all European studs recovered. The Weil stud of Germany, founded by King Wilhelm I, went into considerable decline; by the time the Weil herd was transferred to the Marbach State Stud in 1932, only 17 purebred Arabians remained. The Spanish Civil War and World War II also had a devastating impact on horse breeding throughout Europe. The Veragua stud was destroyed, and its records lost, with the only survivors being the broodmares and the younger horses, who were rescued by Francisco Franco. Crabbet Park, Tersk, and Janów Podlaski survived. Both the Soviet Union and the United States obtained valuable Arabian bloodlines as spoils of war, which they used to strengthen their breeding programs. The Soviets had taken steps to protect their breeding stock at Tersk Stud, and by utilizing horses captured in Poland they were able to re-establish their breeding program soon after the end of World War II. The Americans brought Arabian horses captured in Europe to the United States, mostly to the Pomona U.S. Army Remount station, the former W.K. Kellogg Ranch in California. In the postwar era, Poland, Spain, and Germany developed or re-established many well-respected Arabian stud farms. The studs of Poland in particular were decimated by both the Nazis and the Soviets, but were able to reclaim some of their breeding stock and became particularly world-renowned for their quality Arabian horses, tested rigorously by racing and other performance standards. During the 1950s, the Russians also obtained additional horses from Egypt to augment their breeding programs. After the Cold War While only a few Arabians were exported from behind the Iron Curtain during the Cold War, those who did come to the west caught the eye of breeders worldwide. Improved international relations between eastern Europe and the west led to major imports of Polish and Russian-bred Arabian horses to western Europe and the United States in the 1970s and 1980s. The collapse of the former Soviet Union in 1991, greater political stability in Egypt, and the rise of the European Union all increased international trade in Arabian horses. Organizations such as the World Arabian Horse Association (WAHO) created consistent standards for transferring the registration of Arabian horses between different nations. Today, Arabian horses are traded all over the world. In America The first horses on the American mainland since the end of the Ice Age arrived with the Spanish Conquistadors. Hernán Cortés brought 16 horses of Andalusian, Barb, and Arabian ancestry to Mexico in 1519. Others followed, such as Francisco Vásquez de Coronado, who brought 250 horses of similar breeding to America in 1540. More horses followed with each new arrival of Conquistadors, missionaries, and settlers. Many horses escaped or were stolen, becoming the foundation stock of the American Mustang. Early imports Colonists from England also brought horses of Arabian breeding to the eastern seaboard. One example was Nathaniel Harrison, who imported a horse of Arabian, Barb and Turkish ancestry to America in 1747. One of George Washington's primary mounts during the American Revolutionary War was a gray half-Arabian horse named Blueskin, sired by the stallion "Ranger", also known as "Lindsay's Arabian", said to have been obtained from the Sultan of Morocco. Other Presidents are linked to ownership of Arabian horses; in 1840, President Martin Van Buren received two Arabians from the Sultan of Oman, and in 1877, President Ulysses S. Grant obtained an Arabian stallion, Leopard, and a Barb, Linden Tree, as gifts from Abdul Hamid II, the "Sultan of Turkey". A. Keene Richard was the first American known to have specifically bred Arabian horses. He traveled to the desert in 1853 and 1856 to obtain breeding stock, which he crossed on Thoroughbreds, and also bred purebred Arabians. Unfortunately, his horses were lost during the Civil War and have no known purebred Arabian descendants today. Another major U.S. political figure, William H. Seward purchased four Arabians in Beirut in 1859, prior to becoming Secretary of State to Abraham Lincoln. Leopard is the only stallion imported prior to 1888 who left known purebred descendants in America. In 1888 Randolph Huntington imported the desert-bred Arabian mare *Naomi, and bred her to Leopard, producing Leopard's only purebred Arabian son, Anazeh, who sired eight purebred Arabian foals, four of whom still appear in pedigrees today. Development of purebred breeding in America In 1908, the Arabian Horse Registry of America was established, recording 71 animals, and by 1994, the number had reached half a million. Today there are more Arabians registered in North America than in the rest of the world put together. The origins of the registry date to 1893, when the Hamidie Society sponsored an exhibit of Arabian horses from what today is Syria at the World Fair in Chicago. This exhibition raised considerable interest in Arabian horses. Records are unclear if 40 or 45 horses were imported for the exposition, but seven died in a fire shortly after arrival. The 28 horses that remained at the end of the exhibition stayed in America and were sold at auction when the Hamidie Society went bankrupt. These horses caught the interest of American breeders, including Peter Bradley of the Hingham Stock Farm, who purchased some Hamidie horses at the auction, and Homer Davenport, another admirer of the Hamidie imports. Major Arabian importations to the United States included those of Davenport and Bradley, who teamed up to purchase several stallions and mares directly from the Bedouin in 1906. Spencer Borden of the Interlachen Stud made several importations between 1898 and 1911; and W.R. Brown of the Maynesboro Stud, interested in the Arabian as a cavalry mount, imported many Arabians over a period of years, starting in 1918. Another wave of imports came in the 1920s and 30s when breeders such as W.K. Kellogg, Henry Babson, Roger Selby, James Draper, and others imported Arabian bloodstock from Crabbet Park Stud in England, as well as from Poland, Spain and Egypt. The breeding of Arabians was fostered by the U. S. Army Remount Service, which stood purebred stallions at public stud for a reduced rate. Several Arabians, mostly of Polish breeding, were captured from Nazi Germany and imported to the U.S.A. following World War II. In 1957, two deaths in England led to more sales to the United States: first from Crabbet Stud on the demise of Lady Wentworth, and then from Hanstead with the passing of Gladys Yule. As the tensions of the Cold War eased, more Arabians were imported to America from Poland and Egypt, and in the late 1970s, as political issues surrounding import regulations and the recognition of stud books were resolved, many Arabian horses were imported from Spain and Russia. Modern trends In the 1980s, Arabians became a popular status symbol and were marketed similarly to fine art. Some individuals also used horses as a tax shelter. Prices skyrocketed, especially in the United States, with a record-setting public auction price for a mare named NH Love Potion, who sold for $2.55 million in 1984, and the largest syndication in history for an Arabian stallion, Padron, at $11 million. The potential for profit led to over-breeding of the Arabian. When the Tax Reform Act of 1986 closed the tax-sheltering "passive investment" loophole, limiting the use of horse farms as tax shelters, the Arabian market was particularly vulnerable due to over-saturation and artificially inflated prices, and it collapsed, forcing many breeders into bankruptcy and sending many purebred Arabians to slaughter. Prices recovered slowly, with many breeders moving away from producing "living art" and towards a horse more suitable for amateur owners and many riding disciplines. By 2003, a survey found that 67% of purebred Arabian horses in America are owned for recreational riding purposes. , there are more than 660,000 Arabians that have been registered in the United States, and the US has the largest number of Arabians of any nation in the world. In Australia Early imports Arabian horses were introduced to Australia in the earliest days of European Settlement. Early imports included both purebred Arabians and light Spanish "jennets" from Andalusia, many Arabians also came from India. Based on records describing stallions "of Arabic and Persian blood", the first Arabian horses were probably imported to Australia in several groups between 1788 and 1802. About 1803, a merchant named Robert Campbell imported a bay Arabian stallion, Hector, from India; Hector was said to have been owned by Arthur Wellesley, who later became known as the Duke of Wellington. In 1804 two additional Arabians, also from India, arrived in Tasmania one of whom, White William, sired the first purebred Arabian foal born in Australia, a stallion named Derwent. Throughout the 19th century, many more Arabians came to Australia, though most were used to produce crossbred horses and left no recorded purebred descendants. The first significant imports to be permanently recorded with offspring still appearing in modern purebred Arabian pedigrees were those of James Boucaut, who in 1891 imported several Arabians from Wilfred and Lady Anne Blunt's Crabbet Arabian Stud in England. Purebred Arabians were used to improve racehorses and some of them became quite famous as such; about 100 Arabian sires are included in the Australian Stud Book (for Thoroughbred racehorses). The military was also involved in the promotion of breeding cavalry horses, especially around World War I. They were part of the foundation of several breeds considered uniquely Australian, including the Australian Pony, the Waler and the Australian Stock Horse. In the 20th and 21st centuries In the early 20th century, more Arabian horses, mostly of Crabbet bloodlines, arrived in Australia. The first Arabians of Polish breeding arrived in 1966, and Egyptian lines were first imported in 1970. Arabian horses from the rest of the world followed, and today the Australian Arabian horse registry is the second largest in the world, next to that of the United States. Modern breeding Arabian horses today are found all over the world. They are no longer classified by Bedouin strain, but are informally classified by the nation of origin of famed horses in a given pedigree. Popular types of Arabians are labeled "Polish", "Spanish", "Crabbet", "Russian", "Egyptian", and "Domestic" (describing horses whose ancestors were imported to the United States prior to 1944, including those from programs such as Kellogg, Davenport, Maynesboro, Babson, Dickenson and Selby). In the US, a specific mixture of Crabbet, Maynesboro and Kellogg bloodlines has acquired the copyrighted designation "CMK". Each set of bloodlines has its own devoted followers, with the virtues of each hotly debated. Most debates are between those who value the Arabian most for its refined beauty and those who value the horse for its stamina and athleticism; there are also a number of breeders who specialize in preservation breeding of various bloodlines. Controversies exist over the relative "purity" of certain animals; breeders argue about the genetic "purity" of various pedigrees, discussing whether some horses descend from "impure" animals that cannot be traced to the desert Bedouin. The major factions are as follows: The Arabian Horse Association (AHA) states, "The origin of the purebred Arabian horse was the Arabian desert, and all Arabians ultimately trace their lineage to this source." In essence, all horses accepted for registration in the United States are deemed to be "purebred" Arabians by AHA. The World Arabian Horse Association (WAHO) has the broadest definition of a purebred Arabian. WAHO states, "A Purebred Arabian horse is one which appears in any purebred Arabian Stud Book or Register listed by WAHO as acceptable." By this definition, over 95% of the known purebred Arabian horses in the world are registered in stud books acceptable to WAHO. WAHO also researched the purity question in general, and its findings are on its web site, describing both the research and the political issues surrounding Arabian horse bloodlines, particularly in America. At the other end of the spectrum, organizations focused on bloodlines that are the most meticulously documented to desert sources have the most restrictive definitions. For example, The Asil Club in Europe only accepts "a horse whose pedigree is exclusively based on Bedouin breeding of the Arabian peninsula, without any crossbreeding with non-Arabian horses at any time". Likewise, the Al Khamsa organization takes the position that "The horse...which are called "Al Khamsa Arabian Horses", are those horses in North America that can reasonably be assumed to descend entirely from bedouin Arabian horses bred by horse-breeding bedouin tribes of the deserts of the Arabian Peninsula without admixture from sources unacceptable to Al Khamsa." Most restrictive of all are horses identified as "straight Egyptian" by the Pyramid Society, which must trace in all lines to the desert and also to horses owned or bred by specific Egyptian breeding programs. By this definition, straight Egyptian Arabians constitute only 2% of all Arabian horses in America. Ironically, some pure-blooded desert-bred Arabians in Syria had enormous difficulties being accepted as registrable purebred Arabians because many of the Bedouin who owned them saw no need to obtain a piece of paper to verify the purity of their horses. However, eventually the Syrians developed a stud book for their animals that was accepted by the World Arabian Horse Association (WAHO) in 2007. Influence on other horse breeds Because of the genetic strength of the desert-bred Arabian horse, Arabian bloodlines have played a part in the development of nearly every modern light horse breed, including the Thoroughbred, Orlov Trotter, Morgan, American Saddlebred, American Quarter Horse, and Warmblood breeds such as the Trakehner. Arabian bloodlines have also influenced the development of the Welsh Pony, the Australian Stock Horse, Percheron draft horse, Appaloosa, and the Colorado Ranger Horse. Today, people cross Arabians with other breeds to add refinement, endurance, agility and beauty. In the US, Half-Arabians have their own registry within the Arabian Horse Association, which includes a special section for Anglo-Arabians (Arabian-Thoroughbred crosses). Some crosses originally registered only as Half-Arabians became popular enough to have their own breed registry, including the National Show Horse (an Arabian-Saddlebred cross), the Quarab (Arabian-Quarter Horse), the Pintabian the Welara (Arabian-Welsh Pony), and the Morab (Arabian-Morgan). In addition, some Arabians and Half Arabians have been approved for breeding by some Warmblood registries, particularly the Trakehner registry. There is intense debate over the role the Arabian played in the development of other light horse breeds. Before DNA-based research developed, one hypothesis, based on body types and conformation, suggested the light, "dry", oriental horse adapted to the desert climate had developed prior to domestication; DNA studies of multiple horse breeds now suggest that while domesticated horses arose from multiple mare lines, there is very little variability in the Y-chromosome between breeds. Following domestication of the horse, due to the location of the Middle East as a crossroads of the ancient world, and relatively near the earliest locations of domestication, oriental horses spread throughout Europe and Asia both in ancient and modern times. There is little doubt that humans crossed "oriental" blood on that of other types to create light riding horses; the only actual questions are at what point the "oriental" prototype could be called an "Arabian", how much Arabian blood was mixed with local animals, and at what point in history. For some breeds, such as the Thoroughbred, Arabian influence of specific animals is documented in written stud books. For older breeds, dating the influx of Arabian ancestry is more difficult. For example, while outside cultures, and the horses they brought with them, influenced the predecessor to the Iberian horse in both the time of Ancient Rome and again with the Islamic invasions of the 8th century, it is difficult to trace precise details of the journeys taken by waves of conquerors and their horses as they traveled from the Middle East to North Africa and across Gibraltar to southern Europe. Mitochondrial DNA studies of modern Andalusian horses of the Iberian peninsula and Barb horses of North Africa present convincing evidence that both breeds crossed the Strait of Gibraltar and influenced one another. Though these studies did not compare Andalusian and Barb mtDNA to that of Arabian horses, there is evidence that horses resembling Arabians, whether before or after the breed was called an "Arabian", were part of this genetic mix. Arabians and Barbs, though probably related to one another, are quite different in appearance, and horses of both Arabian and Barb type were present in the Muslim armies that occupied Europe. There is also historical documentation that Islamic invaders raised Arabian horses in Spain prior to the Reconquista; the Spanish also documented imports of Arabian horses in 1847, 1884 and 1885 that were used to improve existing Spanish stock and revive declining equine populations. Uses Arabians are versatile horses that compete in many equestrian fields, including horse racing, the horse show disciplines of saddle seat, Western pleasure, and hunt seat, as well as dressage, cutting, reining, endurance riding, show jumping, eventing, youth events such as equitation, and others. They are used as pleasure riding, trail riding, and working ranch horses for those who are not interested in competition. Competition Arabians dominate the sport of endurance riding because of their stamina. They are the leading breed in competitions such as the Tevis Cup that can cover up to in a day, and they participate in FEI-sanctioned endurance events worldwide, including the World Equestrian Games. There is an extensive series of horse shows in the United States and Canada for Arabian, Half-Arabian, and Anglo-Arabian horses, sanctioned by the USEF in conjunction with the Arabian Horse Association. Classes offered include Western pleasure, reining, hunter type and saddle seat English pleasure, and halter, plus the very popular "Native" costume class. "Sport horse" events for Arabian horses have become popular in North America, particularly after the Arabian Horse Association began hosting a separate Arabian and Half Arabian Sport Horse National Championship in 2003 that by 2004 grew to draw 2000 entries. This competition draws Arabian and part-Arabian horses that perform in hunter, jumper, sport horse under saddle, sport horse in hand, dressage, and combined driving competition. Other nations also sponsor major shows strictly for purebred and partbred Arabians, including Great Britain France, Spain, Poland, and the United Arab Emirates. Purebred Arabians have excelled in open events against other breeds. One of the most famous examples in the field of western riding competition was the Arabian mare Ronteza, who defeated 50 horses of all breeds to win the 1961 Reined Cow Horse championship at the Cow Palace in San Francisco, California. Another Arabian competitive against all breeds was the stallion Aaraf who won an all-breed cutting horse competition at the Quarter Horse Congress in the 1950s. In show jumping and show hunter competition, a number of Arabians have competed successfully against other breeds in open competition, including the purebred gelding Russian Roulette, who has won multiple jumping classes against horses of all breeds on the open circuit, and in eventing, a purebred Arabian competed on the Brazilian team at the 2004 Athens Olympics. Part-Arabians have also appeared at open sport horse events and even Olympic level competition. The Anglo-Arabian Linon was ridden to an Olympic silver medal for France in Dressage in 1928 and 1932, as well as a team gold in 1932, and another French Anglo-Arabian, Harpagon, was ridden to a team gold medal and an individual silver in dressage at the 1948 Olympics. At the 1952 Olympics, the French rider Pierre d'Oriola won the Gold individual medal in show jumping on the Anglo-Arabian Ali Baba. Another Anglo-Arabian, Tamarillo, ridden by William Fox-Pitt, represents the United Kingdom in FEI and Olympic competition, winning many awards, including first place at the 2004 Badminton Horse Trials. More recently a gelding named Theodore O'Connor, nicknamed "Teddy", a 14.1 (or 14.2, sources vary) hand pony of Thoroughbred, Arabian, and Shetland pony breeding, won two gold medals at the 2007 Pan American Games and was finished in the top six at the 2007 and 2008 Rolex Kentucky Three Day CCI competition. Other activities Arabians are involved in a wide variety of activities, including fairs, movies, parades, circuses and other places where horses are showcased. They have been popular in movies, dating back to the silent film era when Rudolph Valentino rode the Kellogg Arabian stallion Jadaan in 1926's Son of the Sheik, and have been seen in many other films, including The Black Stallion featuring the stallion Cass Ole, The Young Black Stallion, which used over 40 Arabians during filming, as well as Hidalgo and the 1959 version of Ben-Hur. Arabians are mascots for football teams, performing crowd-pleasing activities on the field and sidelines. One of the horses who serves as "Traveler", the mascot for the University of Southern California Trojans, has been a purebred Arabian. "Thunder", a stage name for the purebred Arabian stallion J B Kobask, was mascot for the Denver Broncos from 1993 until his retirement in 2004, when the Arabian gelding Winter Solstyce took over as "Thunder II". Cal Poly Pomona's W.K. Kellogg Arabian Horse Center Equestrian Unit has made Arabian horses a regular sight at the annual Tournament of Roses Parade held each New Year's Day in Pasadena, California. Arabians also are used on search and rescue teams and occasionally for police work. Some Arabians are used in polo in the US and Europe, in the Turkish equestrian sport of Cirit (), as well as in circuses, therapeutic horseback riding programs, and on guest ranches.
Biology and health sciences
Horses
null
292744
https://en.wikipedia.org/wiki/Ising%20model
Ising model
The Ising model (or Lenz–Ising model), named after the physicists Ernst Ising and Wilhelm Lenz, is a mathematical model of ferromagnetism in statistical mechanics. The model consists of discrete variables that represent magnetic dipole moments of atomic "spins" that can be in one of two states (+1 or −1). The spins are arranged in a graph, usually a lattice (where the local structure repeats periodically in all directions), allowing each spin to interact with its neighbors. Neighboring spins that agree have a lower energy than those that disagree; the system tends to the lowest energy but heat disturbs this tendency, thus creating the possibility of different structural phases. The model allows the identification of phase transitions as a simplified model of reality. The two-dimensional square-lattice Ising model is one of the simplest statistical models to show a phase transition. The Ising model was invented by the physicist , who gave it as a problem to his student Ernst Ising. The one-dimensional Ising model was solved by alone in his 1924 thesis; it has no phase transition. The two-dimensional square-lattice Ising model is much harder and was only given an analytic description much later, by . It is usually solved by a transfer-matrix method, although there exists a very simple approach relating the model to a non-interacting fermionic quantum field theory. In dimensions greater than four, the phase transition of the Ising model is described by mean-field theory. The Ising model for greater dimensions was also explored with respect to various tree topologies in the late 1970s, culminating in an exact solution of the zero-field, time-independent model for closed Cayley trees of arbitrary branching ratio, and thereby, arbitrarily large dimensionality within tree branches. The solution to this model exhibited a new, unusual phase transition behavior, along with non-vanishing long-range and nearest-neighbor spin-spin correlations, deemed relevant to large neural networks as one of its possible . The Ising problem without an external field can be equivalently formulated as a graph maximum cut (Max-Cut) problem that can be solved via combinatorial optimization. Definition Consider a set of lattice sites, each with a set of adjacent sites (e.g. a graph) forming a -dimensional lattice. For each lattice site there is a discrete variable such that , representing the site's spin. A spin configuration, is an assignment of spin value to each lattice site. For any two adjacent sites there is an interaction . Also a site has an external magnetic field interacting with it. The energy of a configuration is given by the Hamiltonian function where the first sum is over pairs of adjacent spins (every pair is counted once). The notation indicates that sites and are nearest neighbors. The magnetic moment is given by . Note that the sign in the second term of the Hamiltonian above should actually be positive because the electron's magnetic moment is antiparallel to its spin, but the negative term is used conventionally. The configuration probability is given by the Boltzmann distribution with inverse temperature : where , and the normalization constant is the partition function. For a function of the spins ("observable"), one denotes by the expectation (mean) value of . The configuration probabilities represent the probability that (in equilibrium) the system is in a state with configuration . Discussion The minus sign on each term of the Hamiltonian function is conventional. Using this sign convention, Ising models can be classified according to the sign of the interaction: if, for a pair i, j The system is called ferromagnetic or antiferromagnetic if all interactions are ferromagnetic or all are antiferromagnetic. The original Ising models were ferromagnetic, and it is still often assumed that "Ising model" means a ferromagnetic Ising model. In a ferromagnetic Ising model, spins desire to be aligned: the configurations in which adjacent spins are of the same sign have higher probability. In an antiferromagnetic model, adjacent spins tend to have opposite signs. The sign convention of H(σ) also explains how a spin site j interacts with the external field. Namely, the spin site wants to line up with the external field. If: Simplifications Ising models are often examined without an external field interacting with the lattice, that is, h = 0 for all j in the lattice Λ. Using this simplification, the Hamiltonian becomes When the external field is zero everywhere, h = 0, the Ising model is symmetric under switching the value of the spin in all the lattice sites; a nonzero field breaks this symmetry. Another common simplification is to assume that all of the nearest neighbors ⟨ij⟩ have the same interaction strength. Then we can set Jij = J for all pairs i, j in Λ. In this case the Hamiltonian is further simplified to Connection to graph maximum cut A subset S of the vertex set V(G) of a weighted undirected graph G determines a cut of the graph G into S and its complementary subset G\S. The size of the cut is the sum of the weights of the edges between S and G\S. A maximum cut size is at least the size of any other cut, varying S. For the Ising model without an external field on a graph G, the Hamiltonian becomes the following sum over the graph edges E(G) . Here each vertex i of the graph is a spin site that takes a spin value . A given spin configuration partitions the set of vertices into two -depended subsets, those with spin up and those with spin down . We denote by the -depended set of edges that connects the two complementary vertex subsets and . The size of the cut to bipartite the weighted undirected graph G can be defined as where denotes a weight of the edge and the scaling 1/2 is introduced to compensate for double counting the same weights . The identities where the total sum in the first term does not depend on , imply that minimizing in is equivalent to minimizing . Defining the edge weight thus turns the Ising problem without an external field into a graph Max-Cut problem maximizing the cut size , which is related to the Ising Hamiltonian as follows, Questions A significant number of statistical questions to ask about this model are in the limit of large numbers of spins: In a typical configuration, are most of the spins +1 or −1, or are they split equally? If a spin at any given position i is 1, what is the probability that the spin at position j is also 1? If β is changed, is there a phase transition? On a lattice Λ, what is the fractal dimension of the shape of a large cluster of +1 spins? Basic properties and history The most studied case of the Ising model is the translation-invariant ferromagnetic zero-field model on a d-dimensional lattice, namely, Λ = Zd, Jij = 1, h = 0. No phase transition in one dimension In his 1924 PhD thesis, Ising solved the model for the d = 1 case, which can be thought of as a linear horizontal lattice where each site only interacts with its left and right neighbor. In one dimension, the solution admits no phase transition. Namely, for any positive β, the correlations ⟨σiσj⟩ decay exponentially in |i − j|: and the system is disordered. On the basis of this result, he incorrectly concluded that this model does not exhibit phase behaviour in any dimension. Phase transition and exact solution in two dimensions The Ising model undergoes a phase transition between an ordered and a disordered phase in 2 dimensions or more. Namely, the system is disordered for small β, whereas for large β the system exhibits ferromagnetic order: This was first proven by Rudolf Peierls in 1936, using what is now called a Peierls argument. The Ising model on a two-dimensional square lattice with no magnetic field was analytically solved by . Onsager obtained the correlation functions and free energy of the Ising model and announced the formula for the spontaneous magnetization for the 2-dimensional model in 1949 but did not give a derivation. gave the first published proof of this formula, using a limit formula for Fredholm determinants, proved in 1951 by Szegő in direct response to Onsager's work. Correlation inequalities A number of correlation inequalities have been derived rigorously for the Ising spin correlations (for general lattice structures), which have enabled mathematicians to study the Ising model both on and off criticality. Griffiths inequality Given any subset of spins and on the lattice, the following inequality holds, where . With , the special case results. This means that spins are positively correlated on the Ising ferromagnet. An immediate application of this is that the magnetization of any set of spins is increasing with respect to any set of coupling constants . Simon-Lieb inequality The Simon-Lieb inequality states that for any set disconnecting from (e.g. the boundary of a box with being inside the box and being outside), This inequality can be used to establish the sharpness of phase transition for the Ising model. FKG inequality This inequality is proven first for a type of positively-correlated percolation model, of which includes a representation of the Ising model. It is used to determine the critical temperatures of planar Potts model using percolation arguments (which includes the Ising model as a special case). Historical significance One of Democritus' arguments in support of atomism was that atoms naturally explain the sharp phase boundaries observed in materials, as when ice melts to water or water turns to steam. His idea was that small changes in atomic-scale properties would lead to big changes in the aggregate behavior. Others believed that matter is inherently continuous, not atomic, and that the large-scale properties of matter are not reducible to basic atomic properties. While the laws of chemical bonding made it clear to nineteenth century chemists that atoms were real, among physicists the debate continued well into the early twentieth century. Atomists, notably James Clerk Maxwell and Ludwig Boltzmann, applied Hamilton's formulation of Newton's laws to large systems, and found that the statistical behavior of the atoms correctly describes room temperature gases. But classical statistical mechanics did not account for all of the properties of liquids and solids, nor of gases at low temperature. Once modern quantum mechanics was formulated, atomism was no longer in conflict with experiment, but this did not lead to a universal acceptance of statistical mechanics, which went beyond atomism. Josiah Willard Gibbs had given a complete formalism to reproduce the laws of thermodynamics from the laws of mechanics. But many faulty arguments survived from the 19th century, when statistical mechanics was considered dubious. The lapses in intuition mostly stemmed from the fact that the limit of an infinite statistical system has many zero-one laws which are absent in finite systems: an infinitesimal change in a parameter can lead to big differences in the overall, aggregate behavior, as Democritus expected. No phase transitions in finite volume In the early part of the twentieth century, some believed that the partition function could never describe a phase transition, based on the following argument: The partition function is a sum of e−βE over all configurations. The exponential function is everywhere analytic as a function of β. The sum of analytic functions is an analytic function. This argument works for a finite sum of exponentials, and correctly establishes that there are no singularities in the free energy of a system of a finite size. For systems which are in the thermodynamic limit (that is, for infinite systems) the infinite sum can lead to singularities. The convergence to the thermodynamic limit is fast, so that the phase behavior is apparent already on a relatively small lattice, even though the singularities are smoothed out by the system's finite size. This was first established by Rudolf Peierls in the Ising model. Peierls droplets Shortly after Lenz and Ising constructed the Ising model, Peierls was able to explicitly show that a phase transition occurs in two dimensions. To do this, he compared the high-temperature and low-temperature limits. At infinite temperature (β = 0) all configurations have equal probability. Each spin is completely independent of any other, and if typical configurations at infinite temperature are plotted so that plus/minus are represented by black and white, they look like television snow. For high, but not infinite temperature, there are small correlations between neighboring positions, the snow tends to clump a little bit, but the screen stays randomly looking, and there is no net excess of black or white. A quantitative measure of the excess is the magnetization, which is the average value of the spin: A bogus argument analogous to the argument in the last section now establishes that the magnetization in the Ising model is always zero. Every configuration of spins has equal energy to the configuration with all spins flipped. So for every configuration with magnetization M there is a configuration with magnetization −M with equal probability. The system should therefore spend equal amounts of time in the configuration with magnetization M as with magnetization −M. So the average magnetization (over all time) is zero. As before, this only proves that the average magnetization is zero at any finite volume. For an infinite system, fluctuations might not be able to push the system from a mostly plus state to a mostly minus with a nonzero probability. For very high temperatures, the magnetization is zero, as it is at infinite temperature. To see this, note that if spin A has only a small correlation ε with spin B, and B is only weakly correlated with C, but C is otherwise independent of A, the amount of correlation of A and C goes like ε2. For two spins separated by distance L, the amount of correlation goes as εL, but if there is more than one path by which the correlations can travel, this amount is enhanced by the number of paths. The number of paths of length L on a square lattice in d dimensions is since there are 2d choices for where to go at each step. A bound on the total correlation is given by the contribution to the correlation by summing over all paths linking two points, which is bounded above by the sum over all paths of length L divided by which goes to zero when ε is small. At low temperatures (β ≫ 1) the configurations are near the lowest-energy configuration, the one where all the spins are plus or all the spins are minus. Peierls asked whether it is statistically possible at low temperature, starting with all the spins minus, to fluctuate to a state where most of the spins are plus. For this to happen, droplets of plus spin must be able to congeal to make the plus state. The energy of a droplet of plus spins in a minus background is proportional to the perimeter of the droplet L, where plus spins and minus spins neighbor each other. For a droplet with perimeter L, the area is somewhere between (L − 2)/2 (the straight line) and (L/4)2 (the square box). The probability cost for introducing a droplet has the factor e−βL, but this contributes to the partition function multiplied by the total number of droplets with perimeter L, which is less than the total number of paths of length L: So that the total spin contribution from droplets, even overcounting by allowing each site to have a separate droplet, is bounded above by which goes to zero at large β. For β sufficiently large, this exponentially suppresses long loops, so that they cannot occur, and the magnetization never fluctuates too far from −1. So Peierls established that the magnetization in the Ising model eventually defines superselection sectors, separated domains not linked by finite fluctuations. Kramers–Wannier duality Kramers and Wannier were able to show that the high-temperature expansion and the low-temperature expansion of the model are equal up to an overall rescaling of the free energy. This allowed the phase-transition point in the two-dimensional model to be determined exactly (under the assumption that there is a unique critical point). Yang–Lee zeros After Onsager's solution, Yang and Lee investigated the way in which the partition function becomes singular as the temperature approaches the critical temperature. Applications Magnetism The original motivation for the model was the phenomenon of ferromagnetism. Iron is magnetic; once it is magnetized it stays magnetized for a long time compared to any atomic time. In the 19th century, it was thought that magnetic fields are due to currents in matter, and Ampère postulated that permanent magnets are caused by permanent atomic currents. The motion of classical charged particles could not explain permanent currents though, as shown by Larmor. In order to have ferromagnetism, the atoms must have permanent magnetic moments which are not due to the motion of classical charges. Once the electron's spin was discovered, it was clear that the magnetism should be due to a large number of electron spins all oriented in the same direction. It was natural to ask how the electrons' spins all know which direction to point in, because the electrons on one side of a magnet don't directly interact with the electrons on the other side. They can only influence their neighbors. The Ising model was designed to investigate whether a large fraction of the electron spins could be oriented in the same direction using only local forces. Lattice gas The Ising model can be reinterpreted as a statistical model for the motion of atoms. Since the kinetic energy depends only on momentum and not on position, while the statistics of the positions only depends on the potential energy, the thermodynamics of the gas only depends on the potential energy for each configuration of atoms. A coarse model is to make space-time a lattice and imagine that each position either contains an atom or it doesn't. The space of configuration is that of independent bits Bi, where each bit is either 0 or 1 depending on whether the position is occupied or not. An attractive interaction reduces the energy of two nearby atoms. If the attraction is only between nearest neighbors, the energy is reduced by −4JBiBj for each occupied neighboring pair. The density of the atoms can be controlled by adding a chemical potential, which is a multiplicative probability cost for adding one more atom. A multiplicative factor in probability can be reinterpreted as an additive term in the logarithm – the energy. The extra energy of a configuration with N atoms is changed by μN. The probability cost of one more atom is a factor of exp(−βμ). So the energy of the lattice gas is: Rewriting the bits in terms of spins, For lattices where every site has an equal number of neighbors, this is the Ising model with a magnetic field h = (zJ − μ)/2, where z is the number of neighbors. In biological systems, modified versions of the lattice gas model have been used to understand a range of binding behaviors. These include the binding of ligands to receptors in the cell surface, the binding of chemotaxis proteins to the flagellar motor, and the condensation of DNA. Neuroscience The activity of neurons in the brain can be modelled statistically. Each neuron at any time is either active + or inactive −. The active neurons are those that send an action potential down the axon in any given time window, and the inactive ones are those that do not. Following the general approach of Jaynes, a later interpretation of Schneidman, Berry, Segev and Bialek, is that the Ising model is useful for any model of neural function, because a statistical model for neural activity should be chosen using the principle of maximum entropy. Given a collection of neurons, a statistical model which can reproduce the average firing rate for each neuron introduces a Lagrange multiplier for each neuron: But the activity of each neuron in this model is statistically independent. To allow for pair correlations, when one neuron tends to fire (or not to fire) along with another, introduce pair-wise lagrange multipliers: where are not restricted to neighbors. Note that this generalization of Ising model is sometimes called the quadratic exponential binary distribution in statistics. This energy function only introduces probability biases for a spin having a value and for a pair of spins having the same value. Higher order correlations are unconstrained by the multipliers. An activity pattern sampled from this distribution requires the largest number of bits to store in a computer, in the most efficient coding scheme imaginable, as compared with any other distribution with the same average activity and pairwise correlations. This means that Ising models are relevant to any system which is described by bits which are as random as possible, with constraints on the pairwise correlations and the average number of 1s, which frequently occurs in both the physical and social sciences. Spin glasses With the Ising model the so-called spin glasses can also be described, by the usual Hamiltonian where the S-variables describe the Ising spins, while the Ji,k are taken from a random distribution. For spin glasses a typical distribution chooses antiferromagnetic bonds with probability p and ferromagnetic bonds with probability 1 − p (also known as the random-bond Ising model). These bonds stay fixed or "quenched" even in the presence of thermal fluctuations. When p = 0 we have the original Ising model. This system deserves interest in its own; particularly one has "non-ergodic" properties leading to strange relaxation behaviour. Much attention has been also attracted by the related bond and site dilute Ising model, especially in two dimensions, leading to intriguing critical behavior. Artificial neural network Ising model was instrumental in the development of the Hopfield network. The original Ising model is a model for equilibrium. Roy J. Glauber in 1963 studied the Ising model evolving in time, as a process towards thermal equilibrium (Glauber dynamics), adding in the component of time. (Kaoru Nakano, 1971) and (Shun'ichi Amari, 1972), proposed to modify the weights of an Ising model by Hebbian learning rule as a model of associative memory. The same idea was published by (, 1974), who was cited by Hopfield in his 1982 paper. The Sherrington–Kirkpatrick model of spin glass, published in 1975, is the Hopfield network with random initialization. Sherrington and Kirkpatrick found that it is highly likely for the energy function of the SK model to have many local minima. In the 1982 paper, Hopfield applied this recently developed theory to study the Hopfield network with binary activation functions. In a 1984 paper he extended this to continuous activation functions. It became a standard model for the study of neural networks through statistical mechanics. Sea ice The melt pond can be modelled by the Ising model; sea ice topography data bears rather heavily on the results. The state variable is binary for a simple 2D approximation, being either water or ice. Cayley tree topologies and large neural networks In order to investigate an Ising model with potential relevance for large (e.g. with or interactions per node) neural nets, at the suggestion of Krizan in 1979, obtained the exact analytical expression for the free energy of the Ising model on the closed Cayley tree (with an arbitrarily large branching ratio) for a zero-external magnetic field (in the thermodynamic limit) by applying the methodologies of and where is an arbitrary branching ratio (greater than or equal to 2), , , (with representing the nearest-neighbor interaction energy) and there are k (→ ∞ in the thermodynamic limit) generations in each of the tree branches (forming the closed tree architecture as shown in the given closed Cayley tree diagram.) The sum in the last term can be shown to converge uniformly and rapidly (i.e. for z → ∞, it remains finite) yielding a continuous and monotonous function, establishing that, for greater than or equal to 2, the free energy is a continuous function of temperature T. Further analysis of the free energy indicates that it exhibits an unusual discontinuous first derivative at the critical temperature (, .) The spin-spin correlation between sites (in general, m and n) on the tree was found to have a transition point when considered at the vertices (e.g. A and Ā, its reflection), their respective neighboring sites (such as B and its reflection), and between sites adjacent to the top and bottom extreme vertices of the two trees (e.g. A and B), as may be determined from where is equal to the number of bonds, is the number of graphs counted for odd vertices with even intermediate sites (see cited methodologies and references for detailed calculations), is the multiplicity resulting from two-valued spin possibilities and the partition function is derived from . (Note: is consistent with the referenced literature in this section and is equivalent to or utilized above and in earlier sections; it is valued at .) The critical temperature is given by The critical temperature for this model is only determined by the branching ratio and the site-to-site interaction energy , a fact which may have direct implications associated with neural structure vs. its function (in that it relates the energies of interaction and branching ratio to its transitional behavior.) For example, a relationship between the transition behavior of activities of neural networks between sleeping and wakeful states (which may correlate with a spin-spin type of phase transition) in terms of changes in neural interconnectivity () and/or neighbor-to-neighbor interactions (), over time, is just one possible avenue suggested for further experimental investigation into such a phenomenon. In any case, for this Ising model it was established, that “the stability of the long-range correlation increases with increasing or increasing .” For this topology, the spin-spin correlation was found to be zero between the extreme vertices and the central sites at which the two trees (or branches) are joined (i.e. between A and individually C, D, or E.) This behavior is explained to be due to the fact that, as k increases, the number of links increases exponentially (between the extreme vertices) and so even though the contribution to spin correlations decrease exponentially, the correlation between sites such as the extreme vertex (A) in one tree and the extreme vertex in the joined tree (Ā) remains finite (above the critical temperature.) In addition, A and B also exhibit a non-vanishing correlation (as do their reflections) thus lending itself to, for B level sites (with A level), being considered “clusters” which tend to exhibit synchronization of firing. Based upon a review of other classical network models as a comparison, the Ising model on a closed Cayley tree was determined to be the first classical statistical mechanical model to demonstrate both local and long-range sites with non-vanishing spin-spin correlations, while at the same time exhibiting intermediate sites with zero correlation, which indeed was a relevant matter for large neural networks at the time of its consideration. The model's behavior is also of relevance for any other divergent-convergent tree physical (or biological) system exhibiting a closed Cayley tree topology with an Ising-type of interaction. This topology should not be ignored since its behavior for Ising models has been solved exactly, and presumably nature will have found a way of taking advantage of such simple symmetries at many levels of its designs. early on noted the possibility of interrelationships between (1) the classical large neural network model (with similar coupled divergent-convergent topologies) with (2) an underlying statistical quantum mechanical model (independent of topology and with persistence in fundamental quantum states): It was a natural and common belief among early neurophysicists (e.g. Umezawa, Krizan, Barth, etc.) that classical neural models (including those with statistical mechanical aspects) will one day have to be integrated with quantum physics (with quantum statistical aspects), similar perhaps to how the domain of chemistry has historically integrated itself into quantum physics via quantum chemistry. Several additional statistical mechanical problems of interest remain to be solved for the closed Cayley tree, including the time-dependent case and the external field situation, as well as theoretical efforts aimed at understanding interrelationships with underlying quantum constituents and their physics. Numerical simulation The Ising model can often be difficult to evaluate numerically if there are many states in the system. Consider an Ising model with L = |Λ|: the total number of sites on the lattice, σj ∈ {−1, +1}: an individual spin site on the lattice, j = 1, ..., L, S ∈ {−1, +1}L: state of the system. Since every spin site has ±1 spin, there are 2L different states that are possible. This motivates the reason for the Ising model to be simulated using Monte Carlo methods. The Hamiltonian that is commonly used to represent the energy of the model when using Monte Carlo methods is: Furthermore, the Hamiltonian is further simplified by assuming zero external field h, since many questions that are posed to be solved using the model can be answered in absence of an external field. This leads us to the following energy equation for state σ: Given this Hamiltonian, quantities of interest such as the specific heat or the magnetization of the magnet at a given temperature can be calculated. Metropolis algorithm The Metropolis–Hastings algorithm is the most commonly used Monte Carlo algorithm to calculate Ising model estimations. The algorithm first chooses selection probabilities g(μ, ν), which represent the probability that state ν is selected by the algorithm out of all states, given that one is in state μ. It then uses acceptance probabilities A(μ, ν) so that detailed balance is satisfied. If the new state ν is accepted, then we move to that state and repeat with selecting a new state and deciding to accept it. If ν is not accepted then we stay in μ. This process is repeated until some stopping criterion is met, which for the Ising model is often when the lattice becomes ferromagnetic, meaning all of the sites point in the same direction. When implementing the algorithm, one must ensure that g(μ, ν) is selected such that ergodicity is met. In thermal equilibrium a system's energy only fluctuates within a small range. This is the motivation behind the concept of single-spin-flip dynamics, which states that in each transition, we will only change one of the spin sites on the lattice. Furthermore, by using single- spin-flip dynamics, one can get from any state to any other state by flipping each site that differs between the two states one at a time. The maximum amount of change between the energy of the present state, Hμ and any possible new state's energy Hν (using single-spin-flip dynamics) is 2J between the spin we choose to "flip" to move to the new state and that spin's neighbor. Thus, in a 1D Ising model, where each site has two neighbors (left and right), the maximum difference in energy would be 4J. Let c represent the lattice coordination number; the number of nearest neighbors that any lattice site has. We assume that all sites have the same number of neighbors due to periodic boundary conditions. It is important to note that the Metropolis–Hastings algorithm does not perform well around the critical point due to critical slowing down. Other techniques such as multigrid methods, Niedermayer's algorithm, Swendsen–Wang algorithm, or the Wolff algorithm are required in order to resolve the model near the critical point; a requirement for determining the critical exponents of the system. Specifically for the Ising model and using single-spin-flip dynamics, one can establish the following. Since there are L total sites on the lattice, using single-spin-flip as the only way we transition to another state, we can see that there are a total of L new states ν from our present state μ. The algorithm assumes that the selection probabilities are equal to the L states: g(μ, ν) = 1/L. Detailed balance tells us that the following equation must hold: Thus, we want to select the acceptance probability for our algorithm to satisfy If Hν > Hμ, then A(ν, μ) > A(μ, ν). Metropolis sets the larger of A(μ, ν) or A(ν, μ) to be 1. By this reasoning the acceptance algorithm is: The basic form of the algorithm is as follows: Pick a spin site using selection probability g(μ, ν) and calculate the contribution to the energy involving this spin. Flip the value of the spin and calculate the new contribution. If the new energy is less, keep the flipped value. If the new energy is more, only keep with probability Repeat. The change in energy Hν − Hμ only depends on the value of the spin and its nearest graph neighbors. So if the graph is not too connected, the algorithm is fast. This process will eventually produce a pick from the distribution. As a Markov chain It is possible to view the Ising model as a Markov chain, as the immediate probability Pβ(ν) of transitioning to a future state ν only depends on the present state μ. The Metropolis algorithm is actually a version of a Markov chain Monte Carlo simulation, and since we use single-spin-flip dynamics in the Metropolis algorithm, every state can be viewed as having links to exactly L other states, where each transition corresponds to flipping a single spin site to the opposite value. Furthermore, since the energy equation Hσ change only depends on the nearest-neighbor interaction strength J, the Ising model and its variants such the Sznajd model can be seen as a form of a voter model for opinion dynamics. Solutions One dimension The thermodynamic limit exists as long as the interaction decay is with α > 1. In the case of ferromagnetic interaction with 1 < α < 2, Dyson proved, by comparison with the hierarchical case, that there is phase transition at small enough temperature. In the case of ferromagnetic interaction , Fröhlich and Spencer proved that there is phase transition at small enough temperature (in contrast with the hierarchical case). In the case of interaction with α > 2 (which includes the case of finite-range interactions), there is no phase transition at any positive temperature (i.e. finite β), since the free energy is analytic in the thermodynamic parameters. In the case of nearest neighbor interactions, E. Ising provided an exact solution of the model. At any positive temperature (i.e. finite β) the free energy is analytic in the thermodynamics parameters, and the truncated two-point spin correlation decays exponentially fast. At zero temperature (i.e. infinite β), there is a second-order phase transition: the free energy is infinite, and the truncated two-point spin correlation does not decay (remains constant). Therefore, T = 0 is the critical temperature of this case. Scaling formulas are satisfied. Ising's exact solution In the nearest neighbor case (with periodic or free boundary conditions) an exact solution is available. The Hamiltonian of the one-dimensional Ising model on a lattice of L sites with free boundary conditions is where J and h can be any number, since in this simplified case J is a constant representing the interaction strength between the nearest neighbors and h is the constant external magnetic field applied to lattice sites. Then the free energy is and the spin-spin correlation (i.e. the covariance) is where C(β) and c(β) are positive functions for T > 0. For T → 0, though, the inverse correlation length c(β) vanishes. Proof The proof of this result is a simple computation. If h = 0, it is very easy to obtain the free energy in the case of free boundary condition, i.e. when Then the model factorizes under the change of variables This gives Therefore, the free energy is With the same change of variables hence it decays exponentially as soon as T ≠ 0; but for T = 0, i.e. in the limit β → ∞ there is no decay. If h ≠ 0 we need the transfer matrix method. For the periodic boundary conditions case is the following. The partition function is The coefficients can be seen as the entries of a matrix. There are different possible choices: a convenient one (because the matrix is symmetric) is or In matrix formalism where λ1 is the highest eigenvalue of V, while is the other eigenvalue: and . This gives the formula of the free energy above. In the thermodynamics limit for the non-interaction case (J = 0), we got as the answer for the open-boundary Ising model. Comments The energy of the lowest state is −JL, when all the spins are the same. For any other configuration, the extra energy is equal to 2J times the number of sign changes that are encountered when scanning the configuration from left to right. If we designate the number of sign changes in a configuration as k, the difference in energy from the lowest energy state is 2k. Since the energy is additive in the number of flips, the probability p of having a spin-flip at each position is independent. The ratio of the probability of finding a flip to the probability of not finding one is the Boltzmann factor: The problem is reduced to independent biased coin tosses. This essentially completes the mathematical description. From the description in terms of independent tosses, the statistics of the model for long lines can be understood. The line splits into domains. Each domain is of average length exp(2β). The length of a domain is distributed exponentially, since there is a constant probability at any step of encountering a flip. The domains never become infinite, so a long system is never magnetized. Each step reduces the correlation between a spin and its neighbor by an amount proportional to p, so the correlations fall off exponentially. The partition function is the volume of configurations, each configuration weighted by its Boltzmann weight. Since each configuration is described by the sign-changes, the partition function factorizes: The logarithm divided by L is the free energy density: which is analytic away from β = ∞. A sign of a phase transition is a non-analytic free energy, so the one-dimensional model does not have a phase transition. One-dimensional solution with transverse field To express the Ising Hamiltonian using a quantum mechanical description of spins, we replace the spin variables with their respective Pauli matrices. However, depending on the direction of the magnetic field, we can create a transverse-field or longitudinal-field Hamiltonian. The transverse-field Hamiltonian is given by The transverse-field model experiences a phase transition between an ordered and disordered regime at J ~ h. This can be shown by a mapping of Pauli matrices Upon rewriting the Hamiltonian in terms of this change-of-basis matrices, we obtain Since the roles of h and J are switched, the Hamiltonian undergoes a transition at J = h. Renormalization When there is no external field, we can derive a functional equation that satisfies using renormalization. Specifically, let be the partition function with sites. Now we have:where . We sum over each of , to obtainNow, since the cosh function is even, we can solve as . Now we have a self-similarity relation:Taking the limit, we obtainwhere . When is small, we have , so we can numerically evaluate by iterating the functional equation until is small. Two dimensions In the ferromagnetic case there is a phase transition. At low temperature, the Peierls argument proves positive magnetization for the nearest neighbor case and then, by the Griffiths inequality, also when longer range interactions are added. Meanwhile, at high temperature, the cluster expansion gives analyticity of the thermodynamic functions. In the nearest-neighbor case, the free energy was exactly computed by Onsager. The spin-spin correlation functions were computed by McCoy and Wu. Onsager's exact solution obtained the following analytical expression for the free energy of the Ising model on the anisotropic square lattice when the magnetic field in the thermodynamic limit as a function of temperature and the horizontal and vertical interaction energies and , respectively From this expression for the free energy, all thermodynamic functions of the model can be calculated by using an appropriate derivative. The 2D Ising model was the first model to exhibit a continuous phase transition at a positive temperature. It occurs at the temperature which solves the equation In the isotropic case when the horizontal and vertical interaction energies are equal , the critical temperature occurs at the following point When the interaction energies , are both negative, the Ising model becomes an antiferromagnet. Since the square lattice is bi-partite, it is invariant under this change when the magnetic field , so the free energy and critical temperature are the same for the antiferromagnetic case. For the triangular lattice, which is not bi-partite, the ferromagnetic and antiferromagnetic Ising model behave notably differently. Specifically, around a triangle, it is impossible to make all 3 spin-pairs antiparallel, so the antiferromagnetic Ising model cannot reach the minimal energy state. This is an example of geometric frustration. Transfer matrix Start with an analogy with quantum mechanics. The Ising model on a long periodic lattice has a partition function Think of the i direction as space, and the j direction as time. This is an independent sum over all the values that the spins can take at each time slice. This is a type of path integral, it is the sum over all spin histories. A path integral can be rewritten as a Hamiltonian evolution. The Hamiltonian steps through time by performing a unitary rotation between time t and time t + Δt: The product of the U matrices, one after the other, is the total time evolution operator, which is the path integral we started with. where N is the number of time slices. The sum over all paths is given by a product of matrices, each matrix element is the transition probability from one slice to the next. Similarly, one can divide the sum over all partition function configurations into slices, where each slice is the one-dimensional configuration at time 1. This defines the transfer matrix: The configuration in each slice is a one-dimensional collection of spins. At each time slice, T has matrix elements between two configurations of spins, one in the immediate future and one in the immediate past. These two configurations are C1 and C2, and they are all one-dimensional spin configurations. We can think of the vector space that T acts on as all complex linear combinations of these. Using quantum mechanical notation: where each basis vector is a spin configuration of a one-dimensional Ising model. Like the Hamiltonian, the transfer matrix acts on all linear combinations of states. The partition function is a matrix function of T, which is defined by the sum over all histories which come back to the original configuration after N steps: Since this is a matrix equation, it can be evaluated in any basis. So if we can diagonalize the matrix T, we can find Z. T in terms of Pauli matrices The contribution to the partition function for each past/future pair of configurations on a slice is the sum of two terms. There is the number of spin flips in the past slice and there is the number of spin flips between the past and future slice. Define an operator on configurations which flips the spin at site i: In the usual Ising basis, acting on any linear combination of past configurations, it produces the same linear combination but with the spin at position i of each basis vector flipped. Define a second operator which multiplies the basis vector by +1 and −1 according to the spin at position i: T can be written in terms of these: where A and B are constants which are to be determined so as to reproduce the partition function. The interpretation is that the statistical configuration at this slice contributes according to both the number of spin flips in the slice, and whether or not the spin at position i has flipped. Spin flip creation and annihilation operators Just as in the one-dimensional case, we will shift attention from the spins to the spin-flips. The σz term in T counts the number of spin flips, which we can write in terms of spin-flip creation and annihilation operators: The first term flips a spin, so depending on the basis state it either: moves a spin-flip one unit to the right moves a spin-flip one unit to the left produces two spin-flips on neighboring sites destroys two spin-flips on neighboring sites. Writing this out in terms of creation and annihilation operators: Ignore the constant coefficients, and focus attention on the form. They are all quadratic. Since the coefficients are constant, this means that the T matrix can be diagonalized by Fourier transforms. Carrying out the diagonalization produces the Onsager free energy. Onsager's formula for spontaneous magnetization Onsager famously announced the following expression for the spontaneous magnetization M of a two-dimensional Ising ferromagnet on the square lattice at two different conferences in 1948, though without proof where and are horizontal and vertical interaction energies. A complete derivation was only given in 1951 by using a limiting process of transfer matrix eigenvalues. The proof was subsequently greatly simplified in 1963 by Montroll, Potts, and Ward using Szegő's limit formula for Toeplitz determinants by treating the magnetization as the limit of correlation functions. Minimal model At the critical point, the two-dimensional Ising model is a two-dimensional conformal field theory. The spin and energy correlation functions are described by a minimal model, which has been exactly solved. Three dimensions In three as in two dimensions, the most studied case of the Ising model is the translation-invariant model on a cubic lattice with nearest-neighbor coupling in the zero magnetic field. Many theoreticians searched for an analytical three-dimensional solution for many decades, which would be analogous to Onsager's solution in the two-dimensional case. Such a solution has not been found until now, although there is no proof that it may not exist. In three dimensions, the Ising model was shown to have a representation in terms of non-interacting fermionic strings by Alexander Polyakov and Vladimir Dotsenko. This construction has been carried on the lattice, and the continuum limit, conjecturally describing the critical point, is unknown. In three as in two dimensions, Peierls' argument shows that there is a phase transition. This phase transition is rigorously known to be continuous (in the sense that correlation length diverges and the magnetization goes to zero), and is called the critical point. It is believed that the critical point can be described by a renormalization group fixed point of the Wilson-Kadanoff renormalization group transformation. It is also believed that the phase transition can be described by a three-dimensional unitary conformal field theory, as evidenced by Monte Carlo simulations, exact diagonalization results in quantum models, and quantum field theoretical arguments. Although it is an open problem to establish rigorously the renormalization group picture or the conformal field theory picture, theoretical physicists have used these two methods to compute the critical exponents of the phase transition, which agree with the experiments and with the Monte Carlo simulations. This conformal field theory describing the three-dimensional Ising critical point is under active investigation using the method of the conformal bootstrap. This method currently yields the most precise information about the structure of the critical theory (see Ising critical exponents). In 2000, Sorin Istrail of Sandia National Laboratories proved that the spin glass Ising model on a nonplanar lattice is NP-complete. That is, assuming P ≠ NP, the general spin glass Ising model is exactly solvable only in planar cases, so solutions for dimensions higher than two are also intractable. Istrail's result only concerns the spin glass model with spatially varying couplings, and tells nothing about Ising's original ferromagnetic model with equal couplings. Four dimensions and above In any dimension, the Ising model can be productively described by a locally varying mean field. The field is defined as the average spin value over a large region, but not so large so as to include the entire system. The field still has slow variations from point to point, as the averaging volume moves. These fluctuations in the field are described by a continuum field theory in the infinite system limit. Local field The field H is defined as the long wavelength Fourier components of the spin variable, in the limit that the wavelengths are long. There are many ways to take the long wavelength average, depending on the details of how high wavelengths are cut off. The details are not too important, since the goal is to find the statistics of H and not the spins. Once the correlations in H are known, the long-distance correlations between the spins will be proportional to the long-distance correlations in H. For any value of the slowly varying field H, the free energy (log-probability) is a local analytic function of H and its gradients. The free energy F(H) is defined to be the sum over all Ising configurations which are consistent with the long wavelength field. Since H is a coarse description, there are many Ising configurations consistent with each value of H, so long as not too much exactness is required for the match. Since the allowed range of values of the spin in any region only depends on the values of H within one averaging volume from that region, the free energy contribution from each region only depends on the value of H there and in the neighboring regions. So F is a sum over all regions of a local contribution, which only depends on H and its derivatives. By symmetry in H, only even powers contribute. By reflection symmetry on a square lattice, only even powers of gradients contribute. Writing out the first few terms in the free energy: On a square lattice, symmetries guarantee that the coefficients Zi of the derivative terms are all equal. But even for an anisotropic Ising model, where the Zis in different directions are different, the fluctuations in H are isotropic in a coordinate system where the different directions of space are rescaled. On any lattice, the derivative term is a positive definite quadratic form, and can be used to define the metric for space. So any translationally invariant Ising model is rotationally invariant at long distances, in coordinates that make Zij = δij. Rotational symmetry emerges spontaneously at large distances just because there aren't very many low order terms. At higher order multicritical points, this accidental symmetry is lost. Since βF is a function of a slowly spatially varying field, the probability of any field configuration is (omitting higher-order terms): The statistical average of any product of H terms is equal to: The denominator in this expression is called the partition function:and the integral over all possible values of H is a statistical path integral. It integrates exp(βF) over all values of H, over all the long wavelength fourier components of the spins. F is a "Euclidean" Lagrangian for the field H. It is similar to the Lagrangian in of a scalar field in quantum field theory, the difference being that all the derivative terms enter with a positive sign, and there is no overall factor of i (thus "Euclidean"). Dimensional analysis The form of F can be used to predict which terms are most important by dimensional analysis. Dimensional analysis is not completely straightforward, because the scaling of H needs to be determined. In the generic case, choosing the scaling law for H is easy, since the only term that contributes is the first one, This term is the most significant, but it gives trivial behavior. This form of the free energy is ultralocal, meaning that it is a sum of an independent contribution from each point. This is like the spin-flips in the one-dimensional Ising model. Every value of H at any point fluctuates completely independently of the value at any other point. The scale of the field can be redefined to absorb the coefficient A, and then it is clear that A only determines the overall scale of fluctuations. The ultralocal model describes the long wavelength high temperature behavior of the Ising model, since in this limit the fluctuation averages are independent from point to point. To find the critical point, lower the temperature. As the temperature goes down, the fluctuations in H go up because the fluctuations are more correlated. This means that the average of a large number of spins does not become small as quickly as if they were uncorrelated, because they tend to be the same. This corresponds to decreasing A in the system of units where H does not absorb A. The phase transition can only happen when the subleading terms in F can contribute, but since the first term dominates at long distances, the coefficient A must be tuned to zero. This is the location of the critical point: where t is a parameter which goes through zero at the transition. Since t is vanishing, fixing the scale of the field using this term makes the other terms blow up. Once t is small, the scale of the field can either be set to fix the coefficient of the H4 term or the (∇H)2 term to 1. Magnetization To find the magnetization, fix the scaling of H so that λ is one. Now the field H has dimension −d/4, so that H4ddx is dimensionless, and Z has dimension 2 − d/2. In this scaling, the gradient term is only important at long distances for d ≤ 4. Above four dimensions, at long wavelengths, the overall magnetization is only affected by the ultralocal terms. There is one subtle point. The field H is fluctuating statistically, and the fluctuations can shift the zero point of t. To see how, consider H4 split in the following way: The first term is a constant contribution to the free energy, and can be ignored. The second term is a finite shift in t. The third term is a quantity that scales to zero at long distances. This means that when analyzing the scaling of t by dimensional analysis, it is the shifted t that is important. This was historically very confusing, because the shift in t at any finite λ is finite, but near the transition t is very small. The fractional change in t is very large, and in units where t is fixed the shift looks infinite. The magnetization is at the minimum of the free energy, and this is an analytic equation. In terms of the shifted t, For t < 0, the minima are at H proportional to the square root of t. So Landau's catastrophe argument is correct in dimensions larger than 5. The magnetization exponent in dimensions higher than 5 is equal to the mean-field value. When t is negative, the fluctuations about the new minimum are described by a new positive quadratic coefficient. Since this term always dominates, at temperatures below the transition the fluctuations again become ultralocal at long distances. Fluctuations To find the behavior of fluctuations, rescale the field to fix the gradient term. Then the length scaling dimension of the field is 1 − d/2. Now the field has constant quadratic spatial fluctuations at all temperatures. The scale dimension of the H2 term is 2, while the scale dimension of the H4 term is 4 − d. For d < 4, the H4 term has positive scale dimension. In dimensions higher than 4 it has negative scale dimensions. This is an essential difference. In dimensions higher than 4, fixing the scale of the gradient term means that the coefficient of the H4 term is less and less important at longer and longer wavelengths. The dimension at which nonquadratic contributions begin to contribute is known as the critical dimension. In the Ising model, the critical dimension is 4. In dimensions above 4, the critical fluctuations are described by a purely quadratic free energy at long wavelengths. This means that the correlation functions are all computable from as Gaussian averages: valid when x − y is large. The function G(x − y) is the analytic continuation to imaginary time of the Feynman propagator, since the free energy is the analytic continuation of the quantum field action for a free scalar field. For dimensions 5 and higher, all the other correlation functions at long distances are then determined by Wick's theorem. All the odd moments are zero, by ± symmetry. The even moments are the sum over all partition into pairs of the product of G(x − y) for each pair. where C is the proportionality constant. So knowing G is enough. It determines all the multipoint correlations of the field. The critical two-point function To determine the form of G, consider that the fields in a path integral obey the classical equations of motion derived by varying the free energy: This is valid at noncoincident points only, since the correlations of H are singular when points collide. H obeys classical equations of motion for the same reason that quantum mechanical operators obey them—its fluctuations are defined by a path integral. At the critical point t = 0, this is Laplace's equation, which can be solved by Gauss's method from electrostatics. Define an electric field analog by Away from the origin: since G is spherically symmetric in d dimensions, and E is the radial gradient of G. Integrating over a large d − 1 dimensional sphere, This gives: and G can be found by integrating with respect to r. The constant C fixes the overall normalization of the field. G(r) away from the critical point When t does not equal zero, so that H is fluctuating at a temperature slightly away from critical, the two point function decays at long distances. The equation it obeys is altered: For r small compared with , the solution diverges exactly the same way as in the critical case, but the long distance behavior is modified. To see how, it is convenient to represent the two point function as an integral, introduced by Schwinger in the quantum field theory context: This is G, since the Fourier transform of this integral is easy. Each fixed τ contribution is a Gaussian in x, whose Fourier transform is another Gaussian of reciprocal width in k. This is the inverse of the operator ∇2 − t in k-space, acting on the unit function in k-space, which is the Fourier transform of a delta function source localized at the origin. So it satisfies the same equation as G with the same boundary conditions that determine the strength of the divergence at 0. The interpretation of the integral representation over the proper time τ is that the two point function is the sum over all random walk paths that link position 0 to position x over time τ. The density of these paths at time τ at position x is Gaussian, but the random walkers disappear at a steady rate proportional to t so that the Gaussian at time τ is diminished in height by a factor that decreases steadily exponentially. In the quantum field theory context, these are the paths of relativistically localized quanta in a formalism that follows the paths of individual particles. In the pure statistical context, these paths still appear by the mathematical correspondence with quantum fields, but their interpretation is less directly physical. The integral representation immediately shows that G(r) is positive, since it is represented as a weighted sum of positive Gaussians. It also gives the rate of decay at large r, since the proper time for a random walk to reach position τ is r2 and in this time, the Gaussian height has decayed by . The decay factor appropriate for position r is therefore . A heuristic approximation for G(r) is: This is not an exact form, except in three dimensions, where interactions between paths become important. The exact forms in high dimensions are variants of Bessel functions. Symanzik polymer interpretation The interpretation of the correlations as fixed size quanta travelling along random walks gives a way of understanding why the critical dimension of the H4 interaction is 4. The term H4 can be thought of as the square of the density of the random walkers at any point. In order for such a term to alter the finite order correlation functions, which only introduce a few new random walks into the fluctuating environment, the new paths must intersect. Otherwise, the square of the density is just proportional to the density and only shifts the H2 coefficient by a constant. But the intersection probability of random walks depends on the dimension, and random walks in dimension higher than 4 do not intersect. The fractal dimension of an ordinary random walk is 2. The number of balls of size ε required to cover the path increase as ε−2. Two objects of fractal dimension 2 will intersect with reasonable probability only in a space of dimension 4 or less, the same condition as for a generic pair of planes. Kurt Symanzik argued that this implies that the critical Ising fluctuations in dimensions higher than 4 should be described by a free field. This argument eventually became a mathematical proof. 4 − ε dimensions – renormalization group The Ising model in four dimensions is described by a fluctuating field, but now the fluctuations are interacting. In the polymer representation, intersections of random walks are marginally possible. In the quantum field continuation, the quanta interact. The negative logarithm of the probability of any field configuration H is the free energy function The numerical factors are there to simplify the equations of motion. The goal is to understand the statistical fluctuations. Like any other non-quadratic path integral, the correlation functions have a Feynman expansion as particles travelling along random walks, splitting and rejoining at vertices. The interaction strength is parametrized by the classically dimensionless quantity λ. Although dimensional analysis shows that both λ and Z are dimensionless, this is misleading. The long wavelength statistical fluctuations are not exactly scale invariant, and only become scale invariant when the interaction strength vanishes. The reason is that there is a cutoff used to define H, and the cutoff defines the shortest wavelength. Fluctuations of H at wavelengths near the cutoff can affect the longer-wavelength fluctuations. If the system is scaled along with the cutoff, the parameters will scale by dimensional analysis, but then comparing parameters doesn't compare behavior because the rescaled system has more modes. If the system is rescaled in such a way that the short wavelength cutoff remains fixed, the long-wavelength fluctuations are modified. Wilson renormalization A quick heuristic way of studying the scaling is to cut off the H wavenumbers at a point λ. Fourier modes of H with wavenumbers larger than λ are not allowed to fluctuate. A rescaling of length that make the whole system smaller increases all wavenumbers, and moves some fluctuations above the cutoff. To restore the old cutoff, perform a partial integration over all the wavenumbers which used to be forbidden, but are now fluctuating. In Feynman diagrams, integrating over a fluctuating mode at wavenumber k links up lines carrying momentum k in a correlation function in pairs, with a factor of the inverse propagator. Under rescaling, when the system is shrunk by a factor of (1+b), the t coefficient scales up by a factor (1+b)2 by dimensional analysis. The change in t for infinitesimal b is 2bt. The other two coefficients are dimensionless and do not change at all. The lowest order effect of integrating out can be calculated from the equations of motion: This equation is an identity inside any correlation function away from other insertions. After integrating out the modes with Λ < k < (1+b)Λ, it will be a slightly different identity. Since the form of the equation will be preserved, to find the change in coefficients it is sufficient to analyze the change in the H3 term. In a Feynman diagram expansion, the H3 term in a correlation function inside a correlation has three dangling lines. Joining two of them at large wavenumber k gives a change H3 with one dangling line, so proportional to H: The factor of 3 comes from the fact that the loop can be closed in three different ways. The integral should be split into two parts: The first part is not proportional to t, and in the equation of motion it can be absorbed by a constant shift in t. It is caused by the fact that the H3 term has a linear part. Only the second term, which varies from t to t, contributes to the critical scaling. This new linear term adds to the first term on the left hand side, changing t by an amount proportional to t. The total change in t is the sum of the term from dimensional analysis and this second term from operator products: So t is rescaled, but its dimension is anomalous, it is changed by an amount proportional to the value of λ. But λ also changes. The change in λ requires considering the lines splitting and then quickly rejoining. The lowest order process is one where one of the three lines from H3 splits into three, which quickly joins with one of the other lines from the same vertex. The correction to the vertex is The numerical factor is three times bigger because there is an extra factor of three in choosing which of the three new lines to contract. So These two equations together define the renormalization group equations in four dimensions: The coefficient B is determined by the formula and is proportional to the area of a three-dimensional sphere of radius λ, times the width of the integration region bΛ divided by Λ4: In other dimensions, the constant B changes, but the same constant appears both in the t flow and in the coupling flow. The reason is that the derivative with respect to t of the closed loop with a single vertex is a closed loop with two vertices. This means that the only difference between the scaling of the coupling and the t is the combinatorial factors from joining and splitting. Wilson–Fisher fixed point To investigate three dimensions starting from the four-dimensional theory should be possible, because the intersection probabilities of random walks depend continuously on the dimensionality of the space. In the language of Feynman graphs, the coupling does not change very much when the dimension is changed. The process of continuing away from dimension 4 is not completely well defined without a prescription for how to do it. The prescription is only well defined on diagrams. It replaces the Schwinger representation in dimension 4 with the Schwinger representation in dimension 4 − ε defined by: In dimension 4 − ε, the coupling λ has positive scale dimension ε, and this must be added to the flow. The coefficient B is dimension dependent, but it will cancel. The fixed point for λ is no longer zero, but at: where the scale dimensions of t is altered by an amount λB = ε/3. The magnetization exponent is altered proportionately to: which is .333 in 3 dimensions (ε = 1) and .166 in 2 dimensions (ε = 2). This is not so far off from the measured exponent .308 and the Onsager two dimensional exponent .125. Infinite dimensions – mean field The behavior of an Ising model on a fully connected graph may be completely understood by mean-field theory. This type of description is appropriate to very-high-dimensional square lattices, because then each site has a very large number of neighbors. The idea is that if each spin is connected to a large number of spins, only the average ratio of + spins to − spins is important, since the fluctuations about this mean will be small. The mean field H is the average fraction of spins which are + minus the average fraction of spins which are −. The energy cost of flipping a single spin in the mean field H is ±2JNH. It is convenient to redefine J to absorb the factor N, so that the limit N → ∞ is smooth. In terms of the new J, the energy cost for flipping a spin is ±2JH. This energy cost gives the ratio of probability p that the spin is + to the probability 1−p that the spin is −. This ratio is the Boltzmann factor: so that The mean value of the spin is given by averaging 1 and −1 with the weights p and 1 − p, so the mean value is 2p − 1. But this average is the same for all spins, and is therefore equal to H. The solutions to this equation are the possible consistent mean fields. For βJ < 1 there is only the one solution at H = 0. For bigger values of β there are three solutions, and the solution at H = 0 is unstable. The instability means that increasing the mean field above zero a little bit produces a statistical fraction of spins which are + which is bigger than the value of the mean field. So a mean field which fluctuates above zero will produce an even greater mean field, and will eventually settle at the stable solution. This means that for temperatures below the critical value βJ = 1 the mean-field Ising model undergoes a phase transition in the limit of large N. Above the critical temperature, fluctuations in H are damped because the mean field restores the fluctuation to zero field. Below the critical temperature, the mean field is driven to a new equilibrium value, which is either the positive H or negative H solution to the equation. For βJ = 1 + ε, just below the critical temperature, the value of H can be calculated from the Taylor expansion of the hyperbolic tangent: Dividing by H to discard the unstable solution at H = 0, the stable solutions are: The spontaneous magnetization H grows near the critical point as the square root of the change in temperature. This is true whenever H can be calculated from the solution of an analytic equation which is symmetric between positive and negative values, which led Landau to suspect that all Ising type phase transitions in all dimensions should follow this law. The mean-field exponent is universal because changes in the character of solutions of analytic equations are always described by catastrophes in the Taylor series, which is a polynomial equation. By symmetry, the equation for H must only have odd powers of H on the right hand side. Changing β should only smoothly change the coefficients. The transition happens when the coefficient of H on the right hand side is 1. Near the transition: Whatever A and B are, so long as neither of them is tuned to zero, the spontaneous magnetization will grow as the square root of ε. This argument can only fail if the free energy βF is either non-analytic or non-generic at the exact β where the transition occurs. But the spontaneous magnetization in magnetic systems and the density in gasses near the critical point are measured very accurately. The density and the magnetization in three dimensions have the same power-law dependence on the temperature near the critical point, but the behavior from experiments is: The exponent is also universal, since it is the same in the Ising model as in the experimental magnet and gas, but it is not equal to the mean-field value. This was a great surprise. This is also true in two dimensions, where But there it was not a surprise, because it was predicted by Onsager. Low dimensions – block spins In three dimensions, the perturbative series from the field theory is an expansion in a coupling constant λ which is not particularly small. The effective size of the coupling at the fixed point is one over the branching factor of the particle paths, so the expansion parameter is about 1/3. In two dimensions, the perturbative expansion parameter is 2/3. But renormalization can also be productively applied to the spins directly, without passing to an average field. Historically, this approach is due to Leo Kadanoff and predated the perturbative ε expansion. The idea is to integrate out lattice spins iteratively, generating a flow in couplings. But now the couplings are lattice energy coefficients. The fact that a continuum description exists guarantees that this iteration will converge to a fixed point when the temperature is tuned to criticality. Migdal–Kadanoff renormalization Write the two-dimensional Ising model with an infinite number of possible higher order interactions. To keep spin reflection symmetry, only even powers contribute: By translation invariance, Jij is only a function of i-j. By the accidental rotational symmetry, at large i and j its size only depends on the magnitude of the two-dimensional vector i − j. The higher order coefficients are also similarly restricted. The renormalization iteration divides the lattice into two parts – even spins and odd spins. The odd spins live on the odd-checkerboard lattice positions, and the even ones on the even-checkerboard. When the spins are indexed by the position (i,j), the odd sites are those with i + j odd and the even sites those with i + j even, and even sites are only connected to odd sites. The two possible values of the odd spins will be integrated out, by summing over both possible values. This will produce a new free energy function for the remaining even spins, with new adjusted couplings. The even spins are again in a lattice, with axes tilted at 45 degrees to the old ones. Unrotating the system restores the old configuration, but with new parameters. These parameters describe the interaction between spins at distances larger. Starting from the Ising model and repeating this iteration eventually changes all the couplings. When the temperature is higher than the critical temperature, the couplings will converge to zero, since the spins at large distances are uncorrelated. But when the temperature is critical, there will be nonzero coefficients linking spins at all orders. The flow can be approximated by only considering the first few terms. This truncated flow will produce better and better approximations to the critical exponents when more terms are included. The simplest approximation is to keep only the usual J term, and discard everything else. This will generate a flow in J, analogous to the flow in t at the fixed point of λ in the ε expansion. To find the change in J, consider the four neighbors of an odd site. These are the only spins which interact with it. The multiplicative contribution to the partition function from the sum over the two values of the spin at the odd site is: where N± is the number of neighbors which are ±. Ignoring the factor of 2, the free energy contribution from this odd site is: This includes nearest neighbor and next-nearest neighbor interactions, as expected, but also a four-spin interaction which is to be discarded. To truncate to nearest neighbor interactions, consider that the difference in energy between all spins the same and equal numbers + and – is: From nearest neighbor couplings, the difference in energy between all spins equal and staggered spins is 8J. The difference in energy between all spins equal and nonstaggered but net zero spin is 4J. Ignoring four-spin interactions, a reasonable truncation is the average of these two energies or 6J. Since each link will contribute to two odd spins, the right value to compare with the previous one is half that: For small J, this quickly flows to zero coupling. Large J'''s flow to large couplings. The magnetization exponent is determined from the slope of the equation at the fixed point. Variants of this method produce good numerical approximations for the critical exponents when many terms are included, in both two and three dimensions.
Physical sciences
Magnetostatics
Physics
292941
https://en.wikipedia.org/wiki/Polyethylene%20terephthalate
Polyethylene terephthalate
Polyethylene terephthalate (or poly(ethylene terephthalate), PET, PETE, or the obsolete PETP or PET-P), is the most common thermoplastic polymer resin of the polyester family and is used in fibres for clothing, containers for liquids and foods, and thermoforming for manufacturing, and in combination with glass fibre for engineering resins. In 2016, annual production of PET was 56 million tons. The biggest application is in fibres (in excess of 60%), with bottle production accounting for about 30% of global demand. In the context of textile applications, PET is referred to by its common name, polyester, whereas the acronym PET is generally used in relation to packaging. Polyester makes up about 18% of world polymer production and is the fourth-most-produced polymer after polyethylene (PE), polypropylene (PP) and polyvinyl chloride (PVC). PET consists of repeating (C10H8O4) units. PET is commonly recycled, and has the digit 1 (♳) as its resin identification code (RIC). The National Association for PET Container Resources (NAPCOR) defines PET as: "Polyethylene terephthalate items referenced are derived from terephthalic acid (or dimethyl terephthalate) and mono ethylene glycol, wherein the sum of terephthalic acid (or dimethyl terephthalate) and mono ethylene glycol reacted constitutes at least 90 percent of the mass of monomer reacted to form the polymer, and must exhibit a melting peak temperature between 225 °C and 255 °C, as identified during the second thermal scan in procedure 10.1 in ASTM D3418, when heating the sample at a rate of 10 °C/minute." Depending on its processing and thermal history, polyethylene terephthalate may exist both as an amorphous (transparent) and as a semi-crystalline polymer. The semicrystalline material might appear transparent (particle size less than 500 nm) or opaque and white (particle size up to a few micrometers) depending on its crystal structure and particle size. One process for making PET uses bis(2-hydroxyethyl) terephthalate, which can be synthesized by the esterification reaction between terephthalic acid and ethylene glycol with water as a byproduct (this is also known as a condensation reaction), or by transesterification reaction between ethylene glycol and dimethyl terephthalate (DMT) with methanol as a byproduct. Polymerization is through a polycondensation reaction of the monomers (done immediately after esterification/transesterification) with water as the byproduct. Uses Textiles Polyester fibres are widely used in the textile industry. The invention of the polyester fibre is attributed to J. R. Whinfield. It was first commercialized in the 1940s by ICI, under the brand 'Terylene'. Subsequently E. I. DuPont launched the brand 'Dacron'. As of 2022, there are many brands around the world, mostly Asian. Polyester fibres are used in fashion apparel often blended with cotton, as heat insulation layers in thermal wear, sportswear and workwear and automotive upholstery. Rigid packaging Plastic bottles made from PET are widely used for soft drinks, both still and sparkling. For beverages that are degraded by oxygen, such as beer, a multilayer structure is used. PET sandwiches an additional polyvinyl alcohol (PVOH) or polyamide (PA) layer to further reduce its oxygen permeability. Non-oriented PET sheet can be thermoformed to make packaging trays and blister packs. Crystallizable PET withstands freezing and oven baking temperatures. Both amorphous PET and BoPET are transparent to the naked eye. Color-conferring dyes can easily be formulated into PET sheet. PET is permeable to oxygen and carbon dioxide and this imposes shelf life limitations of contents packaged in PET. In the early 2000s, the global PET packaging market grew at a compound annual growth rate of 9% to €17 billion in 2006. Flexible packaging Biaxially oriented PET (BOPET) film (including brands like "Mylar") can be aluminized by evaporating a thin film of metal onto it to reduce its permeability, and to make it reflective and opaque (MPET). These properties are useful in many applications, including flexible food packaging and thermal insulation (such as space blankets). Photovoltaic modules BOPET is used in the backsheet of photovoltaic modules. Most backsheets consist of a layer of BOPET laminated to a fluoropolymer or a layer of UV stabilized BOPET. PET is also used as a substrate in thin film solar cells. Thermoplastic resins PET can be compounded with glass fibre and crystallization accelerators, to make thermoplastic resins. These can be injection moulded into parts such as housings, covers, electrical appliance components and elements of the ignition system. Nanodiamonds PET is stoichiometrically a mixture of carbon and , and therefore has been used in an experiment involving laser-driven shock compression which created nanodiamonds and superionic water. This could be a possible way of producing nanodiamonds commercially. Other applications A waterproofing barrier in undersea cables. As a film base. As a fibre, spliced into bell rope tops to help prevent wear on the ropes as they pass through the ceiling. Since late 2014 as liner material in type IV composite high pressure gas cylinders. PET works as a much better barrier to oxygen than earlier used (LD)PE. As a 3D printing filament, as well as in the 3D printing plastic PETG (polyethylene terephthalate glycol). In 3D printing PETG has become a popular material - used for high-end applications like surgical fracture tables to automotive and aeronautical sectors, among other industrial applications. The surface properties can be modified to make PETG self-cleaning for applications like the fabrication of traffic signs for the manufacture of light-emitting diode LED spotlights. As one of three layers for the creation of glitter; acting as a plastic core coated with aluminum and topped with plastic to create a light reflecting surface, although as of 2021 many glitter manufacturing companies have begun to phase out the use of PET after calls from organizers of festivals to create bio-friendly glitter alternatives. Film for tape applications, such as the carrier for magnetic tape or backing for pressure-sensitive adhesive tapes. Digitalization has caused the virtual disappeance of the magnetic audio and videotape application. Water-resistant paper. History PET was patented in 1941 by John Rex Whinfield, James Tennant Dickson and their employer the Calico Printers' Association of Manchester, England. E. I. DuPont de Nemours in Delaware, United States, first produced Dacron (PET fiber) in 1950 and used the trademark Mylar (boPET film) in June 1951 and received registration of it in 1952. It is still the best-known name used for polyester film. The current owner of the trademark is DuPont Teijin Films. In the Soviet Union, PET was first manufactured in the laboratories of the Institute of High-Molecular Compounds of the USSR Academy of Sciences in 1949, and its name "Lavsan" is an acronym thereof (лаборатории Института высокомолекулярных соединений Академии наук СССР). The PET bottle was invented in 1973 by Nathaniel Wyeth and patented by DuPont. Physical properties PET in its most stable state is a colorless, semi-crystalline resin. However it is intrinsically slow to crystallize compared to other semicrystalline polymers. Depending on processing conditions it can be formed into either non-crystalline (amorphous) or crystalline articles. Its amenability to drawing in manufacturing makes PET useful in fibre and film applications. Like most aromatic polymers, it has better barrier properties than aliphatic polymers. It is strong and impact-resistant. PET is hygroscopic and absorbs water. About 60% crystallization is the upper limit for commercial products, with the exception of polyester fibers. Transparent products can be produced by rapidly cooling molten polymer below the glass transition temperature (Tg) to form a non-crystalline amorphous solid. Like glass, amorphous PET forms when its molecules are not given enough time to arrange themselves in an orderly, crystalline fashion as the melt is cooled. While at room temperature the molecules are frozen in place, if enough heat energy is put back into them afterward by heating the material above Tg, they can begin to move again, allowing crystals to nucleate and grow. This procedure is known as solid-state crystallization. Amorphous PET also crystallizes and becomes opaque when exposed to solvents, such as chloroform or toluene. A more crystalline product can be produced by allowing the molten polymer to cool slowly. Rather than forming one large single crystal, this material has a number of spherulites (crystallized areas) each containing many small crystallites (grains). Light tends to scatter as it crosses the boundaries between crystallites and the amorphous regions between them, causing the resulting solid to be translucent. Orientation also renders polymers more transparent. This is why BOPET film and bottles are both crystalline, to a degree, and transparent. Flavor absorption PET has an affinity for hydrophobic flavors, and drinks sometimes need to be formulated with a higher flavor dosage, compared to those going into glass, to offset the flavor taken up by the container. While heavy gauge PET bottles returned for re-use, as in some EU countries, the propensity of PET to absorb flavors makes it necessary to conduct a "sniffer test" on returned bottles to avoid cross-contamination of flavors. Intrinsic viscosity Different applications of PET require different degrees of polymerization, which can be obtained by modifying the process conditions. The molecular weight of PET is measured by solution viscosity. The preferred method to measure this viscosity is the intrinsic viscosity (IV) of the polymer. Intrinsic viscosity is a dimensionless measurement found by extrapolating the relative viscosity (measured in (dℓ/g)) to zero concentration. Shown below are the IV ranges for common applications: Copolymers PET is often copolymerized with other diols or diacids to optimize the properties for particular applications. PETG For example, cyclohexanedimethanol (CHDM) can be added to the polymer backbone in place of ethylene glycol. Since this building block is much larger (six additional carbon atoms) than the ethylene glycol unit it replaces, it does not fit in with the neighboring chains the way an ethylene glycol unit would. This interferes with crystallization and lowers the polymer's melting temperature. In general, such PET is known as PETG or PET-G (polyethylene terephthalate glycol-modified). It is a clear amorphous thermoplastic that can be injection-molded, sheet-extruded or extruded as filament for 3D printing. PETG can be colored during processing. Isophthalic acid Another common modifier is isophthalic acid, replacing some of the 1,4-(para-) linked terephthalate units. The 1,2-(ortho-) or 1,3-(meta-) linkage produces an angle in the chain, which also disturbs crystallinity. Advantages Such copolymers are advantageous for certain molding applications, such as thermoforming, which is used for example to make tray or blister packaging from co-PET film, or amorphous PET sheet (A-PET/PETA) or PETG sheet. On the other hand, crystallization is important in other applications where mechanical and dimensional stability are important, such as seat belts. For PET bottles, the use of small amounts of isophthalic acid, CHDM, diethylene glycol (DEG) or other comonomers can be useful: if only small amounts of comonomers are used, crystallization is slowed but not prevented entirely. As a result, bottles are obtainable via stretch blow molding ("SBM"), which are both clear and crystalline enough to be an adequate barrier to aromas and even gases, such as carbon dioxide in carbonated beverages. Production Polyethylene terephthalate is produced largely from purified terephthalic acid (PTA), as well as to a lesser extent from (mono-)ethylene glycol (MEG) and dimethyl terephthalate (DMT). As of 2022, ethylene glycol is made from ethene found in natural gas, while terephthalic acid comes from p-xylene made from crude oil. Typically an antimony or titanium compound is used as a catalyst, a phosphite is added as a stabilizer and a bluing agent such as cobalt salt is added to mask any yellowing. Processes Dimethyl terephthalate process In the dimethyl terephthalate (DMT) process, DMT and excess ethylene glycol (MEG) are transesterified in the melt at 150–200 °C with a basic catalyst. Methanol (CH3OH) is removed by distillation to drive the reaction forward. Excess MEG is distilled off at higher temperature with the aid of vacuum. The second transesterification step proceeds at 270–280 °C, with continuous distillation of MEG as well. The reactions can be summarized as follows: First step C6H4(CO2CH3)2 + 2 HOCH2CH2OH → C6H4(CO2CH2CH2OH)2 + 2 CH3OH Second step n C6H4(CO2CH2CH2OH)2 → [(CO)C6H4(CO2CH2CH2O)]n + n HOCH2CH2OH Terephthalic acid process In the terephthalic acid process, MEG and PTA are esterified directly at moderate pressure (2.7–5.5 bar) and high temperature (220–260 °C). Water is eliminated in the reaction, and it is also continuously removed by distillation: n C6H4(CO2H)2 + n HOCH2CH2OH → [(CO)C6H4(CO2CH2CH2O)]n + 2n H2O Bio-PET Bio-PET is the bio-based counterpart of PET. Essentially in Bio-PET, the MEG is manufactured from ethylene derived from sugar cane ethanol. A better process based on oxidation of ethanol has been proposed, and it is also technically possible to make PTA from readily available bio-based furfural. Bottle processing equipment There are two basic molding methods for PET bottles, one-step and two-step. In two-step molding, two separate machines are used. The first machine injection molds the preform, which resembles a test tube, with the bottle-cap threads already molded into place. The body of the tube is significantly thicker, as it will be inflated into its final shape in the second step using stretch blow molding. In the second step, the preforms are heated rapidly and then inflated against a two-part mold to form them into the final shape of the bottle. Preforms (uninflated bottles) are now also used as robust and unique containers themselves; besides novelty candy, some Red Cross chapters distribute them as part of the Vial of Life program to homeowners to store medical history for emergency responders. The two-step process lends itself to third party production remote from the user site. The preforms can be transported and stored by the thousand in a much smaller space than would finished containers, for the second stage to be carried out on the user site on a 'just in time' basis. In one-step machines, the entire process from raw material to finished container is conducted within one machine, making it especially suitable for molding non-standard shapes (custom molding), including jars, flat oval, flask shapes, etc. Its greatest merit is the reduction in space, product handling and energy, and far higher visual quality than can be achieved by the two-step system. Degradation PET is subject to degradation during processing. If the moisture level is too high, hydrolysis will reduce the molecular weight by chain scission, resulting in brittleness. If the residence time and/or melt temperature (temperature at melting) are too high, then thermal degradation or thermooxidative degradation will occur resulting in discoloration and reduced molecular weight, as well as the formation of acetaldehyde, and the formation "gel" or "fish-eye" formations through cross-linking. Mitigation measures include copolymerisation with other monomers like CHDM or isophthalic acid, which lower the melting point and thus the melt temperature of the resin, as well as the addition of polymer stabilisers such as phosphites. Acetaldehyde Acetaldehyde, which can form by degradation of PET after mishandling of the material, is a colorless, volatile substance with a fruity smell. Although it forms naturally in some fruit, it can cause an off-taste in bottled water. As well as high temperatures (PET decomposes above 300 °C or 570 °F) and long barrel residence times, high pressures and high extruder speeds (which cause shear raising the temperature), can also contribute to the production of acetaldehyde. Photo-oxidation can also cause the gradual formation acetaldehyde over the object's lifespan. This proceeds via a Type II Norrish reaction. When acetaldehyde is produced, some of it remains dissolved in the walls of a container and then diffuses into the product stored inside, altering the taste and aroma. This is not such a problem for non-consumables (such as shampoo), for fruit juices (which already contain acetaldehyde), or for strong-tasting drinks like soft drinks. For bottled water, however, low acetaldehyde content is quite important, because if nothing masks the aroma, even extremely low concentrations (10–20 parts per billion in the water) of acetaldehyde can produce an off-taste. Safety and environmental concerns Commentary published in Environmental Health Perspectives in April 2010 suggested that PET might yield endocrine disruptors under conditions of common use and recommended research on this topic. Proposed mechanisms include leaching of phthalates as well as leaching of antimony. An article published in Journal of Environmental Monitoring in April 2012 concludes that antimony concentration in deionized water stored in PET bottles stays within EU's acceptable limit even if stored briefly at temperatures up to 60 °C (140 °F), while bottled contents (water or soft drinks) may occasionally exceed the EU limit after less than a year of storage at room temperature. Antimony Antimony (Sb) is a metalloid element that is used as a catalyst in the form of compounds such as antimony trioxide (Sb2O3) or antimony triacetate in the production of PET. After manufacturing, a detectable amount of antimony can be found on the surface of the product. This residue can be removed with washing. Antimony also remains in the material itself and can, thus, migrate out into food and drinks. Exposing PET to boiling or microwaving can increase the levels of antimony significantly, possibly above US EPA maximum contamination levels. The drinking water limit assessed by WHO is 20 parts per billion (WHO, 2003), and the drinking water limit in the United States is 6 parts per billion. Although antimony trioxide is of low toxicity when taken orally, its presence is still of concern. The Swiss Federal Office of Public Health investigated the amount of antimony migration, comparing waters bottled in PET and glass: The antimony concentrations of the water in PET bottles were higher, but still well below the allowed maximum concentration. The Swiss Federal Office of Public Health concluded that small amounts of antimony migrate from the PET into bottled water, but that the health risk of the resulting low concentrations is negligible (1% of the "tolerable daily intake" determined by the WHO). A later (2006) but more widely publicized study found similar amounts of antimony in water in PET bottles. The WHO has published a risk assessment for antimony in drinking water. Fruit juice concentrates (for which no guidelines are established), however, that were produced and bottled in PET in the UK were found to contain up to 44.7 μg/L of antimony, well above the EU limits for tap water of 5 μg/L. Shed microfibres Clothing sheds microfibres in use, during washing and machine drying. Plastic litter slowly forms small particles. Microplastics which are present on the bottom of the river or seabed can be ingested by small marine life, thus entering the food chain. As PET has a higher density than water, a significant amount of PET microparticles may be precipitated in sewage treatment plants. PET microfibers generated by apparel wear, washing or machine drying can become airborne, and be dispersed into fields, where they are ingested by livestock or plants and end up in the human food supply. SAPEA have declared that such particles 'do not pose a widespread risk'. PET is known to degrade when exposed to sunlight and oxygen. As of 2016, scarce information exists regarding the life-time of the synthetic polymers in the environment. Polyester recycling While most thermoplastics can, in principle, be recycled, PET bottle recycling is more practical than many other plastic applications because of the high value of the resin and the almost exclusive use of PET for widely used water and carbonated soft drink bottling. PET bottles lend themselves well to recycling (see below). In many countries PET bottles are recycled to a substantial degree, for example about 75% in Switzerland. The term rPET is commonly used to describe the recycled material, though it is also referred to as R-PET or post-consumer PET (POSTC-PET). The prime uses for recycled PET are polyester fiber, strapping, and non-food containers. Because of the recyclability of PET and the relative abundance of post-consumer waste in the form of bottles, PET is also rapidly gaining market share as a carpet fiber. PET, like many plastics, is also an excellent candidate for thermal disposal (incineration), as it is composed of carbon, hydrogen, and oxygen, with only trace amounts of catalyst elements (but no sulfur). In general, PET can either be chemically recycled into its original raw materials (PTA, DMT, and EG), destroying the polymer structure completely; mechanically recycled into a different form, without destroying the polymer; or recycled in a process that includes transesterification and the addition of other glycols, polyols, or glycerol to form a new polyol. The polyol from the third method can be used in polyurethane (PU foam) production, or epoxy-based products, including paints. In 2023 a process was announced for using PET as the basis for supercapacitor production. PET, being stoichiometrically carbon and , can be turned into a form of carbon containing sheets and nanospheres, with a very high surface area. The process involves holding a mixture of PET, water, nitric acid, and ethanol at a high temperature and pressure for eight hours, followed by centrifugation and drying. Significant investments were announced in 2021 and 2022 for chemical recycling of PET by glycolysis, methanolysis, and enzymatic recycling to recover monomers. Initially these will also use bottles as feedstock but it is expected that fibres will also be recycled this way in future. PET is also a desirable fuel for waste-to-energy plants, as it has a high calorific value which helps to reduce the use of primary resources for energy generation. Biodegradation At least one species of bacterium in the genus Nocardia can degrade PET with an esterase enzyme. Esterases are enzymes able to cleave the ester bond between two oxygens that links subunits of PET. The initial degradation of PET can also be achieved esterases expressed by Bacillus, as well as Nocardia. Japanese scientists have isolated another bacterium, Ideonella sakaiensis, that possesses two enzymes which can break down the PET into smaller pieces digestible by the bacteria. A colony of I. sakaiensis can disintegrate a plastic film in about six weeks. French researchers report developing an improved PET hydrolase that can depolymerize (break apart) at least 90 percent of PET in 10 hours, breaking it down into individual monomers. Also, an enzyme based on a natural PET-ase was designed with the help of a machine learning algorithm to be able to tolerate pH and temperature changes by the University of Texas at Austin. The PET-ase was found to able to degrade various products and could break them down as fast as 24 hours.
Physical sciences
Polymers
Chemistry
292984
https://en.wikipedia.org/wiki/Lahar
Lahar
A lahar (, from ) is a violent type of mudflow or debris flow composed of a slurry of pyroclastic material, rocky debris and water. The material flows down from a volcano, typically along a river valley. Lahars are often extremely destructive and deadly; they can flow tens of metres per second, they have been known to be up to deep, and large flows tend to destroy any structures in their path. Notable lahars include those at Mount Pinatubo in the Philippines and Nevado del Ruiz in Colombia, the latter of which killed more than 20,000 people in the Armero tragedy. Etymology The word lahar is of Javanese origin. Berend George Escher introduced it as a geological term in 1922. Description The word lahar is a general term for a flowing mixture of water and pyroclastic debris. It does not refer to a particular rheology or sediment concentration. Lahars can occur as normal stream flows (sediment concentration of less than 30%), hyper-concentrated stream flows (sediment concentration between 30 and 60%), or debris flows (sediment concentration exceeding 60%). Indeed, the rheology and subsequent behaviour of a lahar may vary in place and time within a single event, owing to changes in sediment supply and water supply. Lahars are described as 'primary' or 'syn-eruptive' if they occur simultaneously with or are triggered by primary volcanic activity. 'Secondary' or 'post-eruptive' lahars occur in the absence of primary volcanic activity, e.g. as a result of rainfall during pauses in activity or during dormancy. In addition to their variable rheology, lahars vary considerably in magnitude. The Osceola Lahar produced by Mount Rainier in modern-day Washington some 5600 years ago resulted in a wall of mud deep in the White River canyon and covered an area of over , for a total volume of . A debris-flow lahar can erase virtually any structure in its path, while a hyperconcentrated-flow lahar is capable of carving its own pathway, destroying buildings by undermining their foundations. A hyperconcentrated-flow lahar can leave even frail huts standing, while at the same time burying them in mud, which can harden to near-concrete hardness. A lahar's viscosity decreases the longer it flows and can be further thinned by rain, producing a quicksand-like mixture that can remain fluidized for weeks and complicate search and rescue. Lahars vary in speed. Small lahars less than a few metres wide and several centimetres deep may flow a few metres per second. Large lahars hundreds of metres wide and tens of metres deep can flow several tens of metres per second (22 mph or more), much too fast for people to outrun. On steep slopes, lahar speeds can exceed . A lahar can cause catastrophic destruction along a potential path of more than . Lahars from the 1985 Nevado del Ruiz eruption in Colombia caused the Armero tragedy, burying the city of Armero under of mud and debris and killing an estimated 23,000 people. A lahar caused New Zealand's Tangiwai disaster, where 151 people died after a Christmas Eve express train fell into the Whangaehu River in 1953. Lahars have caused 17% of volcano-related deaths between 1783 and 1997. Trigger mechanisms Lahars have several possible causes: Snow and glaciers can be melted by lava or pyroclastic surges during an eruption. Lava can erupt from open vents and mix with wet soil, mud or snow on the slope of the volcano making a very viscous, high energy lahar. The higher up the slope of the volcano, the more gravitational potential energy the flows will have. A flood caused by a glacier, lake breakout, or heavy rainfalls can generate lahars, also called glacier run or jökulhlaup. Water from a crater lake can combine with volcanic material in an eruption. Heavy rainfall can mobilize unconsolidated pyroclastic deposits. In particular, although lahars are typically associated with the effects of volcanic activity, lahars can occur even without any current volcanic activity, as long as the conditions are right to cause the collapse and movement of mud originating from existing volcanic ash deposits. Snow and glaciers can melt during periods of mild to hot weather. Earthquakes underneath or close to the volcano can shake material loose and cause it to collapse, triggering a lahar avalanche. Rainfall can cause the still-hanging slabs of solidified mud to come rushing down the slopes at a speed of more than , causing devastating results. Places at risk Several mountains in the world – including Mount Rainier in the United States, Mount Ruapehu in New Zealand, and Merapi and Galunggung in Indonesia – are considered particularly dangerous due to the risk of lahars. Several towns in the Puyallup River valley in Washington state, including Orting, are built on top of lahar deposits that are only about 500 years old. Lahars are predicted to flow through the valley every 500 to 1,000 years, so Orting, Sumner, Puyallup, Fife, and the Port of Tacoma face considerable risk. The USGS has set up lahar warning sirens in Pierce County, Washington, so that people can flee an approaching debris flow in the event of a Mount Rainier eruption. A lahar warning system has been set up at Mount Ruapehu by the New Zealand Department of Conservation and hailed as a success after it successfully alerted officials to an impending lahar on 18 March 2007. Since mid-June 1991, when violent eruptions triggered Mount Pinatubo's first lahars in 500 years, a system to monitor and warn of lahars has been in operation. Radio-telemetered rain gauges provide data on rainfall in lahar source regions, acoustic flow monitors on stream banks detect ground vibration as lahars pass, and staffed watchpoints further confirm that lahars are rushing down Pinatubo's slopes. This system has enabled warnings to be sounded for most but not all major lahars at Pinatubo, saving hundreds of lives. Physical preventative measures by the Philippine government were not adequate to stop over of mud from flooding many villages around Mount Pinatubo from 1992 through 1998. Scientists and governments try to identify areas with a high risk of lahars based on historical events and computer models. Volcano scientists play a critical role in effective hazard education by informing officials and the public about realistic hazard probabilities and scenarios (including potential magnitude, timing, and impacts); by helping evaluate the effectiveness of proposed risk-reduction strategies; by helping promote acceptance of (and confidence in) hazards information through participatory engagement with officials and vulnerable communities as partners in risk reduction efforts; and by communicating with emergency managers during extreme events. An example of such a model is TITAN2D. These models are directed towards future planning: identifying low-risk regions to place community buildings, discovering how to mitigate lahars with dams, and constructing evacuation plans. Examples Nevado del Ruiz In 1985, the volcano Nevado del Ruiz erupted in central Colombia. As pyroclastic flows erupted from the volcano's crater, they melted the mountain's glaciers, sending four enormous lahars down its slopes at . The lahars picked up speed in gullies and coursed into the six major rivers at the base of the volcano; they engulfed the town of Armero, killing more than 20,000 of its almost 29,000 inhabitants. Casualties in other towns, particularly Chinchiná, brought the overall death toll to over 25,000. Footage and photographs of Omayra Sánchez, a young victim of the tragedy, were published around the world. Other photographs of the lahars and the impact of the disaster captured attention worldwide and led to controversy over the degree to which the Colombian government was responsible for the disaster. Mount Pinatubo Lahars caused most of the deaths of the 1991 eruption of Mount Pinatubo. The initial eruption killed six people, but the lahars killed more than 1500. The eye of Typhoon Yunya passed over the volcano during its eruption on 15 June 1991, and the resulting rain triggered the flow of volcanic ash, boulders, and water down rivers surrounding the volcano. Angeles City in Pampanga and neighbouring cities and towns were damaged by lahars when Sapang Balen Creek and the Abacan River became channels for mudflows and carried them to the heart of the city and surrounding areas. Over of mud inundated and damaged the towns of Castillejos, San Marcelino and Botolan in Zambales, Porac and Mabalacat in Pampanga, Tarlac City, Capas, Concepcion and Bamban in Tarlac. The Bamban Bridge on the MacArthur Highway, a major north–south transportation route, was destroyed, and temporary bridges erected in its place were inundated by subsequent lahars. On the morning of 1 October 1995, pyroclastic material which clung to the slopes of Pinatubo and surrounding mountains rushed down because of heavy rain, and turned into an lahar. This mudflow killed at least 100 people in Barangay Cabalantian in Bacolor. The Philippine government under President Fidel V. Ramos ordered the construction of the FVR Mega Dike in an attempt to protect people from further mudflows. Typhoon Reming triggered additional lahars in the Philippines in 2006.
Physical sciences
Volcanology
Earth science
293065
https://en.wikipedia.org/wiki/Desiccant
Desiccant
A desiccant is a hygroscopic substance that is used to induce or sustain a state of dryness (desiccation) in its vicinity; it is the opposite of a humectant. Commonly encountered pre-packaged desiccants are solids that absorb water. Desiccants for specialized purposes may be in forms other than solid, and may work through other principles, such as chemical bonding of water molecules. They are commonly encountered in foods to retain crispness. Industrially, desiccants are widely used to control the level of water in gas streams. Types of desiccants Although some desiccants are chemically inert, others are extremely reactive and require specialized handling techniques. The most common desiccant is silica gel, an otherwise inert, nontoxic, water-insoluble white solid. Tens of thousands of tons are produced annually for this purpose. Other common desiccants include activated charcoal, calcium sulfate, calcium chloride, and molecular sieves (typically, zeolites). Desiccants may also be categorized by their type, either I, II, III, IV, or V. These types are a function of the shape of the desiccant's moisture sorption isotherm. Alcohols and acetones are also dehydrating agents. Diethylene glycol is an important industrial desiccant. It absorbs water from natural gas, minimizing the formation of methane hydrates, which can block pipes. Performance efficiency One measure of desiccant efficiency is the ratio (or percentage) of water storable in the desiccant relative to the mass of desiccant. Another measure is the residual relative humidity of the air or other fluid being dried. For drying gases, a desiccant's performance can be precisely described by the dew point of the dried product. Colored saturation indicators Sometimes a humidity indicator is included in the desiccant to show, by color changes, the degree of water-saturation of the desiccant. One commonly used indicator is cobalt chloride (), which is blue when anhydrous, but turns purple upon bonding with two water molecules (·2). Further hydration results in the pink hexaaquacobalt(II) chloride complex . However, the use of cobalt chloride raises health concerns, being potentially carcinogenic. Applications Applications of desiccants are dominated by the petrochemical industry. Hydrocarbons, including natural gas, often must be anhydrous or nearly so for processing or for transport. Catalysts that are used to convert some petroleum fractions are generally deactivated by even traces of water. Natural gas tends to form solid methane hydrates which can block pipes. Domestic uses One example of desiccant usage is in the manufacture of insulated windows where zeolite spheroids fill a rectangular spacer tube at the perimeter of the panes of glass. The desiccant helps to prevent the condensation of moisture between the panes. Another use of zeolites is in the "dryer" component of refrigeration systems to absorb water carried by the refrigerant, whether residual water left over from the construction of the system, or water released by the degradation of other materials over time. Bagged desiccants are also commonly used to protect goods in barrier-sealed shipping containers against moisture damage: rust, corrosion, etc. Hygroscopic cargo, such as cocoa, coffee, various nuts and grains, and other foods can be particularly susceptible to mold and rot when exposed to condensation and humidity. Because of this, shippers often take measures by deploying desiccants to protect against loss. Pharmaceutical packaging often includes small packets of desiccant to keep the atmosphere inside the package below critical levels of water vapor. Air conditioning systems can be based on desiccants, as drier air feels more comfortable and absorbing water itself removes heat. Desiccants are used in livestock farming, where, for example, new-born piglets are highly susceptible to hypothermia owing to their wetness. Laboratory uses Desiccants are also used to remove water from solvents. Drying generally involves mixing the solvent with the solid desiccant. Molecular sieves are superior as desiccants relative to chemical drying reagents such as sodium-benzophenone. Sieves offer the advantages of being safe in air and recyclable.
Physical sciences
Phase separations
Chemistry
293079
https://en.wikipedia.org/wiki/Actuator
Actuator
An actuator is a component of a machine that produces force, torque, or displacement, when an electrical, pneumatic or hydraulic input is supplied to it in a system (called an actuating system). The effect is usually produced in a controlled way. An actuator translates such an input signal into the required form of mechanical energy. It is a type of transducer. In simple terms, it is a "mover". An actuator requires a control device (which provides control signal) and a source of energy. The control signal is relatively low in energy and may be voltage, electric current, pneumatic, or hydraulic fluid pressure, or even human power. In the electric, hydraulic, and pneumatic sense, it is a form of automation or automatic control. The displacement achieved is commonly linear or rotational, as exemplified by linear motors and rotary motors, respectively. Rotary motion is more natural for small machines making large displacements. By means of a leadscrew, rotary motion can be adapted to function as a linear actuator (which produces a linear motion, but is not a linear motor). Another broad classification of actuators separates them into two types: incremental-drive actuators and continuous-drive actuators. Stepper motors are one type of incremental-drive actuators. Examples of continuous-drive actuators include DC torque motors, induction motors, hydraulic and pneumatic motors, and piston-cylinder drives (rams). Types of actuators Mechanical An actuator can be just a mechanism that is directly driven by the motions or forces of other parts of the system. An example is the camshafts that drive the intake and exhaust valves in internal combustion engines, driven by the engine itself. Another example is the mechanism that strikes the hours in a traditional grandfather clock or cuckoo clock. Hydraulic A hydraulic actuator typically uses the pressure of a liquid (usually oil) to cause a piston to slide inside a hollow cylindrical tube linear, rotatory or oscillatory motion. In a single acting actuator the fluid pressure is applied to just one side of the piston, so that it applies useful force in only one direction. The opposite motion may be affected by a spring, by gravity, or by other forces present in the system. In a double acting actuator, the return stroke is driven by fluid pressure applied to the opposite side of the piston. Since liquids are nearly impossible to compress, a hydraulic actuator can exert a large force. The drawback of this approach is its limited acceleration. They respond quickly to input changes, have little inertia, can operate continuously over a relatively large working range, and can hold their position without any significant energy input. A hydraulic actuator can be used to displace the rack of a rack and pinion mechanism, causing the pinion to turn. This arrangement is used, for example, to operate valves in pipelines and other industrial fluid transport installations. Pneumatic A pneumatic actuator is similar to a hydraulic one but uses a gas (usually air) instead of a liquid. Compared to hydraulic actuators, pneumatic ones are less complicated because they do not need pipes for the return and recycling of the working fluid. On the other hand, they still need external infrastructure such as compressors, reservoirs, filters, and air treatment subsystems, which often makes them less convenient that electrical and electromechanical actuators. In the first steam engines and in all steam locomotives, steam pressure is used to drive pneumatic actuators to produce a reciprocating motion, which is converted to rotary motion by some sort of crankshaft mechanism. Electric Since 1960, several actuator technologies have been developed. Electric actuators can be classified in the following groups: Electromechanical An electromechanical actuator (EMA) uses mechanical means to convert the rotational force of an ordinary (rotary) electric motor into a linear movement. The mechanism may be a toothed belt or a screw (either a ball or a lead screw or planetary roller screw). The main advantages of electromechanical actuators are their relatively good level of accuracy with respect to pneumatics, their possible long lifecycle and the little maintenance effort required (might require grease). It is possible to reach relatively high force, on the order of 100 kN. The main limitation of these actuators are the reachable speed, the important dimensions and weight they require. The main application of such actuators is mainly seen in health care devices and factory automation. Electrohydraulic Another approach is an electrohydraulic actuator, where the electric motor remains the prime mover but provides torque to operate a hydraulic accumulator that is then used to transmit actuation force in much the same way that diesel engine/hydraulics are typically used in heavy equipment. Electrical energy is used to actuate equipment such as multi-turn valves, or electric-powered construction and excavation equipment. When used to control the flow of fluid through a valve, a brake is typically installed above the motor to prevent the fluid pressure from forcing open the valve. If no brake is installed, the actuator gets activated to reclose the valve, which is slowly forced open again. This sets up an oscillation (open, close, open ...) and the motor and actuator will eventually become damaged. Rotary Electric rotary actuators use a rotary motor to turn the target part over a certain angle. Rotary actuators can have up to a rotation of 360 degrees. This allows it to differ from a linear motor as the linear is bound to a set distance compared to the rotary motor. Rotary motors have the ability to be set at any given degree in a field making the device easier to set up still with durability and a set torque. Rotary motors can be powered by 3 different techniques such as Electric, Fluid, or Manual. However, Fluid powered rotary actuators have 5 sub-sections of actuators such as Scotch Yoke, Vane, Rack-and-Pinion, Helical, and Electrohydraulic. All forms have their own specific design and use allowing the ability to choose multiple angles of degree. Applications for the rotary actuators are just about endless but, will more than likely be found dealing with mostly hydraulic pressured devices and industries. Rotary actuators are even used in the robotics field when seeing robotic arms in industry lines. Anything you see that deals with motion control systems to perform a task in technology is a good chance to be a rotary actuator. Linear A linear electric actuator uses a linear motor, which can be thought as a rotary electric motor which has been cut and unrolled. Thus, instead of producing a rotational movement, it produces a linear force along their length. Because it generally has lower friction losses than the alternatives, a linear electric actuator can last over a hundred million cycles. Linear motors are divided in 3 basic categories: flat linear motor (classic), U-Channel linear motors and Tubular linear motors. Linear motor technology is the best solution in the context of a low load (up to 30Kgs) because it provides the highest level of speed, control and accuracy. In fact, it represents the most desired and versatile technology. Due to the limitations of pneumatics, the current electric actuator technology is a viable solution for specific industry applications and it has been successfully introduced in market segments such as the watchmaking, semiconductor and pharmaceutical industries (as high as 60% of the applications. The growing interest for this technology, can be explained by the following characteristics: High precision (equal or less than 0,1 mm); High cycling rate (greater than 100 cycles/min); Possible usage in clean and highly-regulated environments (no leakages of air, humidity or lubricants allowed);  Need for programmable motion in the situation of complex operations The main disadvantages of linear motors are: They are expensive respect to pneumatics and other electric technologies. They are not easy to integrate in standard machineries due to their important size and high weight. They have a low force density respect to pneumatic and electromechanical actuators. Thermal An actuator may be driven by heat through the expansion that most solid material exhibit when the temperature increases. This principle is commonly used, for example, to operate electric switches in thermostats. Typically, a (non-electronic) thermostat contains a strip with two layers of different metals, that will bend when heated. Thermal actuators may also exploit the properties of shape-memory alloys. Magnetic Some actuators are driven by externally applied magnetic fields. They typically contain parts made of ferromagnetic materials that are strongly attracted to each other when they are magnetized by the external field. An example are the reed switches that may be used as door opening sensors in a building security system. Alternatively, magnetic actuators can use magnetic shape-memory alloys. Soft actuators A soft actuator is made of a flexible material that changes its shape in response to stimuli including mechanical, thermal, magnetic, and electrical. Soft actuators mainly deal with the robotics of humans rather than industry which is what most of the actuators are used for. For most actuators they are mechanically durable yet do not have an ability to adapt compared to soft actuators. The soft actuators apply to mainly safety and healthcare for humans which is why they are able to adapt to environments by disassembling their parts. This is why the driven energy behind soft actuators deal with flexible materials like certain polymers and liquids that are harmless The majority of the existing soft actuators are fabricated using multistep low yield processes such as micro-moulding, solid freeform fabrication, and mask lithography. However, these methods require manual fabrication of devices, post processing/assembly, and lengthy iterations until maturity in the fabrication is achieved. To avoid the tedious and time-consuming aspects of the current fabrication processes, researchers are exploring an appropriate manufacturing approach for effective fabrication of soft actuators. Therefore, special soft systems that can be fabricated in a single step by rapid prototyping methods, such as 3D printing, are utilized to narrow the gap between the design and implementation of soft actuators, making the process faster, less expensive, and simpler. They also enable incorporation of all actuator components into a single structure eliminating the need to use external joints, adhesives, and fasteners. Shape memory polymer (SMP) actuators are the most similar to our muscles, providing a response to a range of stimuli such as light, electrical, magnetic, heat, pH, and moisture changes. They have some deficiencies including fatigue and high response time that have been improved through the introduction of smart materials and combination of different materials by means of advanced fabrication technology. The advent of 3D printers has made a new pathway for fabricating low-cost and fast response SMP actuators. The process of receiving external stimuli like heat, moisture, electrical input, light or magnetic field by SMP is referred to as shape memory effect (SME). SMP exhibits some rewarding features such a low density, high strain recovery, biocompatibility, and biodegradability. Photopolymers or light activated polymers (LAP) are another type of SMP that are activated by light stimuli. The LAP actuators can be controlled remotely with instant response and, without any physical contact, only with the variation of light frequency or intensity. A need for soft, lightweight and biocompatible soft actuators in soft robotics has influenced researchers for devising pneumatic soft actuators because of their intrinsic compliance nature and ability to produce muscle tension. Polymers such as dielectric elastomers (DE), ionic polymer–metal composites (IPMC), ionic electroactive polymers, polyelectrolyte gels, and gel-metal composites are common materials to form 3D layered structures that can be tailored to work as soft actuators. EAP actuators are categorized as 3D printed soft actuators that respond to electrical excitation as deformation in their shape. Examples and applications In engineering, actuators are frequently used as mechanisms to introduce motion, or to clamp an object so as to prevent motion. In electronic engineering, actuators are a subdivision of transducers. They are devices which transform an input signal (mainly an electrical signal) into some form of motion. Examples of actuators Comb drive Digital micromirror device Electric motor Electroactive polymer Hydraulic cylinder Piezoelectric actuator Plasma actuator Pneumatic actuator Screw jack Servomechanism Solenoid Stepper motor Shape-memory alloy Thermal bimorph Hydraulic actuators Trim actuator, in aircraft design Circular to linear conversion Motors are mostly used when circular motions are needed, but can also be used for linear applications by transforming circular to linear motion with a lead screw or similar mechanism. On the other hand, some actuators are intrinsically linear, such as piezoelectric actuators. Conversion between circular and linear motion is commonly made via a few simple types of mechanism including: Screw: Screw jack, ball screw and roller screw actuators all operate on the principle of the simple machine known as the screw. By rotating the actuator's nut, the screw shaft moves in a line. By moving the screw shaft, the nut rotates. Wheel and axle: Hoist, winch, rack and pinion, chain drive, belt drive, rigid chain and rigid belt actuators operate on the principle of the wheel and axle. By rotating a wheel/axle (e.g. drum, gear, pulley or shaft) a linear member (e.g. cable, rack, chain or belt) moves. By moving the linear member, the wheel/axle rotates. Virtual instrumentation In virtual instrumentation, actuators and sensors are the hardware complements of virtual instruments. Performance metrics Performance metrics for actuators include speed, acceleration, and force (alternatively, angular speed, angular acceleration, and torque), as well as energy efficiency and considerations such as mass, volume, operating conditions, and durability, among others. Force When considering force in actuators for applications, two main metrics should be considered. These two are static and dynamic loads. Static load is the force capability of the actuator while not in motion. Conversely, the dynamic load of the actuator is the force capability while in motion. Speed Speed should be considered primarily at a no-load pace, since the speed will invariably decrease as the load amount increases. The rate the speed will decrease will directly correlate with the amount of force and the initial speed. Operating conditions Actuators are commonly rated using the standard IP Code rating system. Those that are rated for dangerous environments will have a higher IP rating than those for personal or common industrial use. Durability This will be determined by each individual manufacturer, depending on usage and quality.
Technology
Components
null
293265
https://en.wikipedia.org/wiki/Noctuidae
Noctuidae
The Noctuidae, commonly known as owlet moths, cutworms or armyworms, are a family of moths. Taxonomically, they are considered the most controversial family in the superfamily Noctuoidea because many of the clades are constantly changing, along with the other families of the Noctuoidea. It was considered the largest family in Lepidoptera for a long time, but after regrouping Lymantriinae, Catocalinae and Calpinae within the family Erebidae, the latter holds this title now. Currently, Noctuidae is the second largest family in Noctuoidea, with about 1,089 genera and 11,772 species. This classification is still contingent, as more changes continue to appear between Noctuidae and Erebidae. Description Adult: Most noctuid adults have wings with a variety of shades of browns, grays, and other varied shades and colourations but some subfamilies, such as Acronictinae and Agaristinae, are very colorful, especially those from tropical regions (e.g. Baorisa hieroglyphica). They are characterized by a structure in the metathorax called the nodular sclerite or epaulette, which separates the tympanum and the conjunctiva in the tympanal organ. It functions to keep parasites (Acari) out of the tympanal cavity. Another characteristic in this group is trifine hindwing venation, by reduction or absence of the second medial vein (M2). Markings present on the wings of noctuid adults can be helpful in distinguishing species. From the basal location to the outer edge (proximal to distal) on the forewing, there is a claviform (club-shaped) stigma, horizontally oriented with the thicker end closer to the wing's outer edge, located posterior to a discal (round) stigma. These are followed distally by a reniform (kidney-shaped) stigma, which is typically oriented with its concave side facing the wing's outer edge. It is often not possible to discern all of the stigmata on all specimens or species. Crossbands or crosslines may be present, oriented longitudinally from the leading to the trailing edge of the wing. Larva: Commonly green or brown; some species present bright colors, such as the camphorweed cucullia moth (Cucullia alfarata). Most are pudgy and smooth with rounded short heads and few setae, but there are some exceptions in some subfamilies (e.g. Acronictinae and Pantheinae). Pupa: The pupae most often range from shiny brown to dark brown. When they newly pupate they are bright brownish orange, but after a few days start to get darker. Eggs: Vary in colors, but all have a spherical shape. Etymology The word Noctuidae is derived from the name of the type genus Noctua, which is the Latin name for the little owl, and the patronymic suffix -idae used typically to form taxonomic family names in animals. The common name "owlet" originally means a small or young owl. The names "armyworms" and "cutworms" are based on the behavior of the larvae of this group, which can occur in destructive swarms and cut the stems of plants. Ecology Distribution and diversity This family is cosmopolitan and can be found worldwide except in the Antarctic region. Some species, such as the setaceous Hebrew character (Xestia c-nigrum), can be found in the Arctic Circle, specifically in the Yukon territory of western Canada, with an elevation 1,702 m above sea level, where the temperature fluctuates between 23/-25 °C (73/-13 °F). Many species of dart moths have been recorded in elevations as high as 4,000 m above sea level (e.g. Xestia elisabetha). Among the places where the number of species has been counted are North America and northern Mexico, with about 2,522 species. 1,576 species are found in Europe, while the other species are distributed worldwide. Mutualism Members of Noctuidae, like other butterflies and moths, perform an important role in plant pollination. Some species have developed a stronger connection with their host plants. For example, the lychnis moth (Hadena bicruris) has a strange mutualistic relationship with pink plants or carnation plants (Caryophyllaceae), in that larvae feed on the plant while the adults pollinate the flowers. Food guilds Herbivory: Caterpillars of most Noctuidae feed on plants; some feed on poisonous plants and are unaffected by their chemical defences; for example, the splendid brocade moth (Lacanobia splendens) feeds on cowbane (Cicuta virosa), a plant that is notoriously toxic to vertebrates. Predation and cannibalism: During the larval stage, some cutworms readily feed on other insects. One such species is the shivering pinion (Lithophane querquera), whose larvae commonly feed on other lepidopteran larvae. Moreover, many noctuid larvae, such as those of the fall armyworm (Spodoptera frugiperda) and of genera such as Heliothis and Helicoverpa, aggressively eat their siblings and often other species of caterpillar. Nectarivory and puddling: Like many Lepidoptera, many species of adult Noctuidae visit flowers for their nectar. They also seek other liquid food resources such as plant juices, honeydew, dung, urea and mud, among others. As is common in members of the order Lepidoptera, courtship in many Noctuidae includes a set of movements in which the female evaluates the male's reproductive fitness. Most noctuid moths produce pheromones that attract the opposite sex. Female pheromones that attract males occur widely and have long been studied, but the study of male pheromones has further to go. Reproduction Noctuid moths commonly begin the reproductive season from spring to fall, and mostly are multivoltine, such as the eastern panthea moth (Panthea furcilla), which reproduces over the year. Nevertheless, some species have just one brood of offspring (univoltine); among the best known is the lesser yellow underwing (Noctua comes). Defense This group has a wide range of both chemical and physical defenses. Among the chemical defenses three types stand out. First, the pyrrolizidine alkaloid sequestration usually present in Arctiinae is also found in a few species of noctuids, including the Spanish moth (Xanthopastis timais). Another chemical defense is formic acid production, which was thought to be present only in Notodontidae, but later was found in caterpillars of Trachosea champa. Finally, the last type of chemical defense is regurgitation of plant compounds, often used by many insects, but the cabbage palm caterpillar (Litoprosopus futilis) produces a toxin called toluquinone that deters predators. On the other hand, the main physical defense in caterpillars and adults alike is mimicry. Most owlet moths have drab colors with a variety of patterns suitable to camouflage their bodies. The second physical defense consists in thousands of secondary setae that surround the body. The subfamilies that present this mechanism are Pantheinae and Acronictinae. The third is aposematism, represented by species of Cucullinae. Finally, all adults have another mechanism for defense: a tympanal organ available to hear the echolocation spread out by bats, so the moths can avoid them. Human importance Agriculture Many species of owlet moths are considered an agricultural problem around the world. Their larvae are typically known as "cutworms" or "armyworms" due to enormous swarms that destroy crops, orchards and gardens every year. The Old World bollworm (Helicoverpa armigera) produces losses in agriculture every year that exceed US$2 billion. Additionally, the variegated cutworm (Peridroma saucia) is described by many as one of the most damaging pests to vegetables. In West Africa, species including Busseola fusca, Heliocheilus albipunctella, Sesamia calamistis, Helicoverpa armigera, and Spodoptera exempta are major pests of staple crops such as pearl millet, sorghum, and maize. Systematics Since molecular analysis began to play a larger role in systematics, the structure of many Lepidoptera groups has been changing and Noctuidae is not an exception. Most recent studies have shown that Noctuidae sensu stricto is a monophyletic group, mainly based on trifine venation. Some clades within Noctuidae sensu lato have yet to be studied. This taxonomic division represents the subfamilies, tribes and subtribes considered so far. Family Noctuidae Latreille, 1809 Subfamily Acontiinae Guenée, 1841 Tribe Acontiini Guenée, 1841 Tribe Armadini Tribe Chamaecleini Subfamily Acronictinae Harris, 1841 Subfamily Aediinae Subfamily Agaristinae Boisduval, 1833 Subfamily Amphipyrinae Guenée, 1837 Tribe Amphipyrini Guenée, 1837 Tribe Psaphidini Grote, 1896 Subtribe Feraliina Poole, 1995 Subtribe Nocloina Poole, 1995 Subtribe Psaphidina Grote, 1896 Subtribe Triocnemidina Poole, 1995 Subfamily Bagisarinae Crumb, 1956 Subfamily Balsinae Grote, 1896 Subfamily Bryophilinae Guenée, 1852 Subfamily Cobubathinae Wagner & Keegan, 2021 Subfamily Condicinae Poole, 1995 Tribe Condicini Poole, 1995 Tribe Leuconyctini Poole, 1995 Subfamily Cropiinae Keegan & Wagner, 2021 Subfamily Cuculliinae Herrich-Schäffer, 1850 Subfamily Dilobinae Subfamily Dyopsinae Subfamily Eriopinae Herrich-Schäffer, 1851 Subfamily Eucocytiinae Subfamily Eustrotiinae Grote, 1882 Subfamily Grotellinae Subfamily Heliothinae Boisduval, 1828 Subfamily Metoponiinae Herrich-Schäffer, 1851 Tribe Cydosiini Kitching & Rawlins, 1998 Subfamily Noctuinae Latreille, 1809 Tribe Actinotiini Beck, 1996 Tribe Apameini Guenée, 1841 Tribe Arzamini Grote, 1883 Tribe Caradrinini Boisduval, 1840 Subtribe Athetiina Fibiger & Lafontaine, 2005 Subtribe Caradrinina Boisduval, 1840 Tribe Dypterygiini Forbes, 1954 Tribe Elaphriini Beck, 1996 Tribe Episemini Tribe Eriopygini Fibiger & Lafontaine, 2005 Tribe Glottulini Guenée, 1852 Tribe Hadenini Guenée, 1837 Tribe Leucaniini Guenée, 1837 Tribe Noctuini Latreille, 1809 Subtribe Agrotina Harris, 1841 Subtribe Axyliina Subtribe Noctuina Latreille, 1809 Tribe Orthosiini Guenée, 1837 Tribe Phlogophorini Hampson, 1918 Tribe Phosphilini Poole, 1995 Tribe Prodeniini Forbes, 1954 Tribe Pseudeustrotiini Beck, 1996 Tribe Tholerini Beck, 1996 Tribe Xylenini Guenée, 1837 Subtribe Antitypina Forbes & Franclemont, 1954 Subtribe Cosmiina Guenée, 1852 Subtribe Ufeina Crumb, 1956 Subtribe Xylenina Guenée, 1837 Subfamily Oncocnemidinae Forbes & Franclemont, 1954 Subfamily Pantheinae Smith, 1898 Subfamily Plusiinae Boisduval, 1828 Tribe Abrostolini Eichlin & Cunningham, 1978 Tribe Argyrogrammatini Eichlin & Cunningham, 1978 Tribe Plusiini Boisduval, 1828 Subtribe Autoplusiina Kitching, 1987 Subtribe Euchalciina Chou & Lu, 1979 Subtribe Plusiina Boisduval, 1828 Subfamily Raphiinae Subfamily Stiriinae Tribe Annaphilini Tribe Stiriini Grote, 1882 Subtribe Annaphilina Mustelin, 2006 Subtribe Azeniina Poole, 1995 Subtribe Grotellina Poole, 1995 Subtribe Stiriina Grote, 1882 Genera with intervening taxonomy not available include: Alastria Epilitha Fabula Lanatopyga Lenisa Neoligia Orohadena Orthomoia Protapamea Proxenus Pseudluperina
Biology and health sciences
Lepidoptera
Animals
293270
https://en.wikipedia.org/wiki/Heat%20wave
Heat wave
A heat wave or heatwave, sometimes described as extreme heat, is a period of abnormally hot weather generally considered to be at least five consecutive days. A heat wave is usually measured relative to the usual climate in the area and to normal temperatures for the season. The main difficulties with this broad definition emerge when one must quantify what the 'normal' temperature state is, and what the spatial extent of the event may or must be. Temperatures that humans from a hotter climate consider normal can be regarded as a heat wave in a cooler area. This would be the case if the warm temperatures are outside the normal climate pattern for that area. Heat waves have become more frequent, and more intense over land, across almost every area on Earth since the 1950s, the increase in frequency and duration being caused by climate change. Heat waves form when a high-pressure area in the upper atmosphere strengthens and remains over a region for several days up to several weeks. This traps heat near the earth's surface. It is usually possible to forecast heat waves, thus allowing the authorities to issue a warning in advance. Heat waves have an impact on the economy. They can reduce labour productivity, disrupt agricultural and industrial processes and damage infrastructure. Severe heat waves have caused catastrophic crop failures and thousands of deaths from hyperthermia. They have increased the risk of wildfires in areas with drought. They can lead to widespread electricity outages because more air conditioning is used. A heat wave counts as extreme weather. It poses danger to human health, because heat and sunlight overwhelm the thermoregulation in humans. Definitions There are several definitions of heat waves: The IPCC defines heatwave as "a period of abnormally hot weather, often defined with reference to a relative temperature threshold, lasting from two days to months." A definition based on the Heat Wave Duration Index is that a heat wave occurs when the daily maximum temperature of more than five consecutive days exceeds the average maximum temperature by , the normal period being 1961–1990. The same definition is used by the World Meteorological Organization. A definition from the Glossary of Meteorology is: "A period of abnormally and uncomfortably hot and usually humid weather." Marine heatwaves are generally described as prolonged discrete periods of unusually warm sea surface temperatures in a specific region. At this time the most commonly accepted definition is that proposed by Hobday et. al. which refers to an algorithm that uses percentile values for temperatures, and defines a threshold set as the 90th percentile for a given day of the year, above which one can say a marine heatwave is occuring. This definition can be used with temperature data acquired anywhere in the world, allowing for comparisons across different observations and latitudes. Definitions by country Europe Denmark defines a national heat wave (hedebølge) as a period of at least 3 consecutive days in which the average maximum temperature across more than half the country exceeds . The Danish Meteorological Institute also has a definition for a "warmth wave" (varmebølge). It defines this using the same criteria for a temperature. Sweden defines a heat wave as at least five days in a row with a daily high exceeding . In Greece, the Hellenic National Meteorological Service defines a heat wave as occurring over three consecutive days with temperatures at 39 °C (102 °F) or higher. In the same period the minimum temperature is or more. During this period, there are either no winds or only weak winds. These conditions occur in a broad area. The Netherlands defines a heat wave as a period of at least five consecutive days in which the maximum temperature in De Bilt exceeds . During this period the maximum temperature in De Bilt must exceed for at least three days. Belgium also uses this definition of a heat wave with Ukkel as a reference point. So does Luxembourg. In the United Kingdom, the Met Office operates a Heat Health Watch system. This places each Local Authority region into one of four levels. Heat wave conditions occur when the maximum daytime temperature and minimum nighttime temperature rise above the threshold for a particular region. The length of time above that threshold determines the level. Level 1 represents normal summer conditions. Level 2 occurs when there is a 60% or higher risk that the temperature will be above the threshold levels for two days and the intervening night. Level 3 arises when the temperature has been above the threshold for the preceding day and night, and there is a 90% or higher chance that it will stay above the threshold in the following day. Level 4 is triggered if conditions are more severe than those of the preceding three levels. Each of the first three levels gives rise to a particular state of readiness and response by the social and health services. Level 4 involves a more widespread response. The threshold for a heat wave occurs when there are at least three days above across much of the country. Greater London has a threshold of . Other regions In the United States, definitions also vary by region. They usually involve a period of at least two or more days of excessively hot weather. In the Northeast, a heat wave is typically when the temperature reaches or exceeds for three consecutive days. This is not always the case. This is because the high temperature ties in with humidity levels to determine a heat index threshold. The same does not apply to drier climates. A heat storm is a Californian term for an extended heat wave. Heat storms occur when the temperature reaches for three or more consecutive days over a wide area (tens of thousands of square miles). The National Weather Service issues heat advisories and excessive heat warnings when it expects unusual periods of hot weather. In Adelaide, South Australia, a heat wave is five consecutive days at or above , or three consecutive days at or over . The Australian Bureau of Meteorology defines a heat wave as three or more days of unusual maximum and minimum temperatures. Before this new Pilot Heatwave Forecast there was no national definition for heat waves or measures of heat wave severity. Marine Heatwaves Marine heatwaves have become a prominent subject of research in recent years, reflecting the fact that since the turn of this century many ocean areas have experienced peaks of temperatures, along with more frequent, more intense, more prolonged warming events than ever met on record. The genesis of marine heatwaves is mainly driven by a combination of oceanic and atmospheric factors, often triggered by high pressure systems that will reduce cloud cover and increase solar absorption by the sea surface. Human-induced climate change appears bound to play a growing role in the development of marine heatwaves, with increasing impacts on marine ecosystems, such as mass mortality in benthic communities, coral bleaching events, disruptions in fishery catches, and shifts in species distributions. Observations It is possible to compare heat waves in different regions of the world with different climates thanks to a general indicator that appeared in 2015. With these indicators, experts estimated heat waves at the global scale from 1901 to 2010. They found a substantial and sharp increase in the number of affected areas in the last two decades. One study in 2021 investigated 13,115 cities. It found that extreme heat exposure of a wet bulb globe temperature above 30 Celsius tripled between 1983 and 2016, and if the effect of population growth (increasing the urban heat island effect) during those years is excluded, the exposure increased a further 50%. The researchers compiled a comprehensive list of past urban extreme heat events. Causes Heat waves form when a high pressure area at an altitude of strengthens and remains over a region for several days and up to several weeks. This is common in summer in both the Northern and Southern Hemispheres. This is because the jet stream 'follows the sun'. The high pressure area is on the equator side of the jet stream in the upper layers of the atmosphere. Weather patterns are generally slower to change in summer than in winter. So, this upper level high pressure also moves slowly. Under high pressure, the air sinks toward the surface. It warms and dries adiabatically. This inhibits convection and prevents the formation of clouds. A reduction of clouds increases the shortwave radiation reaching the surface. A low pressure area at the surface leads to surface wind from lower latitudes that brings warm air, enhancing the warming. The surface winds could also blow from the hot continental interior towards the coastal zone. This would lead to heat waves on the coast. They could also blow from high towards low elevations. This enhances the subsidence or sinking of the air and therefore the adiabatic warming. In the eastern regions of the United States a heat wave can occur when a high pressure system originating in the Gulf of Mexico becomes stationary just off the Atlantic Seaboard. Hot humid air masses form over the Gulf of Mexico and the Caribbean Sea. At the same time hot dry air masses form over the desert Southwest and northern Mexico. The southwest winds on the back side of the high continue to pump hot, humid Gulf air northeastwards. This results in a spell of hot and humid weather for much of the eastern United States and into southeastern Canada. In the Western Cape Province of South Africa, a heat wave can occur when the low-pressure area offshore and the high-pressure area inland combine to form a bergwind. The air warms as it descends from the Karoo interior. The temperature will rise about 10 Celsius from the interior to the coast. Humidity is usually very low. The temperature can be over 40 Celsius in summer. The highest temperature recorded in South Africa (51.5 Celsius) occurred one summer during a berg wind along the Eastern Cape coastline. The level of soil moisture can intensify heat waves in Europe. Low soil moisture leads to a number of complex feedback mechanisms. These in turn can result in increased surface temperatures. One of the main mechanisms is reduced evaporative cooling of the atmosphere. When water evaporates, it consumes energy. So, it will lower the surrounding temperature. If the soil is very dry, then incoming radiation from the sun will warm the air. But there will be little or no cooling effect from moisture evaporating from the soil. Climate change Impacts on human health Heat-related health effects for vulnerable humans Mortality Underreporting of fatalities The number of heat fatalities is probably highly underreported. This is due to a lack of reports and to misreporting. When considering heat-related illnesses as well, actual death tolls from extreme heat may be six times higher than official figures. This is based on studies of California and Japan. Part of the mortality during a heat wave may be due to short-term forward mortality displacement. In some heat waves there is a decrease in overall mortality in the weeks after a heat wave. These compensatory reductions in mortality suggest that heat affects people who would have died anyway, and brings their deaths forward. Social institutions and structures influence the effects of risks. This factor can also help explain the underreporting of heat waves as a health risk. The deadly French heat wave in 2003 showed that heat wave dangers result from a combination of natural and social factors. Social invisibility is one such factor. Heat-related deaths can occur indoors, for instance among elderly people living alone. In these cases it can be challenging to assign heat as a contributing factor. Heat index for temperature and relative humidity The heat index in the table above is a measure of how hot it feels when relative humidity is factored with the actual air temperature. Psychological and sociological effects Excessive heat causes psychological stress as well as physical stress. This can affect performance. It may also lead to an increase in violent crime. High temperatures are associated with increased conflict between individuals and at the social level. In every society, crime rates go up when temperatures go up. This is particularly the case with violent crimes such as assault, murder and rape. In politically unstable countries, high temperatures can exacerbate factors that lead to civil war. High temperatures also have a significant effect on income. A study of countries in the United States found that the economic productivity of individual days declines by about 1.7 percent for each degree Celsius above . Surface ozone (air pollution) High temperatures also make the effects of ozone pollution in urban areas worse. This raises heat-related mortality during heat waves. During heat waves in urban areas, ground level ozone pollution can be 20 percent higher than usual. One study looked at fine particle concentrations and ozone concentrations from 1860 to 2000. It found that the global population-weighted fine particle concentrations increased by 5 percent due to climate change. Near-surface ozone concentrations rose by 2 percent. An investigation to assess the joint mortality effects of ozone and heat during the European heat waves in 2003 concluded that these appear to be additive. Impacts on societies Reduced economic outputs Calculations from 2022 suggest that heat waves will shrink the global economy by about 1 percent decrease by the middle of the 21st century. Heat waves often have complex effects on economies. They reduce labour productivity, disrupt agricultural and industrial processes and damage infrastructure that is not suitable for extreme heat. In 2016, a marine heatwave in Chile and its subsequent harmful algal bloom caused $800 million (USD) in export losses for the aquaculture industry as salmon and shellfish died off. Reduced agricultural outputs Heat waves are a big threat to agricultural production. In 2019 heat waves in the Mulanje region of Malawi involved temperatures as high as . This and a late rain season scorched tea leaves and reduced yields. Farmed animals Infrastructural damage Heat waves cause roads and highways to buckle and melt, water lines to burst, and power transformers to detonate, causing fires. A heat wave can also damage railways, by buckling and kinking rails. This can slow down or delay traffic. It can even lead to cancellations of service when rails are too dangerous to traverse by trains. Power outages Heat waves often lead to spikes in electricity demand because there is more use of air conditioning. This can create power outages, making the problem worse. During the 2006 North American heat wave, thousands of homes and businesses went without power, especially in California. In Los Angeles, electrical transformers failed, leaving thousands without power for as long as five days. The early 2009 southeastern Australia heat wave caused major power disruptions in the city of Melbourne. They left over half a million people without power as the heat wave blew transformers and overloaded a power grid. Impacts on the environment Wildfires A heat wave occurring during a drought can contribute to bushfires and wildfires. This is because a drought dries out vegetation, so it is more likely to catch fire. During the disastrous heat wave that struck Europe in 2003, fires raged through Portugal. They destroyed over of forest and of agricultural land. They caused about €1 billion worth of damage. High end farmlands have irrigation systems to back up crops. Floods Heat waves can also contribute to flooding. Because hot air is able to carry more moisture, heatwaves may be followed by extreme rainfall especially in mid-latitude regions. For example, the record-breaking heat wave that afflicted Pakistan beginning in May 2022 led to glacier melt and moisture flow. These were factors in the devastating floods that began in June and claimed over 1,100 lives. Wild animals on land Researchers have predicted that roughly 10-40% of all land vertebrate species will be affected by heat waves by 2099, depending on the amount of future greenhouse gas emissions. Heatwaves present an additional form of stress and evolutionary pressure for species that already deal with habitat loss and climate change. Species have a thermal range of tolerance that describes the temperatures where they perform best. Temperature conditions that are outside of this range may experience decreased fitness and the inability to reproduce. The species with sufficient genetic variation will be able to ensure some individuals can survive frequent days of high temperatures in the future. Oceans Marine heatwaves may cause mass mortality in fish populations, especially for species that are better adapted to cooler temperatures. Species that have adapted to warmer temperatures may expand their range during a heatwave. These invasive species may outcompete the native species that experience higher mortality during a heatwave, which disrupts ecosystem functioning. Marine heatwaves have also been correlated with negative impacts on foundation species such as coral and kelp. Options for reducing impacts on humans A possible public health measure during heat waves is to set up air-conditioned public cooling centres. Adding air conditioning in schools provides a cooler work place. But it can result in additional greenhouse gas emissions unless solar energy is used. Policymakers, funders and researchers have created the Extreme Heat Resilience Alliance coalition under the Atlantic Council. This advocates for naming heat waves, measuring them, and ranking them to build better awareness of their impacts. Recent examples by country or region Around the world in 2024 India Southeast Asia China A study found the average resident in China was exposed to 16 days of heat waves in 2023, with more than 37,000 heat wave-related deaths. Besides, the number of work hours lost due to heat stress in China was 36.9 billion in 2023, and China’s citizens experienced a 60% surge in lost safe outdoor activity hours, with each person losing 2.2 hours on average each day. The study predicted that by the 2060s, annual heat wave-related mortality is expected to reach 29,000 to 38,000 in China, with a 28% to 37% increase in work hours lost. United States In July 2019, there were over 50 million people in the United States in jurisdictions with heat advisories. Scientists predicted that many records for highest low temperatures would be broken in the days following these warnings. This means the lowest temperature in a 24-hour period will be higher than any low temperature measured before. According to a 2022 study, 107 million people in the US will experience extremely dangerous heat in the year 2053. Heat waves are the most lethal type of weather phenomenon in the United States. Between 1992 and 2001, deaths from excessive heat in the United States numbered 2,190, compared with 880 deaths from floods and 150 from tropical cyclones. About 400 deaths a year on average are directly due to heat in the United States. The 1995 Chicago heat wave, one of the worst in US history, led to approximately 739 heat-related deaths over 5 days. In the United States, the loss of human life in hot spells in summer exceeds that caused by all other weather events. These include lightning, rain, floods, hurricanes, and tornadoes. About 6,200 Americans need hospital treatment each summer, according to data from 2008. This is due to excessive heat, and those at highest risk are poor, uninsured or elderly. The relationship between extreme temperature and mortality in the United States varies by location. Heat is more likely to increase the risk of death in cities in the northern part of the country than in southern regions. As a whole, people in the United States appear to be adapting to hotter temperatures further north each decade. This might be due to better infrastructure, more modern building design and better public awareness.
Physical sciences
Seasons
Earth science
293336
https://en.wikipedia.org/wiki/Ammonium%20chloride
Ammonium chloride
Ammonium chloride is an inorganic chemical compound with the chemical formula , also written as . It is an ammonium salt of hydrogen chloride. It consists of ammonium cations and chloride anions . It is a white crystalline salt that is highly soluble in water. Solutions of ammonium chloride are mildly acidic. In its naturally occurring mineralogic form, it is known as salammoniac. The mineral is commonly formed on burning coal dumps from condensation of coal-derived gases. It is also found around some types of volcanic vents. It is mainly used as fertilizer and a flavouring agent in some types of liquorice. It is a product of the reaction of hydrochloric acid and ammonia. Production It is a product of the Solvay process used to produce sodium carbonate: CO2 + 2 NH3 + 2 NaCl + H2O → 2 NH4Cl + Na2CO3 Not only is that method the principal one for the manufacture of ammonium chloride, but also it is used to minimize ammonia release in some industrial operations. Ammonium chloride is prepared commercially by combining ammonia (NH3) with either hydrogen chloride (gas) or hydrochloric acid (water solution): NH3 + HCl → NH4Cl Ammonium chloride occurs naturally in volcanic regions, forming on volcanic rocks near fume-releasing vents (fumaroles). The crystals deposit directly from the gaseous state and tend to be short-lived, as they dissolve easily in water. Reactions Ammonium chloride appears to sublime upon heating but actually reversibly decomposes into ammonia and hydrogen chloride gas: NH4Cl NH3 + HCl Ammonium chloride reacts with a strong base, like sodium hydroxide, to release ammonia gas: NH4Cl + NaOH → NH3 + NaCl + H2O Similarly, ammonium chloride also reacts with alkali-metal carbonates at elevated temperatures, giving ammonia and alkali-metal chloride: 2 NH4Cl + Na2CO3 → 2 NaCl + CO2 + H2O + 2 NH3 A solution of 5% by mass of ammonium chloride in water has a pH in the range 4.6 to 6.0. Some reactions of ammonium chloride with other chemicals are endothermic, such as its reaction with barium hydroxide and its dissolving in water. Applications Agriculture The dominant application of ammonium chloride is as a nitrogen source in fertilizers (corresponding to 90% of the world production of ammonium chloride) such as chloroammonium phosphate. The main crops fertilized this way are rice and wheat in Asia. When using ammonium chloride as a nitrogen fertilizer for plants, the appropriate concentration is applied to provide sufficient nutrients without causing harm. Ammonium chloride is approximately 26% nitrogen by weight and can be used to supply nitrogen to plants, especially those preferring slightly acidic conditions. The concentration for nitrogen fertilization in solution is between 50 and 100 milligrams of nitrogen per liter of water (mg N/L), which is equivalent to 50–100 parts per million (ppm) nitrogen, which translates to approximately 0.2 to 0.4 grams of ammonium chloride per liter of water. Ammonium chloride can acidify the soil over time, so soil pH is regularly monitored, especially when growing plants sensitive to acidic conditions. Some plants are sensitive to chloride ions (e.g., avocados, beans, grapes), so applying ammonium chloride to such plants should be done with extra caution to prevent chloride toxicity. While ammonium chloride can be beneficial as a nitrogen source, improper use can harm plants and the environment. Ammonium chloride solutions are generally stable and can be stored for a certain period if kept under appropriate conditions, that is in airtight containers (to prevent contamination, evaporation and hydrolysis), away from light (to prevent photodegradation) and heat sources (to reduce microbial growth and chemical degradation), and if contamination is prevented. In agricultural applications the solution us used shortly after preparation, for the following reasons: Nutrient-rich solutions can promote the growth of microorganisms over time, so that microbial activity can alter the chemical composition of the solution, potentially reducing its efficacy as a fertilizer and introducing pathogens to plants. Over time, water can evaporate from the solution, especially if not stored in a tightly sealed container, which increases the concentration of ammonium chloride, and may lead to over-fertilization and potential damage to plants when applied. While ammonium chloride is relatively stable, prolonged storage may lead to minor changes in pH due to ongoing hydrolysis, especially if the solution is exposed to air, potentially impacting plants sensitive to acidity of the soil. If the water used is not distilled or deionized, dissolved minerals and impurities may precipitate over time, altering the nutrient balance of the solution. Pyrotechnics Ammonium chloride was used in pyrotechnics in the 18th century but was superseded by safer and less hygroscopic chemicals. Its purpose was to provide a chlorine donor to enhance the green and blue colours from copper ions in the flame. It had a secondary use to provide white smoke, but its ready double decomposition reaction with potassium chlorate producing the highly unstable ammonium chlorate made its use very dangerous. Metalwork Ammonium chloride is used as a flux in preparing metals to be tin coated, galvanized or soldered. It works as a flux by cleaning the surface of workpieces by reacting with the metal oxides at the surface to form a volatile metal chloride. For that purpose, it is sold in blocks at hardware stores for use in cleaning the tip of a soldering iron, and it can also be included in solder as flux. Medicine Ammonium chloride is used as an expectorant in cough medicine. Its expectorant action is caused by irritative action on the bronchial mucosa, which causes the production of excess respiratory tract fluid, which presumably is easier to cough up. Ammonium salts are an irritant to the gastric mucosa and may induce nausea and vomiting. Ammonium chloride is used as a systemic acidifying agent in treatment of severe metabolic alkalosis, in oral acid loading test to diagnose distal renal tubular acidosis, to maintain the urine at an acid pH in the treatment of some urinary-tract disorders. Food Ammonium chloride, under the name sal ammoniac or salmiak is used as food additive under the E number E510, working as a yeast nutrient in breadmaking and as an acidifier. It is a feed supplement for cattle and an ingredient in nutritive media for yeasts and many microorganisms. Ammonium chloride is used to spice up dark sweets called salty liquorice (popular in the Nordic countries, Benelux and northern Germany), in baking to give cookies a very crisp texture, and in the liquor Salmiakki Koskenkorva for flavouring. In Turkey, Iran, Tajikistan, India, Pakistan and Arab countries it is called "noshader" and is used to improve the crispness of snacks such as samosas and jalebi. In the laboratory Ammonium chloride has been used historically to produce low temperatures in cooling baths. Ammonium chloride solutions with ammonia are used as buffer solutions including ACK (Ammonium-Chloride-Potassium) lysis buffer. In paleontology, ammonium chloride vapor is deposited on fossils, where the substance forms a brilliant white, easily removed and fairly harmless and inert layer of tiny crystals that covers up any coloration the fossil may have, and if lighted at an angle highly enhances contrast in photographic documentation of three-dimensional specimens. The same technique is applied in archaeology to eliminate reflection on glass and similar specimens for photography. In organic synthesis saturated NH4Cl solution is typically used to quench reaction mixtures. It has a lambda transition at 242.8 K and zero pressure. Flotation Giant squid and some other large squid species maintain neutral buoyancy in seawater through an ammonium chloride solution which is found throughout their bodies and is less dense than seawater. This differs from the method of flotation used by most fish, which involves a gas-filled swim bladder. Batteries Around the turn of the 20th century, ammonium chloride was used in aqueous solution as the electrolyte in Leclanché cells that found a commercial use as the "local battery" in subscribers' telephone installations. Those cells later evolved into zinc–carbon batteries still using ammonium chloride as electrolyte. Concrete treatments Ammonium chloride is known to be an aggressive cleaning agent. A penetrating and intense reddish brown color is stained into concrete surfaces with a mixture of ammonium chloride and ferric chloride. Pre-treatment with acid is unnecessary. Photography Ammonium chloride can also be used in the process of making albumen silver prints, commonly known as albumen prints. In traditional photographic printing processes of the 19th century, ammonium chloride served as a key component in preparing the albumen solution used to coat the photographic paper. Albumen printing was the dominant photographic printing technique from the 1850s through the 1890s, prized for its fine detail and rich tonal rendition. The incorporation of ammonium chloride in the albumen solution was a significant factor in the quality and popularity of this photographic process. The process involves mixing egg whites (albumen) with ammonium chloride to create a viscous solution. This mixture is then applied as a thin layer onto paper, which, after drying, forms a smooth and glossy surface. Ammonium chloride acts as a salting agent, contributing chloride ions that are essential for forming light-sensitive silver chloride when the coated paper is subsequently sensitized with a solution of silver nitrate. Upon exposure to light, the silver chloride reduces to metallic silver, creating a visible image. The use of ammonium chloride, as opposed to sodium chloride (common salt), can influence the contrast and tonal range of the final print, often yielding warmer tones and greater image clarity. Other applications Ammonium chloride is used in a ~5% aqueous solution to work on oil wells with clay swelling problems. Other uses include in hair shampoo, in the glue that bonds plywood, and in cleaning products. In hair shampoo, it is used as a thickening agent in ammonium-based surfactant systems such as ammonium lauryl sulfate. Ammonium chloride is used in the textile and leather industry, in dyeing, tanning, textile printing and cotton clustering. In woodworking, a solution of ammonium chloride and water, when applied to unfinished wood, will burn when subjected to a heat gun resulting in a branding iron mark without use of a branding iron. The solution can be painted onto the wood or applied with a common rubber stamp. History Etymology Pliny, in Book XXXI of his Natural History, refers to a salt produced in the Roman province of Cyrenaica named hammoniacum, so called because of its proximity to the nearby Temple of Jupiter Amun (Greek Ἄμμων Ammon). However, the description Pliny gives of the salt does not conform to the properties of ammonium chloride. According to Herbert Hoover's commentary in his English translation of Georgius Agricola's De re metallica, it is likely to have been common sea salt. Nevertheless, that salt ultimately gave ammonia and ammonium compounds their name. Ancient China The earliest mention of ammonium chloride was in 554 in China. At that time, ammonium chloride came from two sources: (1) the vents of underground coal fires in Central Asia, specifically, in the Tian Shan mountains (which extend from Xinjiang province of northwestern China through Kyrgyzstan) as well as in the Alay (or Alai) mountains of southwestern Kyrgyzstan, and (2) the fumaroles of the volcano Mount Taftan in southeastern Iran. (Indeed, the word for ammonium chloride in several Asian languages derives from the Iranian phrase anosh adur (immortal fire), a reference to the underground fires.) Ammonium chloride was then transported along the Silk Road eastwards to China and westwards to the Muslim lands and Europe. Jabirian alchemists Around 800 A.D. the iranian chemist jaber ibn hayan discovered ammonium chloride in the soot that resulted from burning camel dung, and this source became an alternative to those in Central Asia. The Jabirian alchemists were the authors of the Jabirian corpus, tentatively dated to . The word for ammonium chloride in the Jabirian corpus was nošāder, Iranian in origin. Whereas Greek alchemical texts had been almost exclusively focused on the use of mineral substances, Jabirian alchemy pioneered the use of vegetable and animal substances, and so represented an innovative shift towards 'organic chemistry'. In the Jabirian corpus, the production of ammonium chloride from organic substances (such as plants, blood, and hair) is described. These are the oldest known instructions for deriving an inorganic compound from organic substances by chemical means. One of the innovations in Jabirian alchemy was the addition of ammonium chloride to the category of chemical substances known as 'spirits' (i.e., strongly volatile substances). This included both naturally occurring sal ammoniac and synthetic ammonium chloride produced from organic substances. The addition of sal ammoniac to the list of 'spirits' can perhaps also be seen as a product of this new focus on organic chemistry. Late Middle Ages The first attested reference to sal ammoniac as ammonium chloride is in the Pseudo-Geber work De inventione veritatis, where a preparation of sal ammoniac is given in the chapter De Salis armoniaci præparatione, salis armoniaci being a common name in the Middle Ages for sal ammoniac.
Physical sciences
Halide salts
Chemistry
293396
https://en.wikipedia.org/wiki/Flat-panel%20display
Flat-panel display
A flat-panel display (FPD) is an electronic display used to display visual content such as text or images. It is present in consumer, medical, transportation, and industrial equipment. Flat-panel displays are thin, lightweight, provide better linearity and are capable of higher resolution than typical consumer-grade TVs from earlier eras. They are usually less than thick. While the highest resolution for consumer-grade CRT televisions was 1080i, many interactive flat panels in the 2020s are capable of 1080p and 4K resolution. In the 2010s, portable consumer electronics such as laptops, mobile phones, and portable cameras have used flat-panel displays since they consume less power and are lightweight. As of 2016, flat-panel displays have almost completely replaced CRT displays. Most 2010s-era flat-panel displays use LCD or light-emitting diode (LED) technologies, sometimes combined. Most LCD screens are back-lit with color filters used to display colors. In many cases, flat-panel displays are combined with touch screen technology, which allows the user to interact with the display in a natural manner. For example, modern smartphone displays often use OLED panels, with capacitive touch screens. Flat-panel displays can be divided into two display device categories: volatile and static. The former requires that pixels be periodically electronically refreshed to retain their state (e.g. liquid-crystal displays (LCD)), and can only show an image when it has power. On the other hand, static flat-panel displays rely on materials whose color states are bistable, such as displays that make use of e-ink technology, and as such retain content even when power is removed. History The first engineering proposal for a flat-panel TV was by General Electric in 1954 as a result of its work on radar monitors. The publication of their findings gave all the basics of future flat-panel TVs and monitors. But GE did not continue with the R&D required and never built a working flat panel at that time. The first production flat-panel display was the Aiken tube, developed in the early 1950s and produced in limited numbers in 1958. This saw some use in military systems as a heads up display and as an oscilloscope monitor, but conventional technologies overtook its development. Attempts to commercialize the system for home television use ran into continued problems and the system was never released commercially. Dennis Gabor, better known as the inventor of holography, patented a flat-screen CRT in 1958. This was substantially similar to Aiken's concept, and led to a years-long patent battle. By the time the lawsuits were complete, with Aiken's patent applying in the US and Gabor's in the UK, the commercial aspects had long lapsed, and the two became friends. Around this time,Clive Sinclair came across Gabor's work and began an ultimately unsuccessful decade-long effort to commercialize it. The Philco Predicta featured a relatively flat (for its day) cathode-ray tube setup and would be the first commercially released "flat panel" upon its launch in 1958; the Predicta was a commercial failure. The plasma display panel was invented in 1964 at the University of Illinois, according to The History of Plasma Display Panels. Liquid-crystal displays (LC displays, or LCDs) The MOSFET (metal–oxide–semiconductor field-effect transistor, or MOS transistor) was invented by Mohamed M. Atalla and Dawon Kahng at Bell Labs in 1959, and presented in 1960. Building on their work, Paul K. Weimer at RCA developed the thin-film transistor (TFT) in 1962. It was a type of MOSFET distinct from the standard bulk MOSFET. The idea of a TFT-based LCD was conceived by Bernard J. Lechner of RCA Laboratories in 1968. B.J. Lechner, F.J. Marlowe, E.O. Nester and J. Tults demonstrated the concept in 1968 with a dynamic scattering LCD that used standard discrete MOSFETs. The first active-matrix addressed electroluminescent display (ELD) was made using TFTs by T. Peter Brody's Thin-Film Devices department at Westinghouse Electric Corporation in 1968. In 1973, Brody, J. A. Asars and G. D. Dixon at Westinghouse Research Laboratories demonstrated the first thin-film-transistor liquid-crystal display (TFT LCD). Brody and Fang-Chen Luo demonstrated the first flat active-matrix liquid-crystal display (AM LCD) using TFTs in 1974. By 1982, pocket LCD TVs based on LCD technology were developed in Japan. The 2.1-inch Epson ET-10 Epson Elf was the first color LCD pocket TV, released in 1984. In 1988, a Sharp research team led by engineer T. Nagayasu demonstrated a 14-inch full-color LCD, which convinced the electronics industry that LCD would eventually replace CRTs as the standard television display technology. , all modern high-resolution and high-quality electronic visual display devices use TFT-based active-matrix displays. LED displays The first usable LED display was developed by Hewlett-Packard (HP) and introduced in 1968. It was the result of research and development (R&D) on practical LED technology between 1962 and 1968, by a research team under Howard C. Borden, Gerald P. Pighini, and Mohamed M. Atalla, at HP Associates and HP Labs. In February 1969, they introduced the HP Model 5082-7000 Numeric Indicator. It was the first alphanumeric LED display, and was a revolution in digital display technology, replacing the Nixie tube for numeric displays and becoming the basis for later LED displays. In 1977, James P Mitchell prototyped and later demonstrated what was perhaps the earliest monochromatic flat-panel LED television display. Ching W. Tang and Steven Van Slyke at Eastman Kodak built the first practical organic LED (OLED) device in 1987. In 2003, Hynix produced an organic EL driver capable of lighting in 4,096 colors. In 2004, the Sony Qualia 005 was the first LED-backlit LCD. The Sony XEL-1, released in 2007, was the first OLED television. Common types Liquid-crystal display (LCD) Field-effect LCDs are lightweight, compact, portable, cheap, more reliable, and easier on the eyes than CRT screens. LCD screens use a thin layer of liquid crystal, a liquid that exhibits crystalline properties. It is sandwiched between two glass plates carrying transparent electrodes. Two polarizing films are placed at each side of the LCD. By generating a controlled electric field between electrodes, various segments or pixels of the liquid crystal can be activated, causing changes in their polarizing properties. These polarizing properties depend on the alignment of the liquid-crystal layer and the specific field-effect used, being either Twisted Nematic (TN), In-Plane Switching (IPS) or Vertical Alignment (VA). Color is produced by applying appropriate color filters (red, green and blue) to the individual subpixels. LC displays are used in various electronics like watches, calculators, mobile phones, TVs, computer monitors and laptops screens etc. LED-LCD Most earlier large LCD screens were back-lit using a number of CCFL (cold-cathode fluorescent lamps). However, small pocket size devices almost always used LEDs as their illumination source. With the improvement of LEDs, almost all new displays are now equipped with LED backlight technology. The image is still generated by the LCD layer. Plasma panel A plasma display consists of two glass plates separated by a thin gap filled with a gas such as neon. Each of these plates has several parallel electrodes running across it. The electrodes on the two plates are at right angles to each other. A voltage applied between the two electrodes one on each plate causes a small segment of gas at the two electrodes to glow. The glow of gas segments is maintained by a lower voltage that is continuously applied to all electrodes. By 2010, consumer plasma displays had been discontinued by numerous manufacturers. Electroluminescent panel In an electroluminescent display (ELD), the image is created by applying electrical signals to the plates which make the phosphor glow. Organic light-emitting diode An OLED (organic light-emitting diode) is a light-emitting diode (LED) in which the emissive electroluminescent layer is a film of organic compound which emits light in response to an electric current. This layer of organic semiconductor is situated between two electrodes; typically, at least one of these electrodes is transparent. OLEDs are used to create digital displays in devices such as television screens, computer monitors, portable systems such as mobile phones, handheld game consoles and PDAs. Quantum-dot light-emitting diode QLED or quantum dot LED is a flat panel display technology introduced by Samsung under this trademark. Other television set manufacturers such as Sony have used the same technology to enhance the backlighting of LCD TVs already in 2013. Quantum dots create their own unique light when illuminated by a light source of shorter wavelength such as blue LEDs. This type of LED TV enhances the colour gamut of LCD panels, where the image is still generated by the LCD. In the view of Samsung, quantum dot displays for large-screen TVs are expected to become more popular than the OLED displays in the coming years; Firms like Nanoco and Nanosys compete to provide the QD materials. In the meantime, Samsung Galaxy devices such as smartphones are still equipped with OLED displays manufactured by Samsung as well. Samsung explains on their website that the QLED TV they produce can determine what part of the display needs more or less contrast. Samsung also announced a partnership with Microsoft that will promote the new Samsung QLED TV. Volatile Volatile displays require that pixels be periodically refreshed to retain their state, even for a static image. As such, a volatile screen needs electrical power, either from mains electricity (being plugged into a wall socket) or a battery to maintain an image on the display or change the image. This refresh typically occurs many times a second. If this is not done, for example, if there is a power outage, the pixels will gradually lose their coherent state, and the image will "fade" from the screen. Examples The following flat-display technologies have been commercialized in 1990s to 2010s: Plasma display panel (PDP) Active-matrix liquid-crystal display (AMLCD) Rear projection: Digital Light Processing (DLP), LCD, LCOS Electronic paper: E Ink, Gyricon Light-emitting diode display (LED) Active-matrix organic light-emitting diode (AMOLED) Quantum dot display (QLED) Technologies that were extensively researched, but their commercialization was limited or has been ultimately abandoned: Active-matrix electroluminescent display (ELD) Interferometric modulator display (IMOD) Field-emission display (FED) Surface-conduction electron-emitter display (SED, SED-TV) Static Static flat-panel displays rely on materials whose color states are bistable. This means that the image they hold requires no energy to maintain, but instead requires energy to change. This results in a much more energy-efficient display, but with a tendency toward slow refresh rates which are undesirable in an interactive display. Bistable flat-panel displays are beginning deployment in limited applications (cholesteric liquid-crystal displays, manufactured by Magink, in outdoor advertising; electrophoretic displays in e-book reader devices from Sony and iRex; anlabels; interferometric modulator displays in a smartwatch).
Technology
Media and communication
null
293485
https://en.wikipedia.org/wiki/24-cell
24-cell
In four-dimensional geometry, the 24-cell is the convex regular 4-polytope (four-dimensional analogue of a Platonic solid) with Schläfli symbol {3,4,3}. It is also called C24, or the icositetrachoron, octaplex (short for "octahedral complex"), icosatetrahedroid, octacube, hyper-diamond or polyoctahedron, being constructed of octahedral cells. The boundary of the 24-cell is composed of 24 octahedral cells with six meeting at each vertex, and three at each edge. Together they have 96 triangular faces, 96 edges, and 24 vertices. The vertex figure is a cube. The 24-cell is self-dual. The 24-cell and the tesseract are the only convex regular 4-polytopes in which the edge length equals the radius. The 24-cell does not have a regular analogue in three dimensions or any other number of dimensions, either below or above. It is the only one of the six convex regular 4-polytopes which is not the analogue of one of the five Platonic solids. However, it can be seen as the analogue of a pair of irregular solids: the cuboctahedron and its dual the rhombic dodecahedron. Translated copies of the 24-cell can tesselate four-dimensional space face-to-face, forming the 24-cell honeycomb. As a polytope that can tile by translation, the 24-cell is an example of a parallelotope, the simplest one that is not also a zonotope. Geometry The 24-cell incorporates the geometries of every convex regular polytope in the first four dimensions, except the 5-cell, those with a 5 in their Schlӓfli symbol, and the regular polygons with 7 or more sides. In other words, the 24-cell contains all of the regular polytopes made of triangles and squares that exist in four dimensions except the regular 5-cell, but none of the pentagonal polytopes. The geometric relationships among all of these regular polytopes can be observed in a single 24-cell or the 24-cell honeycomb. The 24-cell is the fourth in the sequence of six convex regular 4-polytopes (in order of size and complexity). It can be deconstructed into 3 overlapping instances of its predecessor the tesseract (8-cell), as the 8-cell can be deconstructed into 2 instances of its predecessor the 16-cell. The reverse procedure to construct each of these from an instance of its predecessor preserves the radius of the predecessor, but generally produces a successor with a smaller edge length. Coordinates The 24-cell has two natural systems of Cartesian coordinates, which reveal distinct structure. Squares The 24-cell is the convex hull of its vertices which can be described as the 24 coordinate permutations of: Those coordinates can be constructed as , rectifying the 16-cell with 8 vertices permutations of (±2,0,0,0). The vertex figure of a 16-cell is the octahedron; thus, cutting the vertices of the 16-cell at the midpoint of its incident edges produces 8 octahedral cells. This process also rectifies the tetrahedral cells of the 16-cell which become 16 octahedra, giving the 24-cell 24 octahedral cells. In this frame of reference the 24-cell has edges of length and is inscribed in a 3-sphere of radius . Remarkably, the edge length equals the circumradius, as in the hexagon, or the cuboctahedron. Such polytopes are radially equilateral. The 24 vertices form 18 great squares (3 sets of 6 orthogonal central squares), 3 of which intersect at each vertex. By viewing just one square at each vertex, the 24-cell can be seen as the vertices of 3 pairs of completely orthogonal great squares which intersect at no vertices. Hexagons The 24-cell is self-dual, having the same number of vertices (24) as cells and the same number of edges (96) as faces. If the dual of the above 24-cell of edge length is taken by reciprocating it about its inscribed sphere, another 24-cell is found which has edge length and circumradius 1, and its coordinates reveal more structure. In this frame of reference the 24-cell lies vertex-up, and its vertices can be given as follows: 8 vertices obtained by permuting the integer coordinates: and 16 vertices with half-integer coordinates of the form: all 24 of which lie at distance 1 from the origin. Viewed as quaternions, these are the unit Hurwitz quaternions. The 24-cell has unit radius and unit edge length in this coordinate system. We refer to the system as unit radius coordinates to distinguish it from others, such as the radius coordinates used above. The 24 vertices and 96 edges form 16 non-orthogonal great hexagons, four of which intersect at each vertex. By viewing just one hexagon at each vertex, the 24-cell can be seen as the 24 vertices of 4 non-intersecting hexagonal great circles which are Clifford parallel to each other. The 12 axes and 16 hexagons of the 24-cell constitute a Reye configuration, which in the language of configurations is written as 124163 to indicate that each axis belongs to 4 hexagons, and each hexagon contains 3 axes. Triangles The 24 vertices form 32 equilateral great triangles, of edge length in the unit-radius 24-cell, inscribed in the 16 great hexagons. Each great triangle is a ring linking three completely disjoint great squares. Hypercubic chords The 24 vertices of the 24-cell are distributed at four different chord lengths from each other: , , and . The chords (the 24-cell edges) are the edges of central hexagons, and the chords are the diagonals of central hexagons. The chords are the edges of central squares, and the chords are the diagonals of central squares. Each vertex is joined to 8 others by an edge of length 1, spanning 60° = of arc. Next nearest are 6 vertices located 90° = away, along an interior chord of length . Another 8 vertices lie 120° = away, along an interior chord of length . The opposite vertex is 180° = away along a diameter of length 2. Finally, as the 24-cell is radially equilateral, its center is 1 edge length away from all vertices. To visualize how the interior polytopes of the 24-cell fit together (as described below), keep in mind that the four chord lengths (, , , ) are the long diameters of the hypercubes of dimensions 1 through 4: the long diameter of the square is ; the long diameter of the cube is ; and the long diameter of the tesseract is . Moreover, the long diameter of the octahedron is like the square; and the long diameter of the 24-cell itself is like the tesseract. Geodesics The vertex chords of the 24-cell are arranged in geodesic great circle polygons. The geodesic distance between two 24-cell vertices along a path of edges is always 1, 2, or 3, and it is 3 only for opposite vertices. The edges occur in 16 hexagonal great circles (in planes inclined at 60 degrees to each other), 4 of which cross at each vertex. The 96 distinct edges divide the surface into 96 triangular faces and 24 octahedral cells: a 24-cell. The 16 hexagonal great circles can be divided into 4 sets of 4 non-intersecting Clifford parallel geodesics, such that only one hexagonal great circle in each set passes through each vertex, and the 4 hexagons in each set reach all 24 vertices. The chords occur in 18 square great circles (3 sets of 6 orthogonal planes), 3 of which cross at each vertex. The 72 distinct chords do not run in the same planes as the hexagonal great circles; they do not follow the 24-cell's edges, they pass through its octagonal cell centers. The 72 chords are the 3 orthogonal axes of the 24 octahedral cells, joining vertices which are 2 edges apart. The 18 square great circles can be divided into 3 sets of 6 non-intersecting Clifford parallel geodesics, such that only one square great circle in each set passes through each vertex, and the 6 squares in each set reach all 24 vertices. The chords occur in 32 triangular great circles in 16 planes, 4 of which cross at each vertex. The 96 distinct chords run vertex-to-every-other-vertex in the same planes as the hexagonal great circles. They are the 3 edges of the 32 great triangles inscribed in the 16 great hexagons, joining vertices which are 2 edges apart on a great circle. The chords occur as 12 vertex-to-vertex diameters (3 sets of 4 orthogonal axes), the 24 radii around the 25th central vertex. The sum of the squared lengths of all these distinct chords of the 24-cell is 576 = 242. These are all the central polygons through vertices, but in 4-space there are geodesics on the 3-sphere which do not lie in central planes at all. There are geodesic shortest paths between two 24-cell vertices that are helical rather than simply circular; they correspond to diagonal isoclinic rotations rather than simple rotations. The edges occur in 48 parallel pairs, apart. The chords occur in 36 parallel pairs, apart. The chords occur in 48 parallel pairs, apart. The central planes of the 24-cell can be divided into 4 orthogonal central hyperplanes (3-spaces) each forming a cuboctahedron. The great hexagons are 60 degrees apart; the great squares are 90 degrees or 60 degrees apart; a great square and a great hexagon are 90 degrees and 60 degrees apart. Each set of similar central polygons (squares or hexagons) can be divided into 4 sets of non-intersecting Clifford parallel polygons (of 6 squares or 4 hexagons). Each set of Clifford parallel great circles is a parallel fiber bundle which visits all 24 vertices just once. Each great circle intersects with the other great circles to which it is not Clifford parallel at one diameter of the 24-cell. Great circles which are completely orthogonal or otherwise Clifford parallel do not intersect at all: they pass through disjoint sets of vertices. Constructions Triangles and squares come together uniquely in the 24-cell to generate, as interior features, all of the triangle-faced and square-faced regular convex polytopes in the first four dimensions (with caveats for the 5-cell and the 600-cell). Consequently, there are numerous ways to construct or deconstruct the 24-cell. Reciprocal constructions from 8-cell and 16-cell The 8 integer vertices (±1, 0, 0, 0) are the vertices of a regular 16-cell, and the 16 half-integer vertices (±, ±, ±, ±) are the vertices of its dual, the tesseract (8-cell). The tesseract gives Gosset's construction of the 24-cell, equivalent to cutting a tesseract into 8 cubic pyramids, and then attaching them to the facets of a second tesseract. The analogous construction in 3-space gives the rhombic dodecahedron which, however, is not regular. The 16-cell gives the reciprocal construction of the 24-cell, Cesaro's construction, equivalent to rectifying a 16-cell (truncating its corners at the mid-edges, as described above). The analogous construction in 3-space gives the cuboctahedron (dual of the rhombic dodecahedron) which, however, is not regular. The tesseract and the 16-cell are the only regular 4-polytopes in the 24-cell. We can further divide the 16 half-integer vertices into two groups: those whose coordinates contain an even number of minus (−) signs and those with an odd number. Each of these groups of 8 vertices also define a regular 16-cell. This shows that the vertices of the 24-cell can be grouped into three disjoint sets of eight with each set defining a regular 16-cell, and with the complement defining the dual tesseract. This also shows that the symmetries of the 16-cell form a subgroup of index 3 of the symmetry group of the 24-cell. Diminishings We can facet the 24-cell by cutting through interior cells bounded by vertex chords to remove vertices, exposing the facets of interior 4-polytopes inscribed in the 24-cell. One can cut a 24-cell through any planar hexagon of 6 vertices, any planar rectangle of 4 vertices, or any triangle of 3 vertices. The great circle central planes (above) are only some of those planes. Here we shall expose some of the others: the face planes of interior polytopes. 8-cell Starting with a complete 24-cell, remove 8 orthogonal vertices (4 opposite pairs on 4 perpendicular axes), and the 8 edges which radiate from each, by cutting through 8 cubic cells bounded by edges to remove 8 cubic pyramids whose apexes are the vertices to be removed. This removes 4 edges from each hexagonal great circle (retaining just one opposite pair of edges), so no continuous hexagonal great circles remain. Now 3 perpendicular edges meet and form the corner of a cube at each of the 16 remaining vertices, and the 32 remaining edges divide the surface into 24 square faces and 8 cubic cells: a tesseract. There are three ways you can do this (choose a set of 8 orthogonal vertices out of 24), so there are three such tesseracts inscribed in the 24-cell. They overlap with each other, but most of their element sets are disjoint: they share some vertex count, but no edge length, face area, or cell volume. They do share 4-content, their common core. 16-cell Starting with a complete 24-cell, remove the 16 vertices of a tesseract (retaining the 8 vertices you removed above), by cutting through 16 tetrahedral cells bounded by chords to remove 16 tetrahedral pyramids whose apexes are the vertices to be removed. This removes 12 great squares (retaining just one orthogonal set) and all the edges, exposing chords as the new edges. Now the remaining 6 great squares cross perpendicularly, 3 at each of 8 remaining vertices, and their 24 edges divide the surface into 32 triangular faces and 16 tetrahedral cells: a 16-cell. There are three ways you can do this (remove 1 of 3 sets of tesseract vertices), so there are three such 16-cells inscribed in the 24-cell. They overlap with each other, but all of their element sets are disjoint: they do not share any vertex count, edge length, or face area, but they do share cell volume. They also share 4-content, their common core. Tetrahedral constructions The 24-cell can be constructed radially from 96 equilateral triangles of edge length which meet at the center of the polytope, each contributing two radii and an edge. They form 96 tetrahedra (each contributing one 24-cell face), all sharing the 25th central apex vertex. These form 24 octahedral pyramids (half-16-cells) with their apexes at the center. The 24-cell can be constructed from 96 equilateral triangles of edge length , where the three vertices of each triangle are located 90° = away from each other on the 3-sphere. They form 48 -edge tetrahedra (the cells of the three 16-cells), centered at the 24 mid-edge-radii of the 24-cell. The 24-cell can be constructed directly from its characteristic simplex , the irregular 5-cell which is the fundamental region of its symmetry group F4, by reflection of that 4-orthoscheme in its own cells (which are 3-orthoschemes). Cubic constructions The 24-cell is not only the 24-octahedral-cell, it is also the 24-cubical-cell, although the cubes are cells of the three 8-cells, not cells of the 24-cell, in which they are not volumetrically disjoint. The 24-cell can be constructed from 24 cubes of its own edge length (three 8-cells). Each of the cubes is shared by 2 8-cells, each of the cubes' square faces is shared by 4 cubes (in 2 8-cells), each of the 96 edges is shared by 8 square faces (in 4 cubes in 2 8-cells), and each of the 96 vertices is shared by 16 edges (in 8 square faces in 4 cubes in 2 8-cells). Relationships among interior polytopes The 24-cell, three tesseracts, and three 16-cells are deeply entwined around their common center, and intersect in a common core. The tesseracts and the 16-cells are rotated 60° isoclinically with respect to each other. This means that the corresponding vertices of two tesseracts or two 16-cells are (120°) apart. The tesseracts are inscribed in the 24-cell such that their vertices and edges are exterior elements of the 24-cell, but their square faces and cubical cells lie inside the 24-cell (they are not elements of the 24-cell). The 16-cells are inscribed in the 24-cell such that only their vertices are exterior elements of the 24-cell: their edges, triangular faces, and tetrahedral cells lie inside the 24-cell. The interior 16-cell edges have length . The 16-cells are also inscribed in the tesseracts: their edges are the face diagonals of the tesseract, and their 8 vertices occupy every other vertex of the tesseract. Each tesseract has two 16-cells inscribed in it (occupying the opposite vertices and face diagonals), so each 16-cell is inscribed in two of the three 8-cells. This is reminiscent of the way, in 3 dimensions, two opposing regular tetrahedra can be inscribed in a cube, as discovered by Kepler. In fact it is the exact dimensional analogy (the demihypercubes), and the 48 tetrahedral cells are inscribed in the 24 cubical cells in just that way. The 24-cell encloses the three tesseracts within its envelope of octahedral facets, leaving 4-dimensional space in some places between its envelope and each tesseract's envelope of cubes. Each tesseract encloses two of the three 16-cells, leaving 4-dimensional space in some places between its envelope and each 16-cell's envelope of tetrahedra. Thus there are measurable 4-dimensional interstices between the 24-cell, 8-cell and 16-cell envelopes. The shapes filling these gaps are 4-pyramids, alluded to above. Boundary cells Despite the 4-dimensional interstices between 24-cell, 8-cell and 16-cell envelopes, their 3-dimensional volumes overlap. The different envelopes are separated in some places, and in contact in other places (where no 4-pyramid lies between them). Where they are in contact, they merge and share cell volume: they are the same 3-membrane in those places, not two separate but adjacent 3-dimensional layers. Because there are a total of 7 envelopes, there are places where several envelopes come together and merge volume, and also places where envelopes interpenetrate (cross from inside to outside each other). Some interior features lie within the 3-space of the (outer) boundary envelope of the 24-cell itself: each octahedral cell is bisected by three perpendicular squares (one from each of the tesseracts), and the diagonals of those squares (which cross each other perpendicularly at the center of the octahedron) are 16-cell edges (one from each 16-cell). Each square bisects an octahedron into two square pyramids, and also bonds two adjacent cubic cells of a tesseract together as their common face. As we saw above, 16-cell tetrahedral cells are inscribed in tesseract cubic cells, sharing the same volume. 24-cell octahedral cells overlap their volume with cubic cells: they are bisected by a square face into two square pyramids, the apexes of which also lie at a vertex of a cube. The octahedra share volume not only with the cubes, but with the tetrahedra inscribed in them; thus the 24-cell, tesseracts, and 16-cells all share some boundary volume. As a configuration This configuration matrix represents the 24-cell. The rows and columns correspond to vertices, edges, faces, and cells. The diagonal numbers say how many of each element occur in the whole 24-cell. The non-diagonal numbers say how many of the column's element occur in or at the row's element. Since the 24-cell is self-dual, its matrix is identical to its 180 degree rotation. Symmetries, root systems, and tessellations The 24 root vectors of the D4 root system of the simple Lie group SO(8) form the vertices of a 24-cell. The vertices can be seen in 3 hyperplanes, with the 6 vertices of an octahedron cell on each of the outer hyperplanes and 12 vertices of a cuboctahedron on a central hyperplane. These vertices, combined with the 8 vertices of the 16-cell, represent the 32 root vectors of the B4 and C4 simple Lie groups. The 48 vertices (or strictly speaking their radius vectors) of the union of the 24-cell and its dual form the root system of type F4. The 24 vertices of the original 24-cell form a root system of type D4; its size has the ratio :1. This is likewise true for the 24 vertices of its dual. The full symmetry group of the 24-cell is the Weyl group of F4, which is generated by reflections through the hyperplanes orthogonal to the F4 roots. This is a solvable group of order 1152. The rotational symmetry group of the 24-cell is of order 576. Quaternionic interpretation When interpreted as the quaternions, the F4 root lattice (which is the integral span of the vertices of the 24-cell) is closed under multiplication and is therefore a ring. This is the ring of Hurwitz integral quaternions. The vertices of the 24-cell form the group of units (i.e. the group of invertible elements) in the Hurwitz quaternion ring (this group is also known as the binary tetrahedral group). The vertices of the 24-cell are precisely the 24 Hurwitz quaternions with norm squared 1, and the vertices of the dual 24-cell are those with norm squared 2. The D4 root lattice is the dual of the F4 and is given by the subring of Hurwitz quaternions with even norm squared. Viewed as the 24 unit Hurwitz quaternions, the unit radius coordinates of the 24-cell represent (in antipodal pairs) the 12 rotations of a regular tetrahedron. Vertices of other convex regular 4-polytopes also form multiplicative groups of quaternions, but few of them generate a root lattice. Voronoi cells The Voronoi cells of the D4 root lattice are regular 24-cells. The corresponding Voronoi tessellation gives the tessellation of 4-dimensional Euclidean space by regular 24-cells, the 24-cell honeycomb. The 24-cells are centered at the D4 lattice points (Hurwitz quaternions with even norm squared) while the vertices are at the F4 lattice points with odd norm squared. Each 24-cell of this tessellation has 24 neighbors. With each of these it shares an octahedron. It also has 24 other neighbors with which it shares only a single vertex. Eight 24-cells meet at any given vertex in this tessellation. The Schläfli symbol for this tessellation is {3,4,3,3}. It is one of only three regular tessellations of R4. The unit balls inscribed in the 24-cells of this tessellation give rise to the densest known lattice packing of hyperspheres in 4 dimensions. The vertex configuration of the 24-cell has also been shown to give the highest possible kissing number in 4 dimensions. Radially equilateral honeycomb The dual tessellation of the 24-cell honeycomb {3,4,3,3} is the 16-cell honeycomb {3,3,4,3}. The third regular tessellation of four dimensional space is the tesseractic honeycomb {4,3,3,4}, whose vertices can be described by 4-integer Cartesian coordinates. The congruent relationships among these three tessellations can be helpful in visualizing the 24-cell, in particular the radial equilateral symmetry which it shares with the tesseract. A honeycomb of unit edge length 24-cells may be overlaid on a honeycomb of unit edge length tesseracts such that every vertex of a tesseract (every 4-integer coordinate) is also the vertex of a 24-cell (and tesseract edges are also 24-cell edges), and every center of a 24-cell is also the center of a tesseract. The 24-cells are twice as large as the tesseracts by 4-dimensional content (hypervolume), so overall there are two tesseracts for every 24-cell, only half of which are inscribed in a 24-cell. If those tesseracts are colored black, and their adjacent tesseracts (with which they share a cubical facet) are colored red, a 4-dimensional checkerboard results. Of the 24 center-to-vertex radii of each 24-cell, 16 are also the radii of a black tesseract inscribed in the 24-cell. The other 8 radii extend outside the black tesseract (through the centers of its cubical facets) to the centers of the 8 adjacent red tesseracts. Thus the 24-cell honeycomb and the tesseractic honeycomb coincide in a special way: 8 of the 24 vertices of each 24-cell do not occur at a vertex of a tesseract (they occur at the center of a tesseract instead). Each black tesseract is cut from a 24-cell by truncating it at these 8 vertices, slicing off 8 cubic pyramids (as in reversing Gosset's construction, but instead of being removed the pyramids are simply colored red and left in place). Eight 24-cells meet at the center of each red tesseract: each one meets its opposite at that shared vertex, and the six others at a shared octahedral cell. The red tesseracts are filled cells (they contain a central vertex and radii); the black tesseracts are empty cells. The vertex set of this union of two honeycombs includes the vertices of all the 24-cells and tesseracts, plus the centers of the red tesseracts. Adding the 24-cell centers (which are also the black tesseract centers) to this honeycomb yields a 16-cell honeycomb, the vertex set of which includes all the vertices and centers of all the 24-cells and tesseracts. The formerly empty centers of adjacent 24-cells become the opposite vertices of a unit edge length 16-cell. 24 half-16-cells (octahedral pyramids) meet at each formerly empty center to fill each 24-cell, and their octahedral bases are the 6-vertex octahedral facets of the 24-cell (shared with an adjacent 24-cell). Notice the complete absence of pentagons anywhere in this union of three honeycombs. Like the 24-cell, 4-dimensional Euclidean space itself is entirely filled by a complex of all the polytopes that can be built out of regular triangles and squares (except the 5-cell), but that complex does not require (or permit) any of the pentagonal polytopes. Rotations The regular convex 4-polytopes are an expression of their underlying symmetry which is known as SO(4), the group of rotations about a fixed point in 4-dimensional Euclidean space. The 3 Cartesian bases of the 24-cell There are three distinct orientations of the tesseractic honeycomb which could be made to coincide with the 24-cell honeycomb, depending on which of the 24-cell's three disjoint sets of 8 orthogonal vertices (which set of 4 perpendicular axes, or equivalently, which inscribed basis 16-cell) was chosen to align it, just as three tesseracts can be inscribed in the 24-cell, rotated with respect to each other. The distance from one of these orientations to another is an isoclinic rotation through 60 degrees (a double rotation of 60 degrees in each pair of orthogonal invariant planes, around a single fixed point). This rotation can be seen most clearly in the hexagonal central planes, where every hexagon rotates to change which of its three diameters is aligned with a coordinate system axis. Planes of rotation Rotations in 4-dimensional Euclidean space can be seen as the composition of two 2-dimensional rotations in completely orthogonal planes. Thus the general rotation in 4-space is a double rotation. There are two important special cases, called a simple rotation and an isoclinic rotation. Simple rotations In 3 dimensions a spinning polyhedron has a single invariant central plane of rotation. The plane is an invariant set because each point in the plane moves in a circle but stays within the plane. Only one of a polyhedron's central planes can be invariant during a particular rotation; the choice of invariant central plane, and the angular distance and direction it is rotated, completely specifies the rotation. Points outside the invariant plane also move in circles (unless they are on the fixed axis of rotation perpendicular to the invariant plane), but the circles do not lie within a central plane. When a 4-polytope is rotating with only one invariant central plane, the same kind of simple rotation is happening that occurs in 3 dimensions. One difference is that instead of a fixed axis of rotation, there is an entire fixed central plane in which the points do not move. The fixed plane is the one central plane that is completely orthogonal to the invariant plane of rotation. In the 24-cell, there is a simple rotation which will take any vertex directly to any other vertex, also moving most of the other vertices but leaving at least 2 and at most 6 other vertices fixed (the vertices that the fixed central plane intersects). The vertex moves along a great circle in the invariant plane of rotation between adjacent vertices of a great hexagon, a great square or a great digon, and the completely orthogonal fixed plane is a digon, a square or a hexagon, respectively. Double rotations The points in the completely orthogonal central plane are not constrained to be fixed. It is also possible for them to be rotating in circles, as a second invariant plane, at a rate independent of the first invariant plane's rotation: a double rotation in two perpendicular non-intersecting planes of rotation at once. In a double rotation there is no fixed plane or axis: every point moves except the center point. The angular distance rotated may be different in the two completely orthogonal central planes, but they are always both invariant: their circularly moving points remain within the plane as the whole plane tilts sideways in the completely orthogonal rotation. A rotation in 4-space always has (at least) two completely orthogonal invariant planes of rotation, although in a simple rotation the angle of rotation in one of them is 0. Double rotations come in two chiral forms: left and right rotations. In a double rotation each vertex moves in a spiral along two orthogonal great circles at once. Either the path is right-hand threaded (like most screws and bolts), moving along the circles in the "same" directions, or it is left-hand threaded (like a reverse-threaded bolt), moving along the circles in what we conventionally say are "opposite" directions (according to the right hand rule by which we conventionally say which way is "up" on each of the 4 coordinate axes). In double rotations of the 24-cell that take vertices to vertices, one invariant plane of rotation contains either a great hexagon, a great square, or only an axis (two vertices, a great digon). The completely orthogonal invariant plane of rotation will necessarily contain a great digon, a great square, or a great hexagon, respectively. The selection of an invariant plane of rotation, a rotational direction and angle through which to rotate it, and a rotational direction and angle through which to rotate its completely orthogonal plane, completely determines the nature of the rotational displacement. In the 24-cell there are several noteworthy kinds of double rotation permitted by these parameters. Isoclinic rotations When the angles of rotation in the two invariant planes are exactly the same, a remarkably symmetric transformation occurs: all the great circle planes Clifford parallel to the invariant planes become invariant planes of rotation themselves, through that same angle, and the 4-polytope rotates isoclinically in many directions at once. Each vertex moves an equal distance in four orthogonal directions at the same time. In the 24-cell any isoclinic rotation through 60 degrees in a hexagonal plane takes each vertex to a vertex two edge lengths away, rotates all 16 hexagons by 60 degrees, and takes every great circle polygon (square, hexagon or triangle) to a Clifford parallel great circle polygon of the same kind 120 degrees away. An isoclinic rotation is also called a Clifford displacement, after its discoverer. The 24-cell in the double rotation animation appears to turn itself inside out. It appears to, because it actually does, reversing the chirality of the whole 4-polytope just the way your bathroom mirror reverses the chirality of your image by a 180 degree reflection. Each 360 degree isoclinic rotation is as if the 24-cell surface had been stripped off like a glove and turned inside out, making a right-hand glove into a left-hand glove (or vice versa). In a simple rotation of the 24-cell in a hexagonal plane, each vertex in the plane rotates first along an edge to an adjacent vertex 60 degrees away. But in an isoclinic rotation in two completely orthogonal planes one of which is a great hexagon, each vertex rotates first to a vertex two edge lengths away ( and 120° distant). The double 60-degree rotation's helical geodesics pass through every other vertex, missing the vertices in between. Each chord of the helical geodesic crosses between two Clifford parallel hexagon central planes, and lies in another hexagon central plane that intersects them both. The chords meet at a 60° angle, but since they lie in different planes they form a helix not a triangle. Three chords and 360° of rotation takes the vertex to an adjacent vertex, not back to itself. The helix of chords closes into a loop only after six chords: a 720° rotation twice around the 24-cell on a skew hexagram with edges. Even though all 24 vertices and all the hexagons rotate at once, a 360 degree isoclinic rotation moves each vertex only halfway around its circuit. After 360 degrees each helix has departed from 3 vertices and reached a fourth vertex adjacent to the original vertex, but has not arrived back exactly at the vertex it departed from. Each central plane (every hexagon or square in the 24-cell) has rotated 360 degrees and been tilted sideways all the way around 360 degrees back to its original position (like a coin flipping twice), but the 24-cell's orientation in the 4-space in which it is embedded is now different. Because the 24-cell is now inside-out, if the isoclinic rotation is continued in the same direction through another 360 degrees, the 24 moving vertices will pass through the other half of the vertices that were missed on the first revolution (the 12 antipodal vertices of the 12 that were hit the first time around), and each isoclinic geodesic will arrive back at the vertex it departed from, forming a closed six-chord helical loop. It takes a 720 degree isoclinic rotation for each hexagram2 geodesic to complete a circuit through every second vertex of its six vertices by winding around the 24-cell twice, returning the 24-cell to its original chiral orientation. The hexagonal winding path that each vertex takes as it loops twice around the 24-cell forms a double helix bent into a Möbius ring, so that the two strands of the double helix form a continuous single strand in a closed loop. In the first revolution the vertex traverses one 3-chord strand of the double helix; in the second revolution it traverses the second 3-chord strand, moving in the same rotational direction with the same handedness (bending either left or right) throughout. Although this isoclinic Möbius ring is a circular spiral through all 4 dimensions, not a 2-dimensional circle, like a great circle it is a geodesic because it is the shortest path from vertex to vertex. Clifford parallel polytopes Two planes are also called isoclinic if an isoclinic rotation will bring them together. The isoclinic planes are precisely those central planes with Clifford parallel geodesic great circles. Clifford parallel great circles do not intersect, so isoclinic great circle polygons have disjoint vertices. In the 24-cell every hexagonal central plane is isoclinic to three others, and every square central plane is isoclinic to five others. We can pick out 4 mutually isoclinic (Clifford parallel) great hexagons (four different ways) covering all 24 vertices of the 24-cell just once (a hexagonal fibration). We can pick out 6 mutually isoclinic (Clifford parallel) great squares (three different ways) covering all 24 vertices of the 24-cell just once (a square fibration). Every isoclinic rotation taking vertices to vertices corresponds to a discrete fibration. Two dimensional great circle polygons are not the only polytopes in the 24-cell which are parallel in the Clifford sense. Congruent polytopes of 2, 3 or 4 dimensions can be said to be Clifford parallel in 4 dimensions if their corresponding vertices are all the same distance apart. The three 16-cells inscribed in the 24-cell are Clifford parallels. Clifford parallel polytopes are completely disjoint polytopes. A 60 degree isoclinic rotation in hexagonal planes takes each 16-cell to a disjoint 16-cell. Like all double rotations, isoclinic rotations come in two chiral forms: there is a disjoint 16-cell to the left of each 16-cell, and another to its right. All Clifford parallel 4-polytopes are related by an isoclinic rotation, but not all isoclinic polytopes are Clifford parallels (completely disjoint). The three 8-cells in the 24-cell are isoclinic but not Clifford parallel. Like the 16-cells, they are rotated 60 degrees isoclinically with respect to each other, but their vertices are not all disjoint (and therefore not all equidistant). Each vertex occurs in two of the three 8-cells (as each 16-cell occurs in two of the three 8-cells). Isoclinic rotations relate the convex regular 4-polytopes to each other. An isoclinic rotation of a single 16-cell will generate a 24-cell. A simple rotation of a single 16-cell will not, because its vertices will not reach either of the other two 16-cells' vertices in the course of the rotation. An isoclinic rotation of the 24-cell will generate the 600-cell, and an isoclinic rotation of the 600-cell will generate the 120-cell. (Or they can all be generated directly by an isoclinic rotation of the 16-cell, generating isoclinic copies of itself.) The different convex regular 4-polytopes nest inside each other, and multiple instances of the same 4-polytope hide next to each other in the Clifford parallel spaces that comprise the 3-sphere. For an object of more than one dimension, the only way to reach these parallel subspaces directly is by isoclinic rotation. Rings In the 24-cell there are sets of rings of six different kinds, described separately in detail in other sections of this article. This section describes how the different kinds of rings are intertwined. The 24-cell contains four kinds of geodesic fibers (polygonal rings running through vertices): great circle squares and their isoclinic helix octagrams, and great circle hexagons and their isoclinic helix hexagrams. It also contains two kinds of cell rings (chains of octahedra bent into a ring in the fourth dimension): four octahedra connected vertex-to-vertex and bent into a square, and six octahedra connected face-to-face and bent into a hexagon. 4-cell rings Four unit-edge-length octahedra can be connected vertex-to-vertex along a common axis of length 4. The axis can then be bent into a square of edge length . Although it is possible to do this in a space of only three dimensions, that is not how it occurs in the 24-cell. Although the axes of the four octahedra occupy the same plane, forming one of the 18 great squares of the 24-cell, each octahedron occupies a different 3-dimensional hyperplane, and all four dimensions are utilized. The 24-cell can be partitioned into 6 such 4-cell rings (three different ways), mutually interlinked like adjacent links in a chain (but these links all have a common center). An isoclinic rotation in the great square plane by a multiple of 90° takes each octahedron in the ring to an octahedron in the ring. 6-cell rings Six regular octahedra can be connected face-to-face along a common axis that passes through their centers of volume, forming a stack or column with only triangular faces. In a space of four dimensions, the axis can then be bent 60° in the fourth dimension at each of the six octahedron centers, in a plane orthogonal to all three orthogonal central planes of each octahedron, such that the top and bottom triangular faces of the column become coincident. The column becomes a ring around a hexagonal axis. The 24-cell can be partitioned into 4 such rings (four different ways), mutually interlinked. Because the hexagonal axis joins cell centers (not vertices), it is not a great hexagon of the 24-cell. However, six great hexagons can be found in the ring of six octahedra, running along the edges of the octahedra. In the column of six octahedra (before it is bent into a ring) there are six spiral paths along edges running up the column: three parallel helices spiraling clockwise, and three parallel helices spiraling counterclockwise. Each clockwise helix intersects each counterclockwise helix at two vertices three edge lengths apart. Bending the column into a ring changes these helices into great circle hexagons. The ring has two sets of three great hexagons, each on three Clifford parallel great circles. The great hexagons in each parallel set of three do not intersect, but each intersects the other three great hexagons (to which it is not Clifford parallel) at two antipodal vertices. A simple rotation in any of the great hexagon planes by a multiple of 60° rotates only that hexagon invariantly, taking each vertex in that hexagon to a vertex in the same hexagon. An isoclinic rotation by 60° in any of the six great hexagon planes rotates all three Clifford parallel great hexagons invariantly, and takes each octahedron in the ring to a non-adjacent octahedron in the ring. Each isoclinically displaced octahedron is also rotated itself. After a 360° isoclinic rotation each octahedron is back in the same position, but in a different orientation. In a 720° isoclinic rotation, its vertices are returned to their original orientation. Four Clifford parallel great hexagons comprise a discrete fiber bundle covering all 24 vertices in a Hopf fibration. The 24-cell has four such discrete hexagonal fibrations . Each great hexagon belongs to just one fibration, and the four fibrations are defined by disjoint sets of four great hexagons each. Each fibration is the domain (container) of a unique left-right pair of isoclinic rotations (left and right Hopf fiber bundles). Four cell-disjoint 6-cell rings also comprise each discrete fibration defined by four Clifford parallel great hexagons. Each 6-cell ring contains only 18 of the 24 vertices, and only 6 of the 16 great hexagons, which we see illustrated above running along the cell ring's edges: 3 spiraling clockwise and 3 counterclockwise. Those 6 hexagons running along the cell ring's edges are not among the set of four parallel hexagons which define the fibration. For example, one of the four 6-cell rings in fibration contains 3 parallel hexagons running clockwise along the cell ring's edges from fibration , and 3 parallel hexagons running counterclockwise along the cell ring's edges from fibration , but that cell ring contains no great hexagons from fibration or fibration . The 24-cell contains 16 great hexagons, divided into four disjoint sets of four hexagons, each disjoint set uniquely defining a fibration. Each fibration is also a distinct set of four cell-disjoint 6-cell rings. The 24-cell has exactly 16 distinct 6-cell rings. Each 6-cell ring belongs to just one of the four fibrations. Helical hexagrams and their isoclines Another kind of geodesic fiber, the helical hexagram isoclines, can be found within a 6-cell ring of octahedra. Each of these geodesics runs through every second vertex of a skew hexagram2, which in the unit-radius, unit-edge-length 24-cell has six edges. The hexagram does not lie in a single central plane, but is composed of six linked chords from the six different hexagon great circles in the 6-cell ring. The isocline geodesic fiber is the path of an isoclinic rotation, a helical rather than simply circular path around the 24-cell which links vertices two edge lengths apart and consequently must wrap twice around the 24-cell before completing its six-vertex loop. Rather than a flat hexagon, it forms a skew hexagram out of two three-sided 360 degree half-loops: open triangles joined end-to-end to each other in a six-sided Möbius loop. Each 6-cell ring contains six such hexagram isoclines, three black and three white, that connect even and odd vertices respectively. Each of the three black-white pairs of isoclines belongs to one of the three fibrations in which the 6-cell ring occurs. Each fibration's right (or left) rotation traverses two black isoclines and two white isoclines in parallel, rotating all 24 vertices. Beginning at any vertex at one end of the column of six octahedra, we can follow an isoclinic path of chords of an isocline from octahedron to octahedron. In the 24-cell the edges are great hexagon edges (and octahedron edges); in the column of six octahedra we see six great hexagons running along the octahedra's edges. The chords are great hexagon diagonals, joining great hexagon vertices two edges apart. We find them in the ring of six octahedra running from a vertex in one octahedron to a vertex in the next octahedron, passing through the face shared by the two octahedra (but not touching any of the face's 3 vertices). Each chord is a chord of just one great hexagon (an edge of a great triangle inscribed in that great hexagon), but successive chords belong to different great hexagons. At each vertex the isoclinic path of chords bends 60 degrees in two central planes at once: 60 degrees around the great hexagon that the chord before the vertex belongs to, and 60 degrees into the plane of a different great hexagon entirely, that the chord after the vertex belongs to. Thus the path follows one great hexagon from each octahedron to the next, but switches to another of the six great hexagons in the next link of the hexagram2 path. Followed along the column of six octahedra (and "around the end" where the column is bent into a ring) the path may at first appear to be zig-zagging between three adjacent parallel hexagonal central planes (like a Petrie polygon), but it is not: any isoclinic path we can pick out always zig-zags between two sets of three adjacent parallel hexagonal central planes, intersecting only every even (or odd) vertex and never changing its inherent even/odd parity, as it visits all six of the great hexagons in the 6-cell ring in rotation. When it has traversed one chord from each of the six great hexagons, after 720 degrees of isoclinic rotation (either left or right), it closes its skew hexagram and begins to repeat itself, circling again through the black (or white) vertices and cells. At each vertex, there are four great hexagons and four hexagram isoclines (all black or all white) that cross at the vertex. Four hexagram isoclines (two black and two white) comprise a unique (left or right) fiber bundle of isoclines covering all 24 vertices in each distinct (left or right) isoclinic rotation. Each fibration has a unique left and right isoclinic rotation, and corresponding unique left and right fiber bundles of isoclines. There are 16 distinct hexagram isoclines in the 24-cell (8 black and 8 white). Each isocline is a skew Clifford polygon of no inherent chirality, but acts as a left (or right) isocline when traversed by a left (or right) rotation in different fibrations. Helical octagrams and their isoclines The 24-cell contains 18 helical octagram isoclines (9 black and 9 white). Three pairs of octagram edge-helices are found in each of the three inscribed 16-cells, described elsewhere as the helical construction of the 16-cell. In summary, each 16-cell can be decomposed (three different ways) into a left-right pair of 8-cell rings of -edged tetrahedral cells. Each 8-cell ring twists either left or right around an axial octagram helix of eight chords. In each 16-cell there are exactly 6 distinct helices, identical octagrams which each circle through all eight vertices. Each acts as either a left helix or a right helix or a Petrie polygon in each of the six distinct isoclinic rotations (three left and three right), and has no inherent chirality except in respect to a particular rotation. Adjacent vertices on the octagram isoclines are = 90° apart, so the circumference of the isocline is 4𝝅. An isoclinic rotation by 90° in great square invariant planes takes each vertex to its antipodal vertex, four vertices away in either direction along the isocline, and = 180° distant across the diameter of the isocline. Each of the 3 fibrations of the 24-cell's 18 great squares corresponds to a distinct left (and right) isoclinic rotation in great square invariant planes. Each 60° step of the rotation takes 6 disjoint great squares (2 from each 16-cell) to great squares in a neighboring 16-cell, on 8-chord helical isoclines characteristic of the 16-cell. In the 24-cell, these 18 helical octagram isoclines can be found within the six orthogonal 4-cell rings of octahedra. Each 4-cell ring has cells bonded vertex-to-vertex around a great square axis, and we find antipodal vertices at opposite vertices of the great square. A chord (the diameter of the great square and of the isocline) connects them. Boundary cells describes how the axes of the 24-cell's octahedral cells are the edges of the 16-cell's tetrahedral cells, each tetrahedron is inscribed in a (tesseract) cube, and each octahedron is inscribed in a pair of cubes (from different tesseracts), bridging them. The vertex-bonded octahedra of the 4-cell ring also lie in different tesseracts. The isocline's four diameter chords form an octagram8{4}=4{2} with edges that each run from the vertex of one cube and octahedron and tetrahedron, to the vertex of another cube and octahedron and tetrahedron (in a different tesseract), straight through the center of the 24-cell on one of the 12 axes. The octahedra in the 4-cell rings are vertex-bonded to more than two other octahedra, because three 4-cell rings (and their three axial great squares, which belong to different 16-cells) cross at 90° at each bonding vertex. At that vertex the octagram makes two right-angled turns at once: 90° around the great square, and 90° orthogonally into a different 4-cell ring entirely. The 180° four-edge arc joining two ends of each diameter chord of the octagram runs through the volumes and opposite vertices of two face-bonded tetrahedra (in the same 16-cell), which are also the opposite vertices of two vertex-bonded octahedra in different 4-cell rings (and different tesseracts). The 720° octagram isocline runs through 8 vertices of the four-cell ring and through the volumes of 16 tetrahedra. At each vertex, there are three great squares and six octagram isoclines (three black-white pairs) that cross at the vertex. This is the characteristic rotation of the 16-cell, not the 24-cell's characteristic rotation, and it does not take whole 16-cells of the 24-cell to each other the way the 24-cell's rotation in great hexagon planes does. Characteristic orthoscheme Every regular 4-polytope has its characteristic 4-orthoscheme, an irregular 5-cell. The characteristic 5-cell of the regular 24-cell is represented by the Coxeter-Dynkin diagram , which can be read as a list of the dihedral angles between its mirror facets. It is an irregular tetrahedral pyramid based on the characteristic tetrahedron of the regular octahedron. The regular 24-cell is subdivided by its symmetry hyperplanes into 1152 instances of its characteristic 5-cell that all meet at its center. The characteristic 5-cell (4-orthoscheme) has four more edges than its base characteristic tetrahedron (3-orthoscheme), joining the four vertices of the base to its apex (the fifth vertex of the 4-orthoscheme, at the center of the regular 24-cell). If the regular 24-cell has radius and edge length 𝒍 = 1, its characteristic 5-cell's ten edges have lengths , , around its exterior right-triangle face (the edges opposite the characteristic angles 𝟀, 𝝉, 𝟁), plus , , (the other three edges of the exterior 3-orthoscheme facet the characteristic tetrahedron, which are the characteristic radii of the octahedron), plus , , , (edges which are the characteristic radii of the 24-cell). The 4-edge path along orthogonal edges of the orthoscheme is , , , , first from a 24-cell vertex to a 24-cell edge center, then turning 90° to a 24-cell face center, then turning 90° to a 24-cell octahedral cell center, then turning 90° to the 24-cell center. Reflections The 24-cell can be constructed by the reflections of its characteristic 5-cell in its own facets (its tetrahedral mirror walls). Reflections and rotations are related: a reflection in an even number of intersecting mirrors is a rotation. Consequently, regular polytopes can be generated by reflections or by rotations. For example, any 720° isoclinic rotation of the 24-cell in a hexagonal invariant plane takes each of the 24 vertices to and through 5 other vertices and back to itself, on a skew hexagram2 geodesic isocline that winds twice around the 3-sphere on every second vertex of the hexagram. Any set of four orthogonal pairs of antipodal vertices (the 8 vertices of one of the three inscribed 16-cells) performing half such an orbit visits 3 * 8 = 24 distinct vertices and generates the 24-cell sequentially in 3 steps of a single 360° isoclinic rotation, just as any single characteristic 5-cell reflecting itself in its own mirror walls generates the 24 vertices simultaneously by reflection. Tracing the orbit of one such 16-cell vertex during the 360° isoclinic rotation reveals more about the relationship between reflections and rotations as generative operations. The vertex follows an isocline (a doubly curved geodesic circle) rather than an ordinary great circle. The isocline connects vertices two edge lengths apart, but curves away from the great circle path over the two edges connecting those vertices, missing the vertex in between. Although the isocline does not follow any one great circle, it is contained within a ring of another kind: in the 24-cell it stays within a 6-cell ring of spherical octahedral cells, intersecting one vertex in each cell, and passing through the volume of two adjacent cells near the missed vertex. Chiral symmetry operations A symmetry operation is a rotation or reflection which leaves the object indistinguishable from itself before the transformation. The 24-cell has 1152 distinct symmetry operations (576 rotations and 576 reflections). Each rotation is equivalent to two reflections, in a distinct pair of non-parallel mirror planes. Pictured are sets of disjoint great circle polygons, each in a distinct central plane of the 24-cell. For example, {24/4}=4{6} is an orthogonal projection of the 24-cell picturing 4 of its [16] great hexagon planes. The 4 planes lie Clifford parallel to the projection plane and to each other, and their great polygons collectively constitute a discrete Hopf fibration of 4 non-intersecting great circles which visit all 24 vertices just once. Each row of the table describes a class of distinct rotations. Each rotation class takes the left planes pictured to the corresponding right planes pictured. The vertices of the moving planes move in parallel along the polygonal isocline paths pictured. For example, the rotation class consists of [32] distinct rotational displacements by an arc-distance of = 120° between 16 great hexagon planes represented by quaternion group and a corresponding set of 16 great hexagon planes represented by quaternion group . One of the [32] distinct rotations of this class moves the representative vertex coordinate to the vertex coordinate . In a rotation class each quaternion group may be representative not only of its own fibration of Clifford parallel planes but also of the other congruent fibrations. For example, rotation class takes the 4 hexagon planes of to the 4 hexagon planes of which are 120° away, in an isoclinic rotation. But in a rigid rotation of this kind, all [16] hexagon planes move in congruent rotational displacements, so this rotation class also includes , and . The name is the conventional representation for all [16] congruent plane displacements. These rotation classes are all subclasses of which has [32] distinct rotational displacements rather than [16] because there are two chiral ways to perform any class of rotations, designated its left rotations and its right rotations. The [16] left displacements of this class are not congruent with the [16] right displacements, but enantiomorphous like a pair of shoes. Each left (or right) isoclinic rotation takes [16] left planes to [16] right planes, but the left and right planes correspond differently in the left and right rotations. The left and right rotational displacements of the same left plane take it to different right planes. Each rotation class (table row) describes a distinct left (and right) isoclinic rotation. The left (or right) rotations carry the left planes to the right planes simultaneously, through a characteristic rotation angle. For example, the rotation moves all [16] hexagonal planes at once by = 120° each. Repeated 6 times, this left (or right) isoclinic rotation moves each plane 720° and back to itself in the same orientation, passing through all 4 planes of the left set and all 4 planes of the right set once each. The picture in the isocline column represents this union of the left and right plane sets. In the example it can be seen as a set of 4 Clifford parallel skew hexagrams, each having one edge in each great hexagon plane, and skewing to the left (or right) at each vertex throughout the left (or right) isoclinic rotation. Visualization Cell rings The 24-cell is bounded by 24 octahedral cells. For visualization purposes, it is convenient that the octahedron has opposing parallel faces (a trait it shares with the cells of the tesseract and the 120-cell). One can stack octahedrons face to face in a straight line bent in the 4th direction into a great circle with a circumference of 6 cells. The cell locations lend themselves to a hyperspherical description. Pick an arbitrary cell and label it the "North Pole". Eight great circle meridians (two cells long) radiate out in 3 dimensions, converging at the 3rd "South Pole" cell. This skeleton accounts for 18 of the 24 cells (2 + ). See the table below. There is another related great circle in the 24-cell, the dual of the one above. A path that traverses 6 vertices solely along edges resides in the dual of this polytope, which is itself since it is self dual. These are the hexagonal geodesics described above. One can easily follow this path in a rendering of the equatorial cuboctahedron cross-section. Starting at the North Pole, we can build up the 24-cell in 5 latitudinal layers. With the exception of the poles, each layer represents a separate 2-sphere, with the equator being a great 2-sphere. The cells labeled equatorial in the following table are interstitial to the meridian great circle cells. The interstitial "equatorial" cells touch the meridian cells at their faces. They touch each other, and the pole cells at their vertices. This latter subset of eight non-meridian and pole cells has the same relative position to each other as the cells in a tesseract (8-cell), although they touch at their vertices instead of their faces. The 24-cell can be partitioned into cell-disjoint sets of four of these 6-cell great circle rings, forming a discrete Hopf fibration of four non-intersecting linked rings. One ring is "vertical", encompassing the pole cells and four meridian cells. The other three rings each encompass two equatorial cells and four meridian cells, two from the northern hemisphere and two from the southern. Note this hexagon great circle path implies the interior/dihedral angle between adjacent cells is 180 - 360/6 = 120 degrees. This suggests you can adjacently stack exactly three 24-cells in a plane and form a 4-D honeycomb of 24-cells as described previously. One can also follow a great circle route, through the octahedrons' opposing vertices, that is four cells long. These are the square geodesics along four chords described above. This path corresponds to traversing diagonally through the squares in the cuboctahedron cross-section. The 24-cell is the only regular polytope in more than two dimensions where you can traverse a great circle purely through opposing vertices (and the interior) of each cell. This great circle is self dual. This path was touched on above regarding the set of 8 non-meridian (equatorial) and pole cells. The 24-cell can be equipartitioned into three 8-cell subsets, each having the organization of a tesseract. Each of these subsets can be further equipartitioned into two non-intersecting linked great circle chains, four cells long. Collectively these three subsets now produce another, six ring, discrete Hopf fibration. Parallel projections The vertex-first parallel projection of the 24-cell into 3-dimensional space has a rhombic dodecahedral envelope. Twelve of the 24 octahedral cells project in pairs onto six square dipyramids that meet at the center of the rhombic dodecahedron. The remaining 12 octahedral cells project onto the 12 rhombic faces of the rhombic dodecahedron. The cell-first parallel projection of the 24-cell into 3-dimensional space has a cuboctahedral envelope. Two of the octahedral cells, the nearest and farther from the viewer along the w-axis, project onto an octahedron whose vertices lie at the center of the cuboctahedron's square faces. Surrounding this central octahedron lie the projections of 16 other cells, having 8 pairs that each project to one of the 8 volumes lying between a triangular face of the central octahedron and the closest triangular face of the cuboctahedron. The remaining 6 cells project onto the square faces of the cuboctahedron. This corresponds with the decomposition of the cuboctahedron into a regular octahedron and 8 irregular but equal octahedra, each of which is in the shape of the convex hull of a cube with two opposite vertices removed. The edge-first parallel projection has an elongated hexagonal dipyramidal envelope, and the face-first parallel projection has a nonuniform hexagonal bi-antiprismic envelope. Perspective projections The vertex-first perspective projection of the 24-cell into 3-dimensional space has a tetrakis hexahedral envelope. The layout of cells in this image is similar to the image under parallel projection. The following sequence of images shows the structure of the cell-first perspective projection of the 24-cell into 3 dimensions. The 4D viewpoint is placed at a distance of five times the vertex-center radius of the 24-cell. Related polytopes Three Coxeter group constructions There are two lower symmetry forms of the 24-cell, derived as a rectified 16-cell, with B4 or [3,3,4] symmetry drawn bicolored with 8 and 16 octahedral cells. Lastly it can be constructed from D4 or [31,1,1] symmetry, and drawn tricolored with 8 octahedra each. Related complex polygons The regular complex polygon 4{3}4, or contains the 24 vertices of the 24-cell, and 24 4-edges that correspond to central squares of 24 of 48 octahedral cells. Its symmetry is 4[3]4, order 96. The regular complex polytope 3{4}3, or , in has a real representation as a 24-cell in 4-dimensional space. 3{4}3 has 24 vertices, and 24 3-edges. Its symmetry is 3[4]3, order 72. Related 4-polytopes Several uniform 4-polytopes can be derived from the 24-cell via truncation: truncating at 1/3 of the edge length yields the truncated 24-cell; truncating at 1/2 of the edge length yields the rectified 24-cell; and truncating at half the depth to the dual 24-cell yields the bitruncated 24-cell, which is cell-transitive. The 96 edges of the 24-cell can be partitioned into the golden ratio to produce the 96 vertices of the snub 24-cell. This is done by first placing vectors along the 24-cell's edges such that each two-dimensional face is bounded by a cycle, then similarly partitioning each edge into the golden ratio along the direction of its vector. An analogous modification to an octahedron produces an icosahedron, or "snub octahedron." The 24-cell is the unique convex self-dual regular Euclidean polytope that is neither a polygon nor a simplex. Relaxing the condition of convexity admits two further figures: the great 120-cell and grand stellated 120-cell. With itself, it can form a polytope compound: the compound of two 24-cells. Related uniform polytopes The 24-cell can also be derived as a rectified 16-cell:
Mathematics
Four-dimensional space
null