source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/Morchella%20septimelata
Morchella septimelata is a species of fungus in the family Morchellaceae described as new to science in 2012. Occurring in western North America, it has been collected from British Columbia, Montana and Oregon, where it fruits at elevations of in coniferous forests in years following slight to moderate burning. Morchella septimelata is one of four species of wildfire-adapted morel in western North America, the others being M. capitata, M. sextelata, and M. tomentosa. M. septimelata cannot be reliably distinguished from M. sextelata without the use of DNA analysis. M. septimelata, identified as phylogenetic species "Mel-7", has also been found in Europe, Turkey, China, and Australia but it remains unclear whether dispersal between these distant locations occurred naturally or through accidental introduction by humans. In 2014, Richard et al. clarified the taxonomic status of this species, retaining the name Morchella eximia Boudier (1909) over M. septimelata.
https://en.wikipedia.org/wiki/Ignacio%20Jord%C3%A1n%20Claudio%20de%20Asso%20y%20del%20R%C3%ADo
Ignacio Jordán Claudio de Asso y del Río (June 4, 1742 – May 21, 1814) was a Spanish diplomat, naturalist, lawyer and historian. He sometimes used the pseudonym of Melchor de Azagra. Biography Of noble birth, he received an excellent education, studying Classical Greek and Latin in the college known as the Escuelas Pías of Zaragoza (1755) and philosophy under the Jesuits at the Real e Imperial Colegio de Nobles de Nuestra Señora y Santiago de Cordellas, located in Barcelona (1756). He studied at the University of Cervera, where he graduated with a bachelor of arts in 1760, and at the University of Zaragoza, where he studied jurisprudence, graduating in 1764. He worked as a jurist from 1765 to 1776 and traveled for three years across Europe and from 1771 to 1775 he published in Madrid, collaboratively and also alone, a massive work on jurisprudence (Instituciones del Derecho Civil de Castilla, 1771; El Fuero Viejo de Castilla, 1771 (with Miguel de Manuel y Rodríguez); El ordenamiento de leyes que D. Alfonso XI hizo en las cortes de Alcalá de Henares el año de mil trescientos y cuarenta y ocho, 1774 (also with Miguel de Manuel y Rodríguez)). In 1776, he began work as a diplomat and served as consul at Dunkirk (1776), consul general at Amsterdam (1776–1787) and consul at Bordeaux (1787–1791). While a diplomat, he concerned himself with scientific and economic matters, and published in Amsterdam works on the botany, zoology, and mineralogy of Aragon (Synopsis estirpium indigenarum Aragoniae, 1779; Mantissa stirpium indigenarum Aragoniae, 1781; Introductio in oryctographiam, et zoologiam Aragoniae, 1784; Enumeratio stirpium in Aragonia noviter detectarum, 1784). He described for the first time the fish Argyrosomus regius and Salaria fluviatilis. He also translated and published the work of various Aragonese writers and poets of the medieval and early modern periods (Bibliotheca Arabico-Aragonensis, 1782; Joannis Sobrarii, Carmina, 1783; Clariorum Aragonensium mon
https://en.wikipedia.org/wiki/Pickering%20series
The Pickering series (also known as the Pickering–Fowler series) consists of three lines of singly ionized helium found, usually in absorption, in the spectra of hot stars like Wolf–Rayet stars. The name comes from Edward Charles Pickering and Alfred Fowler. The lines are produced by transitions from a higher energy level of an electron to a level with principal quantum number n = 4. The lines have wavelengths: 4339 Å (n = 10 to n = 4) 4541 Å (n = 9 to n = 4) 4859 Å (n = 8 to n = 4) 5412 Å (n = 7 to n = 4) 6560 Å (n = 6 to n = 4) 10124 Å (n = 5 to n = 4) The transitions from the even-n states overlap with hydrogen lines and are therefore masked in typical absorption stellar spectra. However, they are seen in emission in the spectra of Wolf-Rayet stars, as these stars have little or no hydrogen. In 1896, Pickering published observations of previously unknown lines in the spectra of the star Zeta Puppis. Pickering attributed the observation to a new form of hydrogen with half-integer transition levels. Fowler managed to produce similar lines from a hydrogen–helium mixture in 1912, and supported Pickering's conclusion as to their origin. Niels Bohr, however, included an analysis of the series in his 'trilogy' on atomic structure and concluded that Pickering and Fowler were wrong and that the spectral lines arise instead from ionised helium, He+. Fowler was initially skeptical but was ultimately convinced that Bohr was correct, and by 1915 "spectroscopists had transferred [the Pickering series] definitively [from hydrogen] to helium." Bohr's theoretical work on the Pickering series had demonstrated the need for "a re-examination of problems that seemed already to have been solved within classical theories" and provided important confirmation for his atomic theory. See also Hydrogen spectral series
https://en.wikipedia.org/wiki/Qualified%20digital%20certificate
In the context of Regulation (EU) No 910/2014 (eIDAS), a qualified digital certificate is a public key certificate issued by a trust service provider which has government-issued qualifications. The certificate is designed to ensure the authenticity and data integrity of an electronic signature and its accompanying message and/or attached data. Description eIDAS defines several tiers of electronic signatures that can be used in conducting public sector and private transactions within and across the borders of EU member states. A qualified digital certificate, in addition to other specific services provided by a qualified trust service provider, is required to elevate the status of an electronic signature to that of being considered a qualified electronic signature. Using cryptography, the digital certificate, also known as a public key certificate, contains information to link it to its owner and the digital signature of the trust entity that verifies the authenticity of the content that has been signed. According to eIDAS, to be considered a qualified digital certificate, the certificate must meet the requirements provided in Annex I of Regulation (EU) No 910/2014, including, but not limited to: Identification that the certificate is a qualified certificate for electronic signature Identification of the qualified trust service provider who issued the qualified certificate, including such information Corresponding electronic signature validation data and electronic signature creation data Indication of the certificate's period of validity Unique certificate identity code of the trust service provider Qualified trust service provider's advanced electronic signature or electronic seal Vision The need for non-repudiation and authentication of electronic signatures was originally addressed in the Electronic Signatures Directive 1999/93/EC to help facilitate secure transactions, specifically those that occur across the borders of EU Member states. The eIDAS Regul
https://en.wikipedia.org/wiki/Dense%20regular%20connective%20tissue
Dense regular connective tissue (DRCT) provides connection between different tissues in the human body. The collagen fibers in dense regular connective tissue are bundled in a parallel fashion. DRCT is divided into white fibrous connective tissue and yellow fibrous connective tissue, both of which occur in two forms: cord arrangement and sheath arrangement. In cord arrangement, bundles of collagen and matrix are distributed in regular alternate patterns. In sheath arrangement, collagen bundles and matrix are distributed in irregular patterns, sometimes in the form of a network. It is similar to areolar tissue, but in DRCT elastic fibers are completely absent. Structures formed An example of their use is in tendons, which connect muscle to bone and derive their strength from the regular, longitudinal arrangement of bundles of collagen fibers. Ligaments bind bone to bone and are similar in structure to tendons. Aponeuroses are layers of flat, broad tendons that join muscles and the body parts the muscles act upon, whether it be bone or muscle. Functions Dense regular connective tissue has great tensile strength that resists pulling forces especially well in one direction. DRCT has a very poor blood supply, which is why damaged tendons and ligaments are slow to heal.
https://en.wikipedia.org/wiki/Maurice%20Janet
Maurice Janet (1888–1983) was a French mathematician. Education and career In 1912, as a student he visited the University of Göttingen. He was a professor at the University of Caen. He was an Invited Speaker of the International Congress of Mathematicians in 1924 in Toronto, in 1932 in Zürich, and in 1936 in Oslo. Named in his honor are Janet bases, Janet sequences and a related algorithm in the theory of systems of partial differential equations. In 1926, he proved results that were later generalized by John Forbes Nash Jr. in his embedding theorem. In 1948, Janet was the president of the Société Mathématique de France. He was a close friend of the mathematician Ernest Vessiot. Selected publications Articles Les systèmes d'équations aux dérivées partielles, Journal de mathématiques pures et appliquées 8 ser., t. 3 (1920), pages 65–123. (paper in which what is now called the Janet basis was introduced) Books
https://en.wikipedia.org/wiki/Key%20Biodiversity%20Area
The Key Biodiversity Areas (KBA) approach helps to identify and designate areas of international importance in terms of biodiversity conservation using globally standardised criteria. KBAs extend the Important Bird Area (IBA) concept to other taxonomic groups and are now being identified in many parts of the world, by a range of organisations. Examples include Important Plant Areas (IPAs), Ecologically and Biologically Significant Areas (EBSAs) in the High Seas, Alliance for Zero Extinction (AZE) sites, Prime Butterfly Areas, Important Mammal Areas and Important Sites for Freshwater Biodiversity, with prototype criteria developed for freshwater molluscs and fish and for marine systems. The determination of KBAs often brings sites onto the conservation agenda that hadn't previously been identified as needing protection due to the nature of the two non-exclusive criteria used to determine them; vulnerability; and irreplaceability. The KBA global standard was published in 2016. Objectives Develop technical and conservation capacity within individual countries and on a global scale Develop partnerships between key organisations – both governmental and nongovernmental – concerned with site conservation Build broad understanding of the process, and broad ownership of the final site list Focus any new survey work on the most important gaps in knowledge See also Conservation biology Ecoregions Biodiversity Crisis Ecoregions High conservation value area High-Biodiversity Wilderness Areas Biodiversity Hotspots Biosphere Reserves Site-based conservation Protected Areas
https://en.wikipedia.org/wiki/Ergosphere
The ergosphere is a region located outside a rotating black hole's outer event horizon. Its name was proposed by Remo Ruffini and John Archibald Wheeler during the Les Houches lectures in 1971 and is derived from the Greek word (ergon), which means "work". It received this name because it is theoretically possible to extract energy and mass from this region. The ergosphere touches the event horizon at the poles of a rotating black hole and extends to a greater radius at the equator. A black hole with modest angular momentum has an ergosphere with a shape approximated by an oblate spheroid, while faster spins produce a more pumpkin-shaped ergosphere. The equatorial (maximal) radius of an ergosphere is the Schwarzschild radius, the radius of a non-rotating black hole. The polar (minimal) radius is also the polar (minimal) radius of the event horizon which can be as little as half the Schwarzschild radius for a maximally rotating black hole. Rotation As a black hole rotates, it twists spacetime in the direction of the rotation at a speed that decreases with distance from the event horizon. This process is known as the Lense–Thirring effect or frame-dragging. Because of this dragging effect, an object within the ergosphere cannot appear stationary with respect to an outside observer at a great distance unless that object were to move at faster than the speed of light (an impossibility) with respect to the local spacetime. The speed necessary for such an object to appear stationary decreases at points further out from the event horizon, until at some distance the required speed is negligible. The set of all such points defines the ergosphere surface, called ergosurface. The outer surface of the ergosphere is called the static surface or static limit. This is because world lines change from being time-like outside the static limit to being space-like inside it. It is the speed of light that arbitrarily defines the ergosphere surface. Such a surface would appear as an o
https://en.wikipedia.org/wiki/Operation%20Jackpot%20Nalli%20C.I.D%20999
Operation Jackpot Nalli C.I.D 999 is a 1969 Indian Kannada-language spy thriller film directed and produced by Dorai–Bhagavan. It stars Rajkumar, Narasimharaju and Rekha. The film is the third in the CID 999 franchise, created along the lines of the James Bond and James Bond-styled films, with the first two being Jedara Bale & Goa Dalli CID 999. The success of this movie led to one more sequel - Operation Diamond Racket. This was the debut movie of actress Rekha in a lead role. With the release of this movie, CID 999 became the first character based trilogy movie in India. This movie was praised for its angled lighting, clever use of shadows and well thought out decor such as the atmospheric arches in the lair - all of which added an element of noir to the movie. Premise CID 999 aka Prakash is assigned to investigate about a nuclear scientist named Shekar, who created a formula named Plasma Binson, which is used to destroy any object in earth, and destroy the formula from getting into wrong hands. How does Prakash destroy the formula forms the crux of the plot. Cast Rajkumar as CID 999 Prakash Rekha as Mona Narasimharaju as CID 888 Baby Surekha Soundtrack
https://en.wikipedia.org/wiki/Percy%20Williams%20Bridgman
Percy Williams Bridgman (April 21, 1882 – August 20, 1961) was an American physicist who received the 1946 Nobel Prize in Physics for his work on the physics of high pressures. He also wrote extensively on the scientific method and on other aspects of the philosophy of science. The Bridgman effect, the Bridgman–Stockbarger technique, and the high-pressure mineral bridgmanite are named after him. Biography Early life Bridgman was born in Cambridge, Massachusetts, and grew up in nearby Auburndale. Bridgman's parents were both born in New England. His father, Raymond Landon Bridgman, was "profoundly religious and idealistic" and worked as a newspaper reporter assigned to state politics. His mother, Mary Ann Maria Williams, was described as "more conventional, sprightly, and competitive". Bridgman attended both elementary and high school in Auburndale, where he excelled at competitions in the classroom, on the playground, and while playing chess. Described as both shy and proud, his home life consisted of family music, card games, and domestic and garden chores. The family was deeply religious; reading the Bible each morning and attending a Congregational Church. However, Bridgman later became an atheist. Education and professional life Bridgman entered Harvard University in 1900, and studied physics through to his PhD. From 1910 until his retirement, he taught at Harvard, becoming a full professor in 1919. In 1905, he began investigating the properties of matter under high pressure. A machinery malfunction led him to modify his pressure apparatus; the result was a new device enabling him to create pressures eventually exceeding 100,000 kgf/cm2 (10 GPa; 100,000 atmospheres). This was a huge improvement over previous machinery, which could achieve pressures of only 3,000 kgf/cm2 (0.3 GPa). This new apparatus led to an abundance of new findings, including a study of the compressibility, electric and thermal conductivity, tensile strength and viscosity of more tha
https://en.wikipedia.org/wiki/Nitrofurazone
Nitrofurazone (INN, trade name Furacin) is an antimicrobial organic compound belonging to the nitrofuran class. It is most commonly used as a topical antibiotic ointment. It is effective against gram-positive bacteria, gram-negative bacteria, and can be used in the treatment of trypanosomiasis. Its use in medicine has become less frequent, as safer and more effective products have become available. Nitrofurazone is listed under California Prop 65, and has demonstrated clear evidence to be mutagenic and carcinogenic during animal studies, and has been discontinued for human use in the USA. The substance is pale yellow and crystalline. It was once widely used as an antibiotic for livestock. Medical uses Human use Nitrofurazone was previously available as a prescription in the U.S., and was indicated as a topical solution, topical cream, or topical ointment for the treatment of bacterial skin infections, wounds, burns, and ulcers. It was also used as a prophylactic measure to prevent infection that could potentially result in skin graft rejection. Nitrofurazone is still very popular as a topical solution for the treatment of tonsillitis in Russia. Animal use Nitrofurazone is indicated for topical use in dogs, cats, and horses, for the treatment or prophylactic treatment of superficial bacterial infections, burns, and cutaneous ulcers. Preparations for treating infections, such as fin rot, in ornamental fish are also still commercially available. The use of nitrofurazone, or related compounds, in animals raised for human consumption has been strictly banned. Pharmacokinetics The mechanism of action is not fully understood, but nitrofurazone's antimicrobial properties are suspected to be due to the interference of DNA synthesis in the microorganism by inhibiting certain enzymes that are involved with glycolysis. Other enzymes this may affect include, pyruvate dehydrogenase, citrate synthetase, malate dehydrogenase, glutathione reductase, and pyruvate deca
https://en.wikipedia.org/wiki/Voja%20Antoni%C4%87
Vojislav "Voja" Antonić (, ʾ, 12 July 1952) is a Serbian inventor, journalist, and writer. He is known for creating a build-it-yourself home computer Galaksija and originating a related "Build your own computer Galaksija" initiative with Dejan Ristanović. This initiative encouraged and enlightened thousands of computer enthusiasts during the 1980s in the Socialist Federal Republic of Yugoslavia. Antonić has donated many of his personal creations to the public domain. He was also a magazine editor and contributed to a number of radio shows. Biography While in school, Voja Antonić found a passion for HAM radios. He obtained a licence and a callsign to broadcast his own waves. One day, the state police seized all CB Band units known to operate in the country, creating a new trend for HAM radio units which bored Voja Antonić who decided to move on towards new digital technologies. His first creation with a microprocessor was Conway's Game of Life machine which shows its state using 16x16 matrix of red LEDs. Without a computer, Voja Antonić wrote the code on paper and operated the input in the system byte by byte using rotary switches. LEDs being expensive back then, it took him months to buy and install the last LEDs. A replica of his machine reportedly worked flawlessly almost continuously for 40 years. When personal computers arrived on the market, they were not accessible in Yugoslavia. Voja Antonić asked a friend in the USA to disassemble a TRS-80 Model I and send it to him and received it labelled as "technical junk". He received it, reassembled it, and started his new computer passion. While studying at the Faculty of Dramatic Arts in the late 1970s, he started to build computer systems capable of rendering animations. Prior to the Winter of 1981/1982, the Skiing Federation of Serbia timed the competitors using regular stopwatches and hand signaling. The upcoming Balkan competition required this to be improved and more precise. In 1981, Antonić created a sm
https://en.wikipedia.org/wiki/Trusted%20execution%20environment
A trusted execution environment (TEE) is a secure area of a main processor. It helps code and data loaded inside it to be protected with respect to confidentiality and integrity. Data integrity prevents unauthorized entities from outside the TEE from altering data, while code integrity prevents code in the TEE from being replaced or modified by unauthorized entities, which may also be the computer owner itself as in certain DRM schemes described in SGX. This is done by implementing unique, immutable, and confidential architectural security such as Intel Software Guard Extensions (Intel SGX) which offers hardware-based memory encryption that isolates specific application code and data in memory. Intel SGX allows user-level code to allocate private regions of memory, called enclaves, which are designed to be protected from processes running at higher privilege levels. A TEE as an isolated execution environment provides security features such as isolated execution, integrity of applications executing with the TEE, along with confidentiality of their assets. In general terms, the TEE offers an execution space that provides a higher level of security for trusted applications running on the device than a rich operating system (OS) and more functionality than a 'secure element' (SE). History The Open Mobile Terminal Platform (OMTP) first defined TEE in their "Advanced Trusted Environment:OMTP TR1" standard, defining it as a "set of hardware and software components providing facilities necessary to support Applications" which had to meet the requirements of one of two defined security levels. The first security level, Profile 1, was targeted against only software attacks and while Profile 2, was targeted against both software and hardware attacks. Commercial TEE solutions based on ARM TrustZone technology, conforming to the TR1 standard, were later launched, such as Trusted Foundations developed by Trusted Logic. Work on the OMTP standards ended in mid 2010 when the grou
https://en.wikipedia.org/wiki/Lability
Lability refers to something that is constantly undergoing change or is likely to undergo change. Biochemistry In reference to biochemistry, this is an important concept as far as kinetics is concerned in metalloproteins. This can allow for the rapid synthesis and degradation of substrates in biological systems. Biology Cells Labile cells refer to cells that constantly divide by entering and remaining in the cell cycle. These are contrasted with "stable cells" and "permanent cells". An important example of this is in the epithelium of the cornea, where cells divide at the basal level and move upwards, and the topmost cells die and fall off. Proteins In medicine, the term "labile" means susceptible to alteration or destruction. For example, a heat-labile protein is one that can be changed or destroyed at high temperatures. The opposite of labile in this context is "stable". Soils Compounds or materials that are easily transformed (often by biological activity) are termed labile. For example, labile phosphate is that fraction of soil phosphate that is readily transformed into soluble or plant-available phosphate. Labile organic matter is the soil organic matter that is easily decomposed by microorganisms. Chemistry The term is used to describe a transient chemical species. As a general example, if a molecule exists in a particular conformation for a short lifetime, before adopting a lower energy conformation (structural arrangement), the former molecular structure is said to have 'high lability' (such as C25, a 25-carbon fullerene spheroid). The term is sometimes also used in reference to reactivity – for example, a complex that quickly reaches equilibrium in solution is said to be labile (with respect to that solution). Another common example is the cis effect in organometallic chemistry, which is the labilization of CO ligands in the cis position of octahedral transition metal complexes. See also Emotional lability
https://en.wikipedia.org/wiki/High%20potential%20iron%E2%80%93sulfur%20protein
High potential iron-sulfur proteins (HIPIP) are a class of iron-sulfur proteins. They are ferredoxins that participate in electron transfer in photosynthetic bacteria as well as in Paracoccus denitrificans. Structure The HiPIPs are small proteins, typically containing 63 to 85 amino acid residues. The sequences show significant variation. As shown in the following schematic representation the iron-sulfur cluster is bound by four conserved cysteine residues. [ 4Fe-4S cluster] | | | | xxxxxxxxxxxxxxxxxxxCxCxxxxxxxCxxxxxCxxxx C: conserved cysteine residue involved in the binding of the 4Fe-4S core. [Fe4S4] clusters The [Fe4S4] clusters are abundant cofactors of metalloproteins. They participate in electron-transfer sequences. The core structure for the [Fe4S4] cluster is a cube with alternating Fe and S vertices. These clusters exist in two oxidation states with a small structural change. Two families of [Fe4S4] clusters are known: the ferredoxin (Fd) family and the high-potential iron–suflur protein (HiPIP) family. Both HiPIP and Fd share the same resting state: [Fe4S4]2+, which have the same geometric and spectroscopic features. Differences arise when it comes to their active state: HiPIP forms by oxidation to [Fe4S4]3+, and Fd is formed by reduction to [Fe4S4]+. \underset{(for\ HiPIP)}{[Fe4S4]^3+} <=>[\ce{oxidation}] \underset{(resting\ state)}{[Fe4S4]^2+} <=>[\ce{reduction}] \underset{(for\ Fd)}{[Fe4S4]+} The different oxidation states are explained by the proteins that combined with the [Fe4S4] cluster. Analysis from crystallographic data suggests that HiPIP is capable of preserving its higher oxidation state by forming fewer hydrogen bonds with water. The characteristic fold of the proteins wraps the [Fe4S4] cluster in a hydrophobic core, only being able to form about five conserved H-bond to the cluster ligands from the backbone. In contrast, the protein associated with the Fd's allow
https://en.wikipedia.org/wiki/Sanitary%20epidemiological%20reconnaissance
Sanitary epidemiological reconnaissance, synonym epidemiological reconnaissance is a literal name of a concept and routine of finding out disease potential on a territory of arrival of major contingent. . This is a kind of medical reconnaissance, process of information gathering on possible infectious diseases' origin-sources, the ways and factors of the infection transfer and determining all conditions that could have promoted the spread of infestation among army service personnel. In 1939 Academician E.N.Pavlovsky announced his "doctrine of nidality", so called by Soviet biologists. People can acquire zoonoses and insect-borne diseases when they occupy at certain times of the year natural habitat of a certain pathogen (e.g., plague, tularemia, leptospirosis, arboviruses, tick-borne relapsing fever). The WHO Expert Committee on Zoonoses listed over 100 such diseases. About natural focality of the diseases is known elsewhere. History Historically, Sanitary epidemiological reconnaissance implied collection and transfer of all data available on sanitary and epidemiological situation of the area of possible deployment and action of armed forces, the same data for the neighbouring and enemy armed forces. The aim for the reconnaissance was to clear up the reasons of the specific disease origin—sources of the infection in various extreme situations, including local wars and armed conflicts, the ways of the infection transfer and all factors promoting to the infestation. This practice has been successfully used on plague-endemic territory at the time of the Soviet–Japanese War (1945) in WWII : "Sanitary epidemiological reconnaissance was organized and conducted continuously for the entire depth of the operation. Mobile sanitary epidemiological detachments followed immediately behind the first echelon of tanks and mechanized vehicles of advancing Soviet army should they encounter any particular contagious disease. Withdrawing enemy forces had poisoned many wells and w
https://en.wikipedia.org/wiki/Killing%20of%20Cecil%20the%20lion
Cecil ( 2002 – 2 July 2015) was a male African lion (Panthera leo leo) who lived primarily in the Hwange National Park in Matabeleland North, Zimbabwe. He was being studied and tracked by a research team of the University of Oxford as part of a long-term study. On the night of July 1, 2015, Cecil was lured out of the protected area and wounded with an arrow by Walter Palmer, an American recreational big-game trophy hunter, then tracked and killed with a compound bow the following morning, between 10 and 12 hours later. Cecil was 13 years old when killed. Palmer had purchased a hunting permit and was not charged legally with any crime; authorities in Zimbabwe have said he is still free to visit the country as a tourist, but not as a hunter. Two Zimbabweans (the hunting guide and the owner of the farm where the hunt took place) were briefly arrested but the charges were eventually dismissed by courts. The killing resulted in international media attention, caused outrage among animal conservationists, criticism by politicians and celebrities and a strong negative response against Palmer. Five months after the killing of Cecil, the U.S. Fish and Wildlife Service added lions in India and West and Central Africa to the endangered species list, making it more difficult for United States citizens to kill lions legally on safaris. According to Wayne Pacelle, then President of the Humane Society, Cecil had "changed the atmospherics on the issue of trophy hunting around the world", adding, "I think it gave less wiggle room to regulators." Because of the high level of media attention and the negative reporting about the killing of Cecil, significantly fewer hunters came to Zimbabwe in the months that followed. This led to the country suffering financial losses and a lion overpopulation in the Bubye Valley Conservancy. Background Cecil was named after the British businessman, politician and mining magnate Cecil Rhodes, as was the namesake country of Rhodesia, now Zimbabw
https://en.wikipedia.org/wiki/Touchard%20polynomials
The Touchard polynomials, studied by , also called the exponential polynomials or Bell polynomials, comprise a polynomial sequence of binomial type defined by where is a Stirling number of the second kind, i.e., the number of partitions of a set of size n into k disjoint non-empty subsets. The first few Touchard polynomials are Properties Basic properties The value at 1 of the nth Touchard polynomial is the nth Bell number, i.e., the number of partitions of a set of size n: If X is a random variable with a Poisson distribution with expected value λ, then its nth moment is E(Xn) = Tn(λ), leading to the definition: Using this fact one can quickly prove that this polynomial sequence is of binomial type, i.e., it satisfies the sequence of identities: The Touchard polynomials constitute the only polynomial sequence of binomial type with the coefficient of x equal 1 in every polynomial. The Touchard polynomials satisfy the Rodrigues-like formula: The Touchard polynomials satisfy the recurrence relation and In the case x = 1, this reduces to the recurrence formula for the Bell numbers. A generalization of both this formula and the definition, is a generalization of Spivey's formula Using the umbral notation Tn(x)=Tn(x), these formulas become: The generating function of the Touchard polynomials is which corresponds to the generating function of Stirling numbers of the second kind. Touchard polynomials have contour integral representation: Zeroes All zeroes of the Touchard polynomials are real and negative. This fact was observed by L. H. Harper in 1967. The absolute value of the leftmost zero is bounded from above by although it is conjectured that the leftmost zero grows linearly with the index n. The Mahler measure of the Touchard polynomials can be estimated as follows: where and are the smallest of the maximum two k indices such that and are maximal, respectively. Generalizations Complete Bell polynomial may be viewed as a multivariate g
https://en.wikipedia.org/wiki/Chronic%20actinic%20dermatitis
Chronic actinic dermatitis is a condition where a subject's skin becomes inflamed due to a reaction to sunlight or artificial light. Patients often have other related conditions of the skin that cause dermatitis in response to a variety of stimuli (e.g., flowers, sunscreens, cosmetics, etc.). Symptoms Once affected, the symptoms may not show for several days. The symptoms can be severe burning, itching, swelling and pain in the affected areas. The areas exposed to sunlight may have the appearance of a sunburn - where clothing is worn the skin is protected. There is no known reaction to moonlight, but reflections from windows and mirrors of sunlight can still cause damage. Diagnosis Diagnosis can occur at any age, ranging from soon after birth to adulthood. A GP may refer a patient to a dermatologist if the condition is not showing clear symptoms, and a variety of tests - usually completed at a hospital - can then determine the exact nature and cause of the patient's condition. Reactions, which vary depending on the severity of the case, include rashes, flared 'bumpy' patches, affected areas being extremely hot to touch, and outbreaks shortly (or within 24 hours) after direct or indirect exposure to UVA and/or UVB light. The skin most likely reacts on the upper chest, hands and face, however it is not unlikely for reactions to happen all over the body. The patient may feel burning, stinging or throbbing sensations in these areas, which causes mild, yet uncomfortable pain in some patients. Others liken the pain and sensation to a chemical burn that doesn't go away. It is a mistake to think that the reaction is like a sunburn, it is far deeper in the skin and often requires the use of ingestible steroids as well as topical steroids in order to alleviate the condition to a degree. The best protection is to be fully covered from sunlight, even when cloudy or hazy. The use of UV-rated clothing is suggested as well as a UV-rated umbrella for outdoors. See also List o
https://en.wikipedia.org/wiki/Can%20f%201
Can f 1 or canis familiaris allergen 1 is a lipocalin allergen produced by dogs in their tongue epithelial tissue. It is homologous with the human lipocalin LCN1.
https://en.wikipedia.org/wiki/Plan%20%28archaeology%29
In archaeological excavation, a plan is a drawn record of features and artifacts in the horizontal plane. Overview Archaeological plan can either take the form of a "multi context" plan, which is drawn with many contexts on it to show relationships between these features as part of some phase, or alternatively a single context plan with a single feature is drawn . Excavated features are drawn in three dimensions with the help of drawing conventions such as hachures. Single context planning developed by the Museum of London has become the professional norm. The basic advantage of single context planning is context plans draw on "transparent perma-trace paper" can be overlaid for re-interpretation at a later date. Multi-context Plans as opposed to single context plans can be made of complete sites, trenches or individual features. In the United Kingdom, the scale of the plans is usually 1:20. They are linked to the site recording system by the inclusion of known grid points and height readings, taken with a dumpy level or a total station (see surveying). Excavation of a site by the removal of human made deposits in the reverse order they were created is the preferred method of excavation and is referred to as stratigraphic area excavation "in plan" as opposed to excavation "in section". Plan and section drawings have an interpretive function as well as being part of the recording system, because the draughts-person makes conscious decisions about what should be included or emphasised. Archaeological plan topics The grid It is common and good practice on excavations to lay out a grid of 5m squares so as to facilitate planning. This grid is marked out on-site with grid pegs that form the baselines for tapes and other planning tools to aid the drawing of plans. It is also common practice that planning is done for each context on a separate piece of perma-trace that conforms to these 5m grid squares. This is part of the single context recording system (see
https://en.wikipedia.org/wiki/Heptyl%20acetate
Heptyl acetate (C9H18O2), also known as acetate C-7, is a colorless alcohol-soluble liquid that is the ester formed by the condensation of 1-heptanol and acetic acid. Heptyl acetate is used as a fruit essence flavoring in foods and as a scent in perfumes. It has a woody, fruity, rumlike odor and a spicy, floral taste with a soapy, fatty texture.
https://en.wikipedia.org/wiki/Cohn%27s%20irreducibility%20criterion
Arthur Cohn's irreducibility criterion is a sufficient condition for a polynomial to be irreducible in —that is, for it to be unfactorable into the product of lower-degree polynomials with integer coefficients. The criterion is often stated as follows: If a prime number is expressed in base 10 as (where ) then the polynomial is irreducible in . The theorem can be generalized to other bases as follows: Assume that is a natural number and is a polynomial such that . If is a prime number then is irreducible in . The base 10 version of the theorem is attributed to Cohn by Pólya and Szegő in one of their books while the generalization to any base b is due to Brillhart, Filaseta, and Odlyzko. In 2002, Ram Murty gave a simplified proof as well as some history of the theorem in a paper that is available online. A further generalization of the theorem allowing coefficients larger than digits was given by Filaseta and Gross. In particular, let be a polynomial with non-negative integer coefficients such that is prime. If all coefficients are 49598666989151226098104244512918, then is irreducible over . Moreover, they proved that this bound is also sharp. In other words, coefficients larger than 49598666989151226098104244512918 do not guarantee irreducibility. The method of Filaseta and Gross was also generalized to provide similar sharp bounds for some other bases by Cole, Dunn, and Filaseta. The converse of this criterion is that, if p is an irreducible polynomial with integer coefficients that have greatest common divisor 1, then there exists a base such that the coefficients of p form the representation of a prime number in that base; this is the Bunyakovsky conjecture and its truth or falsity remains an open question. Historical notes Polya and Szegő gave their own generalization but it has many side conditions (on the locations of the roots, for instance) so it lacks the elegance of Brillhart's, Filaseta's, and Odlyzko's generalization. It is clear from c
https://en.wikipedia.org/wiki/ERM%20protein%20family
The ERM protein family consists of three closely related proteins, ezrin, radixin and moesin. The three paralogs, ezrin, radixin and moesin, are present in vertebrates, whereas other species have only one ERM gene. Therefore, in vertebrates these paralogs likely arose by gene duplication. ERM proteins are highly conserved throughout evolution. More than 75% identity is observed in the N-terminal and the C-terminal of vertebrates (ezrin, radixin, moesin), Drosophila (dmoesin) and C. elegans (ERM-1) homologs. Structure ERM molecules contain the following three domains: N-terminal globular domain, also called FERM domain (Band 4.1, ezrin, radixin, moesin). The FERM domain allows ERM proteins to interact with integral proteins of the plasma membrane, or scaffolding proteins localized beneath the plasma membrane. The FERM domain is composed of three subdomains (F1, F2, F3) that are arranged as a cloverleaf. extended alpha-helical domain. charged C-terminal domain. This domain mediates the interaction with F-actin. Ezrin, radixin and moesin also contain a polyproline region between the central helical and C-terminal domains. Function ERM proteins crosslink actin filaments with plasma membranes. They co-localize with CD44 at actin filament-plasma membrane interaction sites, associating with CD44 via their N-terminal domains and with actin filaments via their C-terminal domains. The ERM protein moesin directly binds to microtubules via its N-terminal FERM domain in vitro and stabilizes microtubules at the cell cortex in vivo. This interaction is required for specific ERM-dependent functions in mitosis. Activation ERM proteins are highly regulated proteins. They exist in two forms: the FERM domain is able to interact with the F-actin binding site and this head-to-tail interaction maintains ERM proteins into a folded form; in this state, ERM proteins are inactive for the folding prevents either integral protein binding, or actin-binding. if this head-to-tail intera
https://en.wikipedia.org/wiki/Comparison%20of%20streaming%20media%20software
This is a comparison of streaming media systems. A more complete list of streaming media systems is also available. General The following tables compare general and technical information for a number of streaming media systems both audio and video. Please see the individual systems' linked articles for further information. Operating system support Container format support Information about what digital container formats are supported. Protocol support Information about which internet protocols are supported for broadcasting streaming media content. Features See also Community radio Comparison of video services Content delivery network Digital television Electronic commerce Internet radio Internet radio device Internet television IPTV List of Internet radio stations List of music streaming services Multicast P2PTV Protection of Broadcasts and Broadcasting Organizations Treaty Push technology Streaming media Ustream Webcast Web television
https://en.wikipedia.org/wiki/All%20fourths%20tuning
Among alternative tunings for the guitar, all-fourths tuning is a regular tuning. In contrast, the standard tuning has one irregularity—a major third between the third and second strings—while having perfect fourths between the other successive strings. The standard tuning's irregular major-third is replaced by a perfect fourth in all-fourths tuning, which has the open notes E2-A2-D3-G3-C4-F4. Among regular tunings, this all-fourths tuning best approximates the standard tuning. In all guitar tunings, the higher-octave version of a chord can be found by translating a chord by twelve frets higher along the fretboard. In every regular tuning, for example in all-fourths tuning, chords and intervals can be moved also diagonally. For all-fourths tuning, all twelve major chords (in the first or open positions) are generated by two chords, the open F major chord and the D major chord. The regularity of chord-patterns reduces the number of finger positions that need to be memorized. Jazz musician Stanley Jordan plays guitar in all-fourths tuning; he has stated that all-fourths tuning "simplifies the fingerboard, making it logical". Among all regular tunings, all-fourths tuning E-A-D-G-C-F is the best approximation of standard tuning, which is more popular. All-fourths tuning is traditionally used for the bass guitar; it is also used for the bajo sexto. Allan Holdsworth stated that if he were to learn the guitar again he would tune it in all-fourths. Relation with all-fifths tuning All-fourths tuning is closely related to all-fifths tuning. All-fourths tuning is based on the perfect fourth (five semitones), and all-fifths tuning is based on the perfect fifth (seven semitones). The perfect-fifth and perfect-fourth intervals are inversions of one another, and the chords of all-fourth and all-fifths are paired as inverted chords. Consequently, chord charts for all-fifths tunings may be used for left-handed all-fourths tuning. See also Scordatura, alternative tunings of
https://en.wikipedia.org/wiki/SUPER-UX
SUPER-UX was a version of the Unix operating system from NEC that is used on its SX series of supercomputers. History The initial version of SUPER-UX was based on UNIX System V version 3.1 with features from BSD 4.3. The version for the NEC SX-9 was based on SVR4.2MP with BSD enhancements. Features SUPER-UX is a 64-bit UNIX operating system. It supports the Supercomputer File System (SFS). Earth Simulator The Earth Simulator uses a custom OS called "ESOS" (Earth Simulator Operating System) based on SUPER-UX. It has many enhanced features custom designed for the Earth Simulator which are not in the regular SUPER-UX OS. See also EWS-UX
https://en.wikipedia.org/wiki/Clean%20feed%20%28television%29
In television technology, a clean feed is a video signal that does not have added graphics and text. This video signal is used in sport production to allow different television stations to add their own digital on-screen graphic image on a common signal, or in news broadcasting to produce two or more different streams, each one with the same picture but in different languages. A clean feed is a signal which has not come from the main output of the video switcher, such as the output of a vision mixer before the downstream keyer stage - the clean feed is identical to the main program output but without any captions keyed into it. Modern production equipment can actually put different keys on multiple outputs, allowing them to go to the clean feed or not. The most sophisticated vision mixers (or production switchers, according to the American nomenclature) can generate a clean feed output for any of their mix/effects (ME) buses. The term clean feed is also used to refer to backhaul feeds of television programming sent via communication satellite or other transport (such as a national fiber-optic network) sent from another TV station or remote television production truck on-location, which does not carry any television advertisements or break bumpers, or in some cases, lower-third graphics or superimposed chyron text. See also Mix-minus Television technology Television terminology
https://en.wikipedia.org/wiki/Personalized%20medicine
Personalized medicine, also referred to as precision medicine, is a medical model that separates people into different groups—with medical decisions, practices, interventions and/or products being tailored to the individual patient based on their predicted response or risk of disease. The terms personalized medicine, precision medicine, stratified medicine and P4 medicine are used interchangeably to describe this concept though some authors and organisations use these expressions separately to indicate particular nuances. While the tailoring of treatment to patients dates back at least to the time of Hippocrates, the term has risen in usage in recent years given the growth of new diagnostic and informatics approaches that provide understanding of the molecular basis of disease, particularly genomics. This provides a clear evidence base on which to stratify (group) related patients. Among the 14 Grand Challenges for Engineering, an initiative sponsored by National Academy of Engineering (NAE), personalized medicine has been identified as a key and prospective approach to "achieve optimal individual health decisions", therefore overcoming the challenge to "Engineer better medicines". Development of concept In personalised medicine, diagnostic testing is often employed for selecting appropriate and optimal therapies based on the context of a patient's genetic content or other molecular or cellular analysis. The use of genetic information has played a major role in certain aspects of personalized medicine (e.g. pharmacogenomics), and the term was first coined in the context of genetics, though it has since broadened to encompass all sorts of personalization measures, including the use of proteomics, imaging analysis, nanoparticle-based theranostics, among others. Relationship to personalized medicine Precision medicine (PM) is a medical model that proposes the customization of healthcare, with medical decisions, treatments, practices, or products being tailored to
https://en.wikipedia.org/wiki/Form%20%28botany%29
In botanical nomenclature, a form (forma, plural formae) is one of the "secondary" taxonomic ranks, below that of variety, which in turn is below that of species; it is an infraspecific taxon. If more than three ranks are listed in describing a taxon, the "classification" is being specified, but only three parts make up the "name" of the taxon: a genus name, a specific epithet, and an infraspecific epithet. The abbreviation "f." or the full "forma" should be put before the infraspecific epithet to indicate the rank. It is not italicised. For example: Acanthocalycium spiniflorum f. klimpelianum or Acanthocalycium spiniflorum forma klimpelianum (Weidlich & Werderm.) Donald Crataegus aestivalis (Walter) Torr. & A.Gray var. cerasoides Sarg. f. luculenta Sarg. is a classification of a plant whose name is: Crataegus aestivalis (Walter) Torr. & A.Gray f. luculenta Sarg. A form usually designates a group with a noticeable morphological deviation. The usual taxonomic practice is that the individuals classified within the form are not necessarily known to be closely related (they may not form a clade). For instance, white-flowered plants of species that usually have coloured flowers can be grouped and named (e.g., as "f. alba"). Formae apomicticae are sometimes named among plants that reproduce asexually, by apomixis. There are theoretically countless numbers of forms based on minor genetic differences, and only a few that have particular significance are likely to be named. See also Form (zoology) Forma specialis, an informal rank used for a parasitic form adapted to a particular host Trinomial nomenclature Variety (botany) Subvariety Plant variety (disambiguation) Cultivar Hybrid (biology) Race (biology)
https://en.wikipedia.org/wiki/ANNNI%20model
In statistical physics, the axial (or anisotropic) next-nearest neighbor Ising model, usually known as the ANNNI model, is a variant of the Ising model. In the ANNNI model, competing ferromagnetic and antiferromagnetic exchange interactions couple spins at nearest and next-nearest neighbor sites along one of the crystallographic axes of the lattice. The model is a prototype for complicated spatially modulated magnetic superstructures in crystals. To describe experimental results on magnetic orderings in erbium, the model was introduced in 1961 by Roger Elliott from the University of Oxford. The model has given its name in 1980 by Michael E. Fisher and Walter Selke, who analysed it first by Monte Carlo methods, and then by low temperature series expansions, showing the fascinating complexity of its phase diagram, including devil's staircases and a Lifshitz point. Indeed, it provides, for two- and three-dimensional systems, a theoretical basis for understanding numerous experimental observations on commensurate and incommensurate structures, as well as accompanying phase transitions, in various magnets, alloys, adsorbates, polytypes, multiferroics, and other solids. Further possible applications range from modeling of cerebral cortex to quantum information.
https://en.wikipedia.org/wiki/Marta%20Civil
Marta Civil is an American mathematics educator. Her research involves understanding the cultural background of minority schoolchildren, particularly Hispanic and Latina/o students in the Southwestern United States, and using that understanding to promote parent engagement and focus mathematics teaching on students' individual strengths. She is the Roy F. Graesser Endowed Professor at the University of Arizona, where she holds appointments in the department of mathematics, the department of mathematics education, and the department of teaching, learning, and sociocultural studies. Education and career Civil earned her Ph.D. at the University of Illinois at Urbana–Champaign in 1990. Her dissertation, Doing and Talking about Mathematics: A Study of Preservice Elementary Teachers, was supervised by Peter George Braunfeld. In 2011 she moved from the University of Arizona to the University of North Carolina, to become Frank A. Daniels Distinguished Professor of Mathematics Education, but returned to Arizona in 2014 to become the Graesser Professor. Books Civil is co-editor of the books Transnational and Borderland Studies in Mathematics Education (Routledge, 2011), Latinos/as and Mathematics Education: Research on Learning and Teaching in Classrooms and Communities (Information Age, 2011), Cases for Mathematics Teacher Educators: Facilitating Conversations about Inequities in Mathematics Classrooms (Information Age, 2016), and Access & Equity: Promoting High-Quality Mathematics in Grades 3-5 (National Council of Teachers of Mathematics, 2018). Recognition In 2013 TODOS: Mathematics for All gave Civil their Iris M. Carl Equity and Leadership Award. She is included in a deck of playing cards featuring notable women mathematicians published by the Association of Women in Mathematics. She received the 2021 National Council of Teachers of Mathematics (NCTM) Lifetime Achievement Award.
https://en.wikipedia.org/wiki/McVittie%20metric
In the general theory of relativity, the McVittie metric is the exact solution of Einstein's field equations describing a black hole or massive object immersed in an expanding cosmological spacetime. The solution was first fully obtained by George McVittie in the 1930s, while investigating the effect of the, then recently discovered, expansion of the Universe on a mass particle. The simplest case of a spherically symmetric solution to the field equations of General Relativity with a cosmological constant term, the Schwarzschild-De Sitter spacetime, arises as a specific case of the McVittie metric, with positive 3-space scalar curvature and constant Hubble parameter . Metric In isotropic coordinates, the McVittie metric is given by where is the usual line element for the euclidean sphere, M is identified as the mass of the massive object, is the usual scale factor found in the FLRW metric, which accounts for the expansion of the space-time; and is a curvature parameter related to the scalar curvature of the 3-space as which is related to the curvature of the 3-space exactly as in the FLRW spacetime. It is generally assumed that , otherwise the Universe is undergoing a contraction. One can define the time-dependent mass parameter , which accounts for the mass density inside the expanding, comoving radius at time , to write the metric in a more succinct way Causal structure and singularities From here on, it is useful to define . For McVittie metrics with the general expanding FLRW solutions properties and , the spacetime has the property of containing at least two singularities. One is a cosmological, null-like naked singularity at the smallest positive root of the equation . This is interpreted as the black hole event-horizon in the case where . For the case, there is an event horizon at , but no singularity, which is extinguished by the existence of an asymptotic Schwarzschild-De Sitter phase of the spacetime. The second singularity li
https://en.wikipedia.org/wiki/Grain
A grain is a small, hard, dry fruit (caryopsis) – with or without an attached hull layer – harvested for human or animal consumption. A grain crop is a grain-producing plant. The two main types of commercial grain crops are cereals and legumes. After being harvested, dry grains are more durable than other staple foods, such as starchy fruits (plantains, breadfruit, etc.) and tubers (sweet potatoes, cassava, and more). This durability has made grains well suited to industrial agriculture, since they can be mechanically harvested, transported by rail or ship, stored for long periods in silos, and milled for flour or pressed for oil. Thus, the grain market is a major global commodity market that includes crops such as maize, rice, soybeans, wheat and other grains. Grains and cereal Grains and cereal are synonymous with caryopses, the fruits of the grass family. In agronomy and commerce, seeds or fruits from other plant families are called grains if they resemble caryopses. For example, amaranth is sold as "grain amaranth", and amaranth products may be described as "whole grains". The pre-Hispanic civilizations of the Andes had grain-based food systems, but at higher elevations none of the grains was a cereal. All three grains native to the Andes (kaniwa, kiwicha, and quinoa) are broad-leafed plants rather than grasses such as corn, rice, and wheat. Classification Cereal grains A cereal is any grass cultivated for the edible components of its grain (botanically, a type of fruit called a caryopsis), composed of the endosperm, germ, and bran. The term may also refer to the resulting grain itself (specifically "cereal grain"). Cereal grain crops are grown in greater quantities and provide more food energy worldwide than any other type of crop and are therefore staple crops. Edible grains from other plant families, such as buckwheat, quinoa and chia, are referred to as pseudocereals. In their natural, unprocessed, whole grain form, cereals are a rich source of vitam
https://en.wikipedia.org/wiki/Center%20of%20curvature
In geometry, the center of curvature of a curve is found at a point that is at a distance from the curve equal to the radius of curvature lying on the normal vector. It is the point at infinity if the curvature is zero. The osculating circle to the curve is centered at the centre of curvature. Cauchy defined the center of curvature C as the intersection point of two infinitely close normal lines to the curve. The locus of centers of curvature for each point on the curve comprise the evolute of the curve. This term is generally used in physics regarding the study of lenses and mirrors (see radius of curvature (optics)). It can also be defined as the spherical distance between the point at which all the rays falling on a lens or mirror either seems to converge to (in the case of convex lenses and concave mirrors) or diverge from (in the case of concave lenses or convex mirrors) and the lens/mirror itself. See also Curvature Differential geometry of curves
https://en.wikipedia.org/wiki/Conic%20section
A conic section, conic or a quadratic curve is a curve obtained from a cone's surface intersecting a plane. The three types of conic section are the hyperbola, the parabola, and the ellipse; the circle is a special case of the ellipse, though it was sometimes called as a fourth type. The ancient Greek mathematicians studied conic sections, culminating around 200 BC with Apollonius of Perga's systematic work on their properties. The conic sections in the Euclidean plane have various distinguishing properties, many of which can be used as alternative definitions. One such property defines a non-circular conic to be the set of those points whose distances to some particular point, called a focus, and some particular line, called a directrix, are in a fixed ratio, called the eccentricity. The type of conic is determined by the value of the eccentricity. In analytic geometry, a conic may be defined as a plane algebraic curve of degree 2; that is, as the set of points whose coordinates satisfy a quadratic equation in two variables which can be written in the form The geometric properties of the conic can be deduced from its equation. In the Euclidean plane, the three types of conic sections appear quite different, but share many properties. By extending the Euclidean plane to include a line at infinity, obtaining a projective plane, the apparent difference vanishes: the branches of a hyperbola meet in two points at infinity, making it a single closed curve; and the two ends of a parabola meet to make it a closed curve tangent to the line at infinity. Further extension, by expanding the real coordinates to admit complex coordinates, provides the means to see this unification algebraically. Euclidean geometry The conic sections have been studied for thousands of years and have provided a rich source of interesting and beautiful results in Euclidean geometry. Definition A conic is the curve obtained as the intersection of a plane, called the cutting plane, with the su
https://en.wikipedia.org/wiki/Block%20cipher%20mode%20of%20operation
In cryptography, a block cipher mode of operation is an algorithm that uses a block cipher to provide information security such as confidentiality or authenticity. A block cipher by itself is only suitable for the secure cryptographic transformation (encryption or decryption) of one fixed-length group of bits called a block. A mode of operation describes how to repeatedly apply a cipher's single-block operation to securely transform amounts of data larger than a block. Most modes require a unique binary sequence, often called an initialization vector (IV), for each encryption operation. The IV has to be non-repeating and, for some modes, random as well. The initialization vector is used to ensure distinct ciphertexts are produced even when the same plaintext is encrypted multiple times independently with the same key. Block ciphers may be capable of operating on more than one block size, but during transformation the block size is always fixed. Block cipher modes operate on whole blocks and require that the last part of the data be padded to a full block if it is smaller than the current block size. There are, however, modes that do not require padding because they effectively use a block cipher as a stream cipher. Historically, encryption modes have been studied extensively in regard to their error propagation properties under various scenarios of data modification. Later development regarded integrity protection as an entirely separate cryptographic goal. Some modern modes of operation combine confidentiality and authenticity in an efficient way, and are known as authenticated encryption modes. History and standardization The earliest modes of operation, ECB, CBC, OFB, and CFB (see below for all), date back to 1981 and were specified in FIPS 81, DES Modes of Operation. In 2001, the US National Institute of Standards and Technology (NIST) revised its list of approved modes of operation by including AES as a block cipher and adding CTR mode in SP800-38A, Recommen
https://en.wikipedia.org/wiki/Viability%20assay
A viability assay is an assay that is created to determine the ability of organs, cells or tissues to maintain or recover a state of survival. Viability can be distinguished from the all-or-nothing states of life and death by the use of a quantifiable index that ranges between the integers of 0 and 1 or, if more easily understood, the range of 0% and 100%. Viability can be observed through the physical properties of cells, tissues, and organs. Some of these include mechanical activity, motility, such as with spermatozoa and granulocytes, the contraction of muscle tissue or cells, mitotic activity in cellular functions, and more. Viability assays provide a more precise basis for measurement of an organism's level of vitality. Viability assays can lead to more findings than the difference of living versus nonliving. These techniques can be used to assess the success of cell culture techniques, cryopreservation techniques, the toxicity of substances, or the effectiveness of substances in mitigating effects of toxic substances. Common methods Though simple visual techniques of observing viability can be useful, it can be difficult to thoroughly measure an organism's/part of an organism's viability merely using the observation of physical properties. However, there are a variety of common protocols utilized for further observation of viability using assays. Tetrazolium reduction: One useful way to locate and measure viability is to complete a Tetrazolium Reduction Assay. The tetrazolium aspect of this assay, which utilizes both positive and negative charges in its formula, promotes the distinction of cell viability in a specimen. Resazurin reduction: Resazurin Reduction Assays perform very closely to that of a tetrazolium assay, except they use the power of redox to fuel their ability to represent cell viability. Protease viability marker: One can look at protease function in specimens if they wish to target viability in cells; this practice in research is known as
https://en.wikipedia.org/wiki/Tremolo
In music, tremolo (), or tremolando (), is a trembling effect. There are two types of tremolo: either a rapid repetition of a note, or a variation in volume. Types of tremolo Rapid reiteration The first is a rapid reiteration: Of a single note, particularly used on bowed string instruments, by rapidly moving the bow back and forth; plucked strings such as on a harp, where it is called bisbigliando () or "whispering". Tremolo picking, on traditionally plucked string instruments including guitar and mandolin, is the rapid articulation of single notes or a group of notes with a plectrum (pick) or with fingers. Tremolo playing sustains notes that would otherwise rapidly decay (fade to silence). Between two notes or chords in alternation, an imitation (not to be confused with a trill) of the preceding that is more common on keyboard instruments. Mallet instruments such as the marimba are capable of either method. A roll on any percussion instrument, whether tuned or untuned. Amplitude variation A second type of tremolo is a variation in amplitude: As produced on organs by tremulants Using electronic effects in guitar amplifiers and effects pedals which rapidly turn the volume of a signal up and down, creating a "shuddering" effect An imitation of the same by strings in which pulsations are taken in the same bow direction A vocal technique involving a wide or slow vibrato, not to be confused with the trillo or "Monteverdi trill" Tremolo is sometimes used interchangeably with vibrato. However, a tremolo is a variation of volume (or amplitude); as contrasted with vibrato, which is a variation of pitch (or frequency). Some electric guitars (in particular the Fender Stratocaster) use a lever called a "tremolo arm" or "whammy bar" that allows a performer to lower or (usually, to some extent) raise the pitch of a note or chord, an effect properly termed vibrato or "pitch bend". This non-standard use of the term "tremolo" refers to pitch rather than amplitude. How
https://en.wikipedia.org/wiki/Vegetative%20phase%20change
Vegetative phase change is the juvenile-to-adult transition in plants. This transition is distinct from the reproductive transition and is most prolonged and pronounced in woody species. Manipulating phase change may be an important avenue for plant improvement. In the model plant Arabidopsis thaliana, vegetative phase change is relatively subtle: leaves become more curled, with an increased number of abaxial trichomes, and increased serration. Studies in A. thaliana and maize identified microRNA MIR156 genes as master regulators of phase change, through their regulation of SQUAMOSA-PROMOTER-BINDING-LIKE (SBP/SPL) transcription factors. This gene regulatory circuit appears to be conserved (with variations) in all land plants, including mosses. See also Plant morphology Heteroblasty
https://en.wikipedia.org/wiki/Lu%C3%ADs%20Cruls
Luíz Cruls or Luís Cruls or Louis Ferdinand Cruls (21 January 1848 – 21 June 1908) was a Belgian-Brazilian astronomer and geodesist. He was Director of the Brazilian National Observatory from 1881 to 1908, led the commission charged with the survey and selection of a future site for the capital of Brazil in the Central Plateau, and was co-discoverer of the Great Comet of 1882. Cruls was also an active proponent of efforts to accurately measure solar parallax and towards that end led a Brazilian team in their observations of 1882 Transit of Venus in Punta Arenas, Chile. Early life Cruls was born in 1848 in Diest, Belgium, the son of Philippe Augustin Guillaume Cruls (a civil engineer) and Anne Elizabeth Jordens. From 1863 to 1868, Cruls studied civil engineering at the University of Ghent. In 1869 he undertook training as a military engineer and officer, graduating as a 2nd Lieutenant. Cruls served in the Belgian army, attaining the rank of 1st Lieutenant, until 1873 or 1874 (sources disagree). Likely inspired by Brazilian friends at University (including Caetano de Almeida Furquim, a fellow engineer), Cruls resigned his commission and set out for Brazil on 5 September 1874. During the trans-Atlantic crossing on the steamer Orénoque, Cruls met and became friends with Joaquim Nabuco, a journalist and abolitionist, and also the son of Jose Thomas Nabuco, an influential Brazilian politician. Nabuco's connections were to provide Cruls with access to the highest levels of Brazilian society. Brazil, Belgium, and back again Within weeks after Cruls' arrival in Brazil, Joaquim Nabuco and his father arranged for him to be presented to Dom Pedro II, Emperor of Brazil, and more importantly, to meet Buarque de Macedo, the Director General of the Ministry of Public Works. This latter meeting led to Cruls being hired as an engineer by the Commission of the Empire General Charter (Comissão da Carta Geral do Império) in the Geodesy section. In January 1875 Cruls was forced to
https://en.wikipedia.org/wiki/Electrical%20termination
In electronics, electrical termination is the practice of ending a transmission line with a device that matches the characteristic impedance of the line. Termination prevents signals from reflecting off the end of the transmission line. Reflections at the ends of unterminated transmission lines cause distortion, which can produce ambiguous digital signal levels and misoperation of digital systems. Reflections in analog signal systems cause such effects as video ghosting, or power loss in radio transmitter transmission lines. Transmission lines Signal termination often requires the installation of a terminator at the beginning and end of a wire or cable to prevent an RF signal from being reflected back from each end, causing interference, or power loss. The terminator is usually placed at the end of a transmission line or daisy chain bus (such as in SCSI), and is designed to match the AC impedance of the cable and hence minimize signal reflections, and power losses. Less commonly, a terminator is also placed at the driving end of the wire or cable, if not already part of the signal-generating equipment. Radio frequency currents tend to reflect from discontinuities in the cable, such as connectors and joints, and travel back down the cable toward the source, causing interference as primary reflections. Secondary reflections can also occur at the cable starts, allowing interference to persist as repeated echoes of old data. These reflections also act as bottlenecks, preventing the signal power from reaching the destination. Transmission line cables require impedance matching to carry electromagnetic signals with minimal reflections and power losses. The distinguishing feature of most transmission line cables is that they have uniform cross-sectional dimensions along their length, giving them a uniform electrical characteristic impedance. Signal terminators are designed to specifically match the characteristic impedances at both cable ends. For many systems, the ter
https://en.wikipedia.org/wiki/Sacrococcygeal%20membrane
The sacrococcygeal membrane is a tough fibrous membrane about 10mm long which extends from the inferior tip of the sacrum to the body of the coccyx in humans. It covers the inferior limit of the epidural space and is analogous to the ligamentum flavum found at other levels in the spine. It can be found at the apex of an equilateral triangle whose base is formed by the dimples overlying the sacro-iliac joints. The cornua of the sacrum may be palpated with a finger; the sacrococcygeal membrane lies between and inferior to these. Thorax (human anatomy)
https://en.wikipedia.org/wiki/Sum-product%20number
A sum-product number in a given number base is a natural number that is equal to the product of the sum of its digits and the product of its digits. There are a finite number of sum-product numbers in any given base . In base 10, there are exactly four numbers : 0, 1, 135, and 144. Definition Let be a natural number. We define the sum-product function for base , , to be the following: where is the number of digits in the number in base , and is the value of each digit of the number. A natural number is a number if it is a fixed point for , which occurs if . The natural numbers 0 and 1 are trivial numbers for all , and all other numbers are nontrivial numbers. For example, the number 144 in base 10 is a sum-product number, because , , and . A natural number is a sociable sum-product number if it is a periodic point for , where for a positive integer , and forms a cycle of period . A number is a sociable number with , and an amicable number is a sociable number with All natural numbers are preperiodic points for , regardless of the base. This is because for any given digit count , the minimum possible value of is and the maximum possible value of is The maximum possible digit sum is therefore and the maximum possible digit product is Thus, the function value is This suggests that or dividing both sides by , Since this means that there will be a maximum value where because of the exponential nature of and the linearity of Beyond this value , always. Thus, there are a finite number of numbers, and any natural number is guaranteed to reach a periodic point or a fixed point less than making it a preperiodic point. The number of iterations needed for to reach a fixed point is the function's persistence of , and undefined if it never reaches a fixed point. Any integer shown to be a sum-product number in a given base must, by definition, also be a Harshad number in that base. Sum-product numbers and cycles of Fb for specif
https://en.wikipedia.org/wiki/Sparse%20identification%20of%20non-linear%20dynamics
Sparse identification of nonlinear dynamics (SINDy) is a data-driven algorithm for obtaining dynamical systems from data. Given a series of snapshots of a dynamical system and its corresponding time derivatives, SINDy performs a sparsity-promoting regression (such as LASSO) on a library of nonlinear candidate functions of the snapshots against the derivatives to find the governing equations. This procedure relies on the assumption that most physical systems only have a few dominant terms which dictate the dynamics, given an appropriately selected coordinate system and quality training data. It has been applied to identify the dynamics of fluids, based on proper orthogonal decomposition, as well as other complex dynamical systems, such as biological networks. Mathematical Overview First, consider a dynamical system of the form where is a state vector (snapshot) of the system at time and the function defines the equations of motion and constraints of the system. The time derivative may be either prescribed or numerically approximated from the snapshots. With and sampled at equidistant points in time (), these can be arranged into matrices of the form and similarly for . Next, a library of nonlinear candidate functions of the columns of is constructed, which may be constant, polynomial, or more exotic functions (like trigonometric and rational terms, and so on): The number of possible model structures from this library is combinatorically high. is then substituted by and a vector of coefficients determining the active terms in : Because only a few terms are expected to be active at each point in time, an assumption is made that admits a sparse representation in . This then becomes an optimization problem in finding a sparse which optimally embeds . In other words, a parsimonious model is obtained by performing least squares regression on the system with sparsity-promoting () regularization where is a regularization parameter. Finally, the sparse
https://en.wikipedia.org/wiki/Terminal%20node%20controller
A terminal node controller (TNC) is a device used by amateur radio operators to participate in AX.25 packet radio networks. It is similar in function to the Packet Assembler/Disassemblers used on X.25 networks, with the addition of a modem to convert baseband digital signals to audio tones. The first TNC, the VADCG board, was originally developed by Doug Lockhart, VE7APU, of Vancouver, British Columbia. Amateur Radio TNCs were first developed in 1978 in Canada by the Montreal Amateur Radio Club and the Vancouver Area Digital Communications group. These never gained much popularity because only a bare printed circuit board was made available and builders had to gather up a large number of components. In 1983, the Tucson Amateur Packet Radio (TAPR) association produced complete kits for their TNC-1 design. This was later available as the Heathkit HD-4040. A few years later, the improved TNC-2 became available, and it was licensed to commercial manufacturers such as MFJ. In 1986, the improved "TNC+" was designed to run programs and protocols developed for the original TNC board. TNC+ also included an assembler and a version of Forth (STOIC), which runs on the TNC+ itself, to support developing new programs and protocols. Description A typical model consists of a microprocessor, a modem, and software (in EPROM) that implements the AX.25 protocol and provides a command line interface to the user. (Commonly, this software provides other functionality as well, such as a basic bulletin board system to receive messages while the operator is away.) Because the TNC contains all the intelligence needed to communicate over an AX.25 network, no external computer is required. All of the network's resources can be accessed using a dumb terminal. The TNC connects to the terminal and a radio transceiver. Data from the terminal is formatted into AX.25 packets and modulated into audio signals (in traditional applications) for transmission by the radio. Received signals
https://en.wikipedia.org/wiki/Post-tech
Post-tech, (or post-technology, post-digital-technology) is type of technology that is more concerned about being human than about technology. It advocates a design that is not merely focused on efficiency and exploiting users by increasing their time spent with digital devices and technology itself but to support the user's focus and intent, well-being, and independence (from technology). With this interstitial spaces could also be created, similar to what Michel Foucault describes as Heterotopia (space). See also Human-centered computing (discipline) Human-computer interaction Attention economy
https://en.wikipedia.org/wiki/Borders%20of%20the%20oceans
The borders of the oceans are the limits of Earth's oceanic waters. The definition and number of oceans can vary depending on the adopted criteria. The principal divisions (in descending order of area) of the five oceans are the Pacific Ocean, Atlantic Ocean, Indian Ocean, Southern (Antarctic) Ocean, and Arctic Ocean. Smaller regions of the oceans are called seas, gulfs, bays, straits, and other terms. Geologically, an ocean is an area of oceanic crust covered by water. See also the list of seas article for the seas included in each ocean area. Overview Though generally described as several separate oceans, the world's oceanic waters constitute one global, interconnected body of salt water sometimes referred to as the World Ocean or Global Ocean. This concept of a continuous body of water with relatively free interchange among its parts is of fundamental importance to oceanography. The major oceanic divisions are defined in part by the continents, various archipelagos, and other criteria. The principal divisions (in descending order of area) are the: Pacific Ocean, Atlantic Ocean, Indian Ocean, Southern (Antarctic) Ocean, and Arctic Ocean. Smaller regions of the oceans are called seas, gulfs, bays, straits, and other terms. Geologically, an ocean is an area of oceanic crust covered by water. Oceanic crust is the thin layer of solidified volcanic basalt that covers the Earth's mantle. Continental crust is thicker but less dense. From this perspective, the Earth has three oceans: the World Ocean, the Caspian Sea, and the Black Sea. The latter two were formed by the collision of Cimmeria with Laurasia. The Mediterranean Sea is at times a discrete ocean because tectonic plate movement has repeatedly broken its connection to the World Ocean through the Strait of Gibraltar. The Black Sea is connected to the Mediterranean through the Bosporus, but the Bosporus is a natural canal cut through continental rock some 7,000 years ago, rather than a piece of oceanic sea floo
https://en.wikipedia.org/wiki/Gegenbauer%20polynomials
In mathematics, Gegenbauer polynomials or ultraspherical polynomials C(x) are orthogonal polynomials on the interval [−1,1] with respect to the weight function (1 − x2)α–1/2. They generalize Legendre polynomials and Chebyshev polynomials, and are special cases of Jacobi polynomials. They are named after Leopold Gegenbauer. Characterizations A variety of characterizations of the Gegenbauer polynomials are available. The polynomials can be defined in terms of their generating function : The polynomials satisfy the recurrence relation : Gegenbauer polynomials are particular solutions of the Gegenbauer differential equation : When α = 1/2, the equation reduces to the Legendre equation, and the Gegenbauer polynomials reduce to the Legendre polynomials. When α = 1, the equation reduces to the Chebyshev differential equation, and the Gegenbauer polynomials reduce to the Chebyshev polynomials of the second kind. They are given as Gaussian hypergeometric series in certain cases where the series is in fact finite: (Abramowitz & Stegun p. 561). Here (2α)n is the rising factorial. Explicitly, From this it is also easy to obtain the value at unit argument: They are special cases of the Jacobi polynomials : in which represents the rising factorial of . One therefore also has the Rodrigues formula Orthogonality and normalization For a fixed α > -1/2, the polynomials are orthogonal on [−1, 1] with respect to the weighting function (Abramowitz & Stegun p. 774) To wit, for n ≠ m, They are normalized by Applications The Gegenbauer polynomials appear naturally as extensions of Legendre polynomials in the context of potential theory and harmonic analysis. The Newtonian potential in Rn has the expansion, valid with α = (n − 2)/2, When n = 3, this gives the Legendre polynomial expansion of the gravitational potential. Similar expressions are available for the expansion of the Poisson kernel in a ball . It follows that the quantities are spherical harmonics,
https://en.wikipedia.org/wiki/OS-level%20virtualization
OS-level virtualization is an operating system (OS) virtualization paradigm in which the kernel allows the existence of multiple isolated user space instances, called containers (LXC, Solaris containers, AIX WPARs, HP-UX SRP Containers, Docker, Podman), zones (Solaris containers), virtual private servers (OpenVZ), partitions, virtual environments (VEs), virtual kernels (DragonFly BSD), or jails (FreeBSD jail or chroot jail). Such instances may look like real computers from the point of view of programs running in them. A computer program running on an ordinary operating system can see all resources (connected devices, files and folders, network shares, CPU power, quantifiable hardware capabilities) of that computer. However, programs running inside of a container can only see the container's contents and devices assigned to the container. On Unix-like operating systems, this feature can be seen as an advanced implementation of the standard chroot mechanism, which changes the apparent root folder for the current running process and its children. In addition to isolation mechanisms, the kernel often provides resource-management features to limit the impact of one container's activities on other containers. Linux containers are all based on the virtualization, isolation, and resource management mechanisms provided by the Linux kernel, notably Linux namespaces and cgroups. The term container, while most popularly referring to OS-level virtualization systems, is sometimes ambiguously used to refer to fuller virtual machine environments operating in varying degrees of concert with the host OS, e.g., Microsoft's Hyper-V containers. A more historic overview of virtualization in general since 1960 can be found in the Timeline of virtualization development. Operation On ordinary operating systems for personal computers, a computer program can see (even though it might not be able to access) all the system's resources. They include: Hardware capabilities that can be emplo
https://en.wikipedia.org/wiki/Ulam%20number
In mathematics, the Ulam numbers comprise an integer sequence devised by and named after Stanislaw Ulam, who introduced it in 1964. The standard Ulam sequence (the (1, 2)-Ulam sequence) starts with U1 = 1 and U2 = 2. Then for n > 2, Un is defined to be the smallest integer that is the sum of two distinct earlier terms in exactly one way and larger than all earlier terms. Examples As a consequence of the definition, 3 is an Ulam number (1 + 2); and 4 is an Ulam number (1 + 3). (Here 2 + 2 is not a second representation of 4, because the previous terms must be distinct.) The integer 5 is not an Ulam number, because 5 = 1 + 4 = 2 + 3. The first few terms are 1, 2, 3, 4, 6, 8, 11, 13, 16, 18, 26, 28, 36, 38, 47, 48, 53, 57, 62, 69, 72, 77, 82, 87, 97, 99, 102, 106, 114, 126, 131, 138, 145, 148, 155, 175, 177, 180, 182, 189, 197, 206, 209, 219, 221, 236, 238, 241, 243, 253, 258, 260, 273, 282, ... . There are infinitely many Ulam numbers. For, after the first n numbers in the sequence have already been determined, it is always possible to extend the sequence by one more element: is uniquely represented as a sum of two of the first n numbers, and there may be other smaller numbers that are also uniquely represented in this way, so the next element can be chosen as the smallest of these uniquely representable numbers. Ulam is said to have conjectured that the numbers have zero density, but they seem to have a density of approximately 0.07398. Properties Apart from 1 + 2 = 3 any subsequent Ulam number cannot be the sum of its two prior consecutive Ulam numbers. Proof: Assume that for n > 2, Un−1 + Un = Un+1 is the required sum in only one way; then so does Un−2 + Un produce a sum in only one way, and it falls between Un and Un+1. This contradicts the condition that Un+1 is the next smallest Ulam number. For n > 2, any three consecutive Ulam numbers (Un−1, Un, Un+1) as integer sides will form a triangle. Proof: The previous property states that for n > 2, Un−2 +
https://en.wikipedia.org/wiki/Infrared%20window
The infrared atmospheric window refers to a region of the Infrared spectrum where there is relatively little absorption of terrestrial thermal radiation by atmospheric gases. The window plays an important role in the atmospheric greenhouse effect by maintaining the balance between incoming solar radiation and outgoing IR to space. In the Earth's atmosphere this window is roughly the region between 8 and 14 μm although it can be narrowed or closed at times and places of high humidity because of the strong absorption in the water vapor continuum or because of blocking by clouds. It covers a substantial part of the spectrum from surface thermal emission which starts at roughly 5 μm. Principally it is a large gap in the absorption spectrum of water vapor. Carbon dioxide plays an important role in setting the boundary at the long wavelength end. Ozone partly blocks transmission in the middle of the window. The importance of the infrared atmospheric window in the atmospheric energy balance was discovered by George Simpson in 1928, based on G. Hettner's 1918 laboratory studies of the gap in the absorption spectrum of water vapor. In those days, computers were not available, and Simpson notes that he used approximations; he writes about the need for this in order to calculate outgoing IR radiation: "There is no hope of getting an exact solution; but by making suitable simplifying assumptions . . . ." Nowadays, accurate line-by-line computations are possible, and careful studies of the spectroscopy of infrared atmospheric gases have been published. Mechanisms in the infrared atmospheric window The principal natural greenhouse gases in order of their importance are water vapor , carbon dioxide , ozone , methane and nitrous oxide . The concentration of the least common of these, , is about 400 ppbV. Other gases which contribute to the greenhouse effect are present at pptV levels. These include the chlorofluorocarbons (CFCs), halons and hydrofluorocarbons (HFC and HCFCs). A
https://en.wikipedia.org/wiki/Eigenface
An eigenface () is the name given to a set of eigenvectors when used in the computer vision problem of human face recognition. The approach of using eigenfaces for recognition was developed by Sirovich and Kirby and used by Matthew Turk and Alex Pentland in face classification. The eigenvectors are derived from the covariance matrix of the probability distribution over the high-dimensional vector space of face images. The eigenfaces themselves form a basis set of all images used to construct the covariance matrix. This produces dimension reduction by allowing the smaller set of basis images to represent the original training images. Classification can be achieved by comparing how faces are represented by the basis set. History The eigenface approach began with a search for a low-dimensional representation of face images. Sirovich and Kirby showed that principal component analysis could be used on a collection of face images to form a set of basis features. These basis images, known as eigenpictures, could be linearly combined to reconstruct images in the original training set. If the training set consists of M images, principal component analysis could form a basis set of N images, where N < M. The reconstruction error is reduced by increasing the number of eigenpictures; however, the number needed is always chosen less than M. For example, if you need to generate a number of N eigenfaces for a training set of M face images, you can say that each face image can be made up of "proportions" of all the K "features" or eigenfaces: Face image1 = (23% of E1) + (2% of E2) + (51% of E3) + ... + (1% En). In 1991 M. Turk and A. Pentland expanded these results and presented the eigenface method of face recognition. In addition to designing a system for automated face recognition using eigenfaces, they showed a way of calculating the eigenvectors of a covariance matrix such that computers of the time could perform eigen-decomposition on a large number of face images. Face
https://en.wikipedia.org/wiki/Biological%20systems%20engineering
Biological systems engineering or Biosystems engineering is a broad-based engineering discipline with particular emphasis on non-medical biology. It can be thought of as a subset of the broader notion of biological engineering or bio-technology though not in the respects that pertain to biomedical engineering as biosystems engineering tends to focus less on medical applications than on agriculture, ecosystems, and food science. The discipline focuses broadly on environmentally sound and sustainable engineering solutions to meet societies' ecologically related needs. Biosystems engineering integrates the expertise of fundamental engineering fields with expertise from non-engineering disciplines. Background and organization Many college and university biological engineering departments have a history of being grounded in agricultural engineering and have only in the past two decades or so changed their names to reflect the movement towards more diverse biological based engineering programs. This major is sometimes called agricultural and biological engineering, biological and environmental engineering, etc., in different universities, generally reflecting interests of local employment opportunities. Since biological engineering covers a wide spectrum, many departments now offer specialization options. Depending on the department and the specialization options offered within each program, curricula may overlap with other related fields. There are a number of different titles for BSE-related departments at various universities. The professional societies commonly associated with many Biological Engineering programs include the American Society of Agricultural and Biological Engineers (ASABE) and the Institute of Biological Engineering (IBE), which generally encompasses BSE. Some program also participate in the Biomedical Engineering Society (BMES) and the American Institute of Chemical Engineers (AIChE). A biological systems engineer has a background in what bot
https://en.wikipedia.org/wiki/Bahcall%E2%80%93Wolf%20cusp
Bahcall–Wolf cusp refers to a particular distribution of stars around a massive black hole at the center of a galaxy or globular cluster. If the nucleus containing the black hole is sufficiently old, exchange of orbital energy between stars drives their distribution toward a characteristic form, such that the density of stars, ρ, varies with distance from the black hole, r, as So far, no clear example of a Bahcall–Wolf cusp has been found in any galaxy or star cluster. This may be due in part to the difficulty of resolving such a feature. Distribution of stars around a supermassive black hole Supermassive black holes reside in galactic nuclei. The total mass of the stars in a nucleus is roughly equal to the mass of the supermassive black hole. In the case of the Milky Way, the mass of the supermassive black hole is about 4 million Solar masses, and the number of stars in the nucleus is about ten million. The stars move around the supermassive black hole in elliptical orbits, similar to the orbits that planets follow around the Sun. The orbital energy of a star is where m is the star's mass, v is the star's velocity, r is its distance from the supermassive black hole, and M is the supermassive black hole's mass. A star's energy remains nearly constant for many orbital periods. But after roughly one relaxation time, most of the stars in the nucleus will have exchanged energy with other stars, causing their orbits to change. Bahcall and Wolf showed that once this has taken place, the distribution of orbital energies has the form which corresponds to the density ρ=ρ0 r −7/4. The figure shows how the density of stars evolves toward the Bahcall–Wolf form. The fully formed cusp extends outward to a distance of roughly one-fifth the supermassive black hole's influence radius. It is believed that relaxation times in the nuclei of small, dense galaxies are short enough for Bahcall–Wolf cusps to form. The Galactic Center The influence radius of the supermassive black
https://en.wikipedia.org/wiki/Degeneracy%20%28mathematics%29
In mathematics, a degenerate case is a limiting case of a class of objects which appears to be qualitatively different from (and usually simpler than) the rest of the class, and the term degeneracy is the condition of being a degenerate case. The definitions of many classes of composite or structured objects often implicitly include inequalities. For example, the angles and the side lengths of a triangle are supposed to be positive. The limiting cases, where one or several of these inequalities become equalities, are degeneracies. In the case of triangles, one has a degenerate triangle if at least one side length or angle is zero. Equivalently, it becomes a "line segment". Often, the degenerate cases are the exceptional cases where changes to the usual dimension or the cardinality of the object (or of some part of it) occur. For example, a triangle is an object of dimension two, and a degenerate triangle is contained in a line, which makes its dimension one. This is similar to the case of a circle, whose dimension shrinks from two to zero as it degenerates into a point. As another example, the solution set of a system of equations that depends on parameters generally has a fixed cardinality and dimension, but cardinality and/or dimension may be different for some exceptional values, called degenerate cases. In such a degenerate case, the solution set is said to be degenerate. For some classes of composite objects, the degenerate cases depend on the properties that are specifically studied. In particular, the class of objects may often be defined or characterized by systems of equations. In most scenarios, a given class of objects may be defined by several different systems of equations, and these different systems of equations may lead to different degenerate cases, while characterizing the same non-degenerate cases. This may be the reason for which there is no general definition of degeneracy, despite the fact that the concept is widely used and defined (if need
https://en.wikipedia.org/wiki/Castelnuovo%E2%80%93Mumford%20regularity
In algebraic geometry, the Castelnuovo–Mumford regularity of a coherent sheaf F over projective space is the smallest integer r such that it is r-regular, meaning that whenever . The regularity of a subscheme is defined to be the regularity of its sheaf of ideals. The regularity controls when the Hilbert function of the sheaf becomes a polynomial; more precisely dim is a polynomial in m when m is at least the regularity. The concept of r-regularity was introduced by , who attributed the following results to : An r-regular sheaf is s-regular for any . If a coherent sheaf is r-regular then is generated by its global sections. Graded modules A related idea exists in commutative algebra. Suppose is a polynomial ring over a field k and M is a finitely generated graded R-module. Suppose M has a minimal graded free resolution and let be the maximum of the degrees of the generators of . If r is an integer such that for all j, then M is said to be r-regular. The regularity of M is the smallest such r. These two notions of regularity coincide when F is a coherent sheaf such that contains no closed points. Then the graded module is finitely generated and has the same regularity as F. See also Hilbert scheme Quot scheme
https://en.wikipedia.org/wiki/Programming%20by%20example
In computer science, programming by example (PbE), also termed programming by demonstration or more generally as demonstrational programming, is an end-user development technique for teaching a computer new behavior by demonstrating actions on concrete examples. The system records user actions and infers a generalized program that can be used on new examples. PbE is intended to be easier to do than traditional computer programming, which generally requires learning and using a programming language. Many PbE systems have been developed as research prototypes, but few have found widespread real-world application. More recently, PbE has proved to be a useful paradigm for creating scientific work-flows. PbE is used in two independent clients for the BioMOBY protocol: Seahawk and Gbrowse moby. Also the programming by demonstration (PbD) term has been mostly adopted by robotics researchers for teaching new behaviors to the robot through a physical demonstration of the task. The usual distinction in literature between these terms is that in PbE the user gives a prototypical product of the computer execution, such as a row in the desired results of a query; while in PbD the user performs a sequence of actions that the computer must repeat, generalizing it to be used in different data sets. For final users, to automate a workflow in a complex tool (e.g. Photoshop), the most simple case of PbD is the macro recorder. See also Query by Example Automated machine learning Example-based machine translation Inductive programming Lapis (text editor), which allows simultaneous editing of similar items in multiple selections created by example Programming by demonstration Test-driven development
https://en.wikipedia.org/wiki/Tiversa
Tiversa is an American cybersecurity firm headquartered in Pittsburgh, Pennsylvania. It was founded by a retired chiropractor and real estate entrepreneur named Robert Boback in 2004. The company specialized in trawling the deep web, investigating peer-to-peer networks, and helping businesses counteract data breaches and other cybersecurity risks. Its main product was EagleVision X1, a piece of software that monitored the deep web -- the parts of the Internet that are not easily accessible to general browsers, such as peer-to-peer networks -- for sensitive data. History Before entering the cybersecurity field, Boback was a chiropractor and real estate entrepreneur. He started Tiversa in 2004 as a two-person shop. Tiversa quickly obtained a high-profile board of advisers, including Maynard Webb (former eBay executive and chairman of Yahoo), Howard Schmidt (Obama-era cybersecurity chief), and Wesley Clark (former Supreme Allied Commander of NATO). Marine One hack In 2009, Tiversa claimed to have discovered a major security breach involving then-President Barack Obama's helicopter, Marine One. The breach involved the leak to Iran of sensitive procurement information about the helicopter as well as the helicopter's blueprints. According to Tiversa's CEO, the breach was caused by a defense contractor employee whose daughter downloaded a peer-to-peer file-sharing client onto a disused laptop which contained the sensitive materials. This discovery made national news, but a whistleblower later claimed that the Iranian hack was actually fabricated by Tiversa employees. Boback, the CEO of Tiversa, denied the allegation. LabMD scandal In May 2008, a Tiversa executive contacted LabMD (a urology testing laboratory) claiming to have discovered evidence of a major data breach and offered to sell LabMD monitoring services to counteract the breach. When the head of LabMD declined to purchase the monitoring services, Tiversa allegedly leaked information about the breach to t
https://en.wikipedia.org/wiki/Retriangulation%20of%20Great%20Britain
The Retriangulation of Great Britain was a triangulation project carried out between 1935 and 1962 that sought to improve the accuracy of maps of Great Britain. Data gathered from the retriangulation replaced data gathered during the Principal Triangulation of Great Britain, which had been performed between 1783 and 1851. The work was designed to form a complete new survey control network for the whole country, and to unify the mapping of the United Kingdom from local county projections into a single national datum projection and reference system. Its completion led to the establishment of the OSGB36 datum and Ordnance Survey National Grid in use today. History and overview The retriangulation was begun in 1935 by the Director General of the Ordnance Survey, Major-General Malcolm MacLeod. It was directed by the cartographer and mathematician Martin Hotine, head of the Trigonometrical and Levelling Division (TLD). The work was halted by the outbreak of World War II in 1939, by which time the primary triangulation network covered all of England and Wales, but only as far as the Moray Firth in Scotland. Secondary triangulation had commenced in 1938, and after the end of the war, the retriangulation work was focused on secondary and lower-order survey work, to expedite the completion of new large-scale surveys. The wartime priorities of the TLD were focused on survey work in connection with the war effort, such as airfield and military construction, survey and computations for anti-aircraft and coastal battery positions, and survey of radiolocation sites. One-third of the Ordnance Survey staff were called up during the war, and the headquarters in Southampton was bombed and badly damaged. Staff were relocated to the Home Counties, where they produced 1:25,000 scale maps of France, Italy, Germany and most of the rest of Europe in preparation for invasion. Primary triangulation observations were not resumed until 1949, and completed in 1952. A problem during the P
https://en.wikipedia.org/wiki/Very%20short%20patch%20repair
Very short patch (VSP) repair is a DNA repair system that removes GT mismatches created by the deamination of 5-methylcytosine to thymine. This system exists because the glycosylases which normally target deaminated bases cannot target thymine (it being one of the regular four bases in DNA). The components of the system are MutS, which binds to the GT mismatch, the VSR endonuclease, which cuts the DNA, and MutL, which recruits the UvrD helicase. VSR (very short patch repair) endonucleases occur in a variety of bacteria. They work by cutting, or rather, making a nick in DNA if the base pair is mutated or damaged. Function Mutations in the base pairs of DNA can be harmful to the organism. In particular, C to T mutations occur quite often due to methylation of cytosine. Hence, the VSR endonucleases have a function to protect the cell from damage caused by mutated DNA. Mechanism VSR recognises a TG mismatched base pair, generated after spontaneous deamination of methylated cytosines, and it creates a nick on a single strand by cleaving the phosphate backbone on the 5' side of the thymine. Then DNA Polymerase I removes the T and some nucleotides on the 3' strand and then resynthesises the patch. Additionally, GT mismatches can lead to C-to-T transition mutations if not repaired. VSR repairs the mismatches in favour of the G-containing strand. In Escherichia coli, this endonuclease nicks double-stranded DNA within the sequence CT(AT)GN or NT(AT)GG next to the thymidine residue, which is mismatched to 2'-deoxyguanosine. The incision is mismatch-dependent and strand specific. Structure The structure of VSR is similar to the core structure of restriction endonucleases, which have a 3-layer alpha/beta/alpha topology. VSR has three aromatic residues (Phe67, Trp68 and Trp86), which intercalate into the major groove, bending the DNA and separating the two strands. The N-terminal domain stabilizes the interaction between the protein and the cleaved product, thereby prote
https://en.wikipedia.org/wiki/Treen%20Cliff
Treen Cliff is a Site of Special Scientific Interest (SSSI) located on the Penwith Peninsula in Cornwall, England, UK, south-west of Penzance. First notified in 1951, with a revision in 1973, and a further notification on 1 July 1986, it is in area, stretching from grid reference SW387220 to SW402225. Designated for both for its biological and geological interest, part of the site, Treryn Dinas, is a Scheduled Ancient Monument consisting of a "cliff castle" with four ramparts and ditches and the Logan Rock. It is within the Cornwall Area of Outstanding Natural Beauty (AONB), the Penwith Heritage Coast and is part owned and managed by the National Trust. Description The SSSI extends from Porthcurno beach in the west to Penberth Cove in the east. Several rare plant species occur and the site is of particular importance for its maritime heath. At the last site assessment on 12 August 2010, the SSSI was described as ″Unfavourable recovering″ because scrub and bracken (Pteridium aquilinum) encroachment are diminishing the interest, and the site needs to be grazed. Removal of sycamore (Acer pseudoplatanus) would enhance the site as well as the control of other invasive species. Maritime cliff habitat types, National Vegetation Classification, (NVC) MC1 and MC5 need to be added to the notifiable features. Biological interest Plant communities The main habitats are maritime heath and maritime grassland, heath, scrub, flush and deciduous woodland. The maritime heath is dominated by heather (Calluna vulgaris), bell heather (Erica cinerea) and western gorse (Ulex gallii). Maritime grassland occurs on the steeper cliff slopes and is dominated by red fescue (Festuca rubra), with spring squill (Scilla verna), wild carrot (Daucus carota), kidney vetch (Anthyllis vulneraria) and sea plantain (Plantago maritima). The Red Data Book western clover (Trifolium occidentale) and the nationally rare hare's foot clover (Trifolium arvense) grow here as well as hairy bird's-foot tref
https://en.wikipedia.org/wiki/Mega-Merger
Mega-merger is a distributed algorithm aimed at solving the election problem in generic connected undirected graph. Introduction Mega-Merger was developed by Robert Gray Gallager at MIT in 1983. It applies a distributed divide and conquer approach mixed with a rank-based conquer strategy. The algorithm is usually presented through a village-city analogy. Each node in the graph indicates a village, while the edges that connect them are the roads and a rooted spanning tree in a sub-graph is a city. The whole graph is then a mega-city. Mega-Merger pushes villages to bind together to form cities according to each other's rank and edges. Cities are then formed by alliances or by conquering/absorption. Pre-requisites Mega-Merger builds a minimum spanning tree over connected graphs provided: Total reliability: No message is lost in transmission. UI (unique initiator): A single node starts the protocol. Bi-directional communications channels: Each edge is bi-directional, communications can travel in both directions. No further restrictions are necessary. Algorithm The algorithm assigns to each village a name and a rank, the former usually unique. The latter states the number of friendly mergers that the city has gone through, and the larger it is, the more powerful a city is considered. Moreover, to each edge is assigned a weight: each village/city has a minimum-weight edge also called merge link, that is the edge whose traversal has minimum cost. The algorithm proceeds in consecutive stages until a mega-city is formed. Each city C computes its own merge link and sends a request for merging across . The request is handled by in the following ways: Friendly merge: : If the cities share the same merge link and have same rank, a friendly merge occurs, and the two cities merge into one. A new name is picked for the newly created city, a ruling village is picked and the path from the previous ruler to the node in the merge link is re-oriented such that it leads to
https://en.wikipedia.org/wiki/Resultant
In mathematics, the resultant of two polynomials is a polynomial expression of their coefficients that is equal to zero if and only if the polynomials have a common root (possibly in a field extension), or, equivalently, a common factor (over their field of coefficients). In some older texts, the resultant is also called the eliminant. The resultant is widely used in number theory, either directly or through the discriminant, which is essentially the resultant of a polynomial and its derivative. The resultant of two polynomials with rational or polynomial coefficients may be computed efficiently on a computer. It is a basic tool of computer algebra, and is a built-in function of most computer algebra systems. It is used, among others, for cylindrical algebraic decomposition, integration of rational functions and drawing of curves defined by a bivariate polynomial equation. The resultant of n homogeneous polynomials in n variables (also called multivariate resultant, or Macaulay's resultant for distinguishing it from the usual resultant) is a generalization, introduced by Macaulay, of the usual resultant. It is, with Gröbner bases, one of the main tools of elimination theory. Notation The resultant of two univariate polynomials and is commonly denoted or In many applications of the resultant, the polynomials depend on several indeterminates and may be considered as univariate polynomials in one of their indeterminates, with polynomials in the other indeterminates as coefficients. In this case, the indeterminate that is selected for defining and computing the resultant is indicated as a subscript: or The degrees of the polynomials are used in the definition of the resultant. However, a polynomial of degree may also be considered as a polynomial of higher degree where the leading coefficients are zero. If such a higher degree is used for the resultant, it is usually indicated as a subscript or a superscript, such as or Definition The resultant of two u
https://en.wikipedia.org/wiki/Institute%20of%20Arctic%20Biology
The Institute of Arctic Biology or IAB of the University of Alaska Fairbanks, is located in Fairbanks, Alaska, US. The institute was established in 1963 by the Board of Regents of the University of Alaska, with Laurence Irving serving as its founding director. The mission of IAB is to advance basic and applied knowledge of high-latitude biological systems through research, education, and service. The Institute supports faculty, post-doctoral, and graduate research in wildlife biology and management, ecology, evolutionary biology, physiology, genetics, biomedicine, bioinformatics, and computational biology. IAB faculty hold joint appointments within other departments at UAF in the College of Natural Science and Mathematics and the School of Natural Resources and Agricultural Sciences. Important facilities and research programs that the Institute of Arctic Biology supports are: The Toolik Field Station; part of the LTER network, is a world-renowned Arctic climate change research station located in the northern foothills of the Brooks Range, Alaska, US. The Center for Alaska Native Health Research The Alaska Basic Neuroscience Program Center for Molecular and Genetic Studies of Hibernation The Alaska Geobotany Center The Resilience and Adaptation Program The Bonanza Creek Long-Term Ecological Research Site The Alaska Cooperative Fish and Wildlife Research Unit; administered by IAB, began in 1950 and is part of a nationwide cooperative program to promote research and graduate student training in the ecology and management of fish, wildlife, and their habitats. See also LTER NEON
https://en.wikipedia.org/wiki/Mathieu%20transformation
The Mathieu transformations make up a subgroup of canonical transformations preserving the differential form The transformation is named after the French mathematician Émile Léonard Mathieu. Details In order to have this invariance, there should exist at least one relation between and only (without any involved). where . When a Mathieu transformation becomes a Lagrange point transformation. See also Canonical transformation
https://en.wikipedia.org/wiki/Swim%20diaper
A swim diaper or swim nappy is a diaper that is made for those who have fecal incontinence (usually babies or toddlers), which is worn underneath a bathing suit, or as a bathing suit. Swim diapers can be reusable and disposable. They are not intended to be absorbent, but only to contain solid waste (feces); the lack of absorbency prevents the swim diaper from swelling with water. Types Often reusable swim diapers are lined with a fiber which encourages the solid waste to cling to the fiber without an absorbency layer. A snug fit in the legs and waist are key to function. Brands such as Splash About and The Honest Co. use tightly knit polyester or neoprene as their material. One disadvantage of a reusable swim diaper is that they must be washed to be reused. On the other hand, a disposable swim diaper is only partially biodegradable and repeated purchases may cost more than reuse. A popular brand of disposable swim diapers is Little Swimmers, marketed under the Kimberly-Clark Huggies brand. Procter & Gamble produces the rival brand Pampers Splashers. Both are sold in three sizes: small (16–26 lb or 7–12 kg), medium (24–34 lb or 11–15 kg) and large (over 32 lb or 14 kg+). Due to their design for swimwear, they are not as absorbent and not intended for regular diapering. Splash About offer the 'Happy Nappy' reusable neoprene swim nappy which can be washed and reworn. Swim diapers at public pools Some public pools require swim diapers for use by young children and the incontinent out of hygiene concerns. For the same reason, other pools do not allow swim diapers at all. Sick people who are incontinent, including babies and children, who do not wear swim diapers may transmit E. coli from their fecal matter. When not used properly, or when using inferior products, health experts caution that swim diapers may not protect pool water against communicable diseases, such as norovirus.
https://en.wikipedia.org/wiki/Henry%20Mann
Henry Berthold Mann (27 October 1905, Vienna – 1 February 2000, Tucson) was a professor of mathematics and statistics at the Ohio State University. Mann proved the Schnirelmann-Landau conjecture in number theory, and as a result earned the 1946 Cole Prize. He and his student developed the ("Mann-Whitney") U-statistic of nonparametric statistics. Mann published the first mathematical book on the design of experiments: . Early life of a number theorist Born in Vienna, Austria-Hungary, to a Jewish family, Mann earned his Ph.D. degree in mathematics in 1935 from the University of Vienna under the supervision of Philipp Furtwängler. Mann immigrated to the United States in 1938, and lived in New York, supporting himself by tutoring students. In additive number theory, Mann proved the Schnirelmann–Landau conjecture on the asymptotic density of sumsets in 1942. By doing so he established Mann's theorem and earned the 1946 Cole Prize. Statistics In 1942 the Carnegie Foundation awarded Mann a fellowship to learn statistics while assisting the operations research group of Harold Hotelling at Columbia University. His group also supported Abraham Wald, and Wald and Mann collaborated on several papers. In statistics, Mann is known for the ("Mann–Whitney") U-statistic and its associated hypothesis test for nonparametric statistics. Collaborating with Wald, Mann developed the Mann–Wald theorem of asymptotic statistics and econometrics. Mann wrote the first mathematical book on the design of experiments , whose principles allowed later statisticians to design and to analyze customized experiments. Like contemporary "self-help" and "how to" books, the earlier books gave easy-to-follow examples but little theory beyond exhortations to follow three principles of Ronald A. Fisher—to "replicate", to "establish control" (for example with blocking), and to "randomize" (assignment of treatments to units). Earlier books provided useful examples of designed experiments along with the des
https://en.wikipedia.org/wiki/Hsien%20Chung%20Wang
Hsien Chung (or Hsien-Chung or Hsien-chung) Wang ( 王宪钟 Wang Xian Zhong; 18 April 1918 in Beijing – 25 June 1978 in New York City) was a Chinese-American mathematician, specializing in differential geometry, Lie groups, and algebraic topology. Biography Part of a family, from Shandong Province, that had produced distinguished scholars for several generations, Hsien Chung Wang studied in Tianjin at Nankai High School, where he had an outstanding academic record. In 1936 he matriculated at Tsing Hua University in Beijing. After completing his Ph.D., Wang went to the United States. He was an Invited Speaker at the ICM in 1958 in Edinburgh. He was a Guggenheim Fellow for the academic year 1960–1961. The Wang sequence used in algebraic topology is named in his honor. He married in 1956 and was the father of three daughters. His doctoral students include J. Stephen Halperin. Selected publications with S. S. Chern: "Differential geometry in symplectic space." I, Sci. Rep. Nat. Tsing Hua Univ 4 (1947): 453–477. "Axiom of the plane in a general space of paths." Annals of Mathematics (1948): 731–737. "The homology groups of the fibre-bundles over a sphere." Duke Math. J 16 (1949): 33–38. "Homogeneous spaces with non-vanishing Euler characteristics." Annals of Mathematics (1949): 925–953. "A problem of PA Smith." Proceedings of the American Mathematical Society 1, no. 1 (1950): 18–19. "A remark on transformation groups leaving fixed an end point." Proceedings of the American Mathematical Society 3, no. 4 (1952): 548–549. "One-dimensional cohomology group of locally compact metrically homogeneous space." Duke Mathematical Journal 19, no. 2 (1952): 303–310. "Complex parallisable manifolds." Proceedings of the American Mathematical Society 5, no. 5 (1954): 771–776. with Kentaro Yano: "A class of affinely connected spaces." Transactions of the American Mathematical Society 80, no. 1 (1955): 72–92. "Discrete subgroups of solvable Lie groups I." Annals of Mathematics (19
https://en.wikipedia.org/wiki/Sinus%20%28botany%29
In botany, a sinus is a space or indentation between two lobes or teeth, usually on a leaf. The term is also used in mycology. For example, one of the defining characteristics of North American species in the Morchella elata clade of morels is the presence of a sinus where the cap attaches to the stipe. See also Leaf shape Sulcus (morphology)
https://en.wikipedia.org/wiki/Fold%20of%20left%20vena%20cava
The fold of the left vena cava, ligament of the left vena cava, or vestigial fold of Marshall, is a triangular fold of the serous pericardium that lies between the left pulmonary artery and subjacent pulmonary vein. It is formed by the folding of the serous layer over the remnant of the lower part of the left superior vena cava (duct of Cuvier), which becomes obliterated during fetal life, and remains as a fibrous band stretching from the highest left intercostal vein to the left atrium, where it is continuous with a small cardiac vein, the vein of the left atrium (oblique vein of Marshall), which opens into the coronary sinus.
https://en.wikipedia.org/wiki/Linked%20field
In mathematics, a linked field is a field for which the quadratic forms attached to quaternion algebras have a common property. Linked quaternion algebras Let F be a field of characteristic not equal to 2. Let A = (a1,a2) and B = (b1,b2) be quaternion algebras over F. The algebras A and B are linked quaternion algebras over F if there is x in F such that A is equivalent to (x,y) and B is equivalent to (x,z). The Albert form for A, B is It can be regarded as the difference in the Witt ring of the ternary forms attached to the imaginary subspaces of A and B. The quaternion algebras are linked if and only if the Albert form is isotropic. Linked fields The field F is linked if any two quaternion algebras over F are linked. Every global and local field is linked since all quadratic forms of degree 6 over such fields are isotropic. The following properties of F are equivalent: F is linked. Any two quaternion algebras over F are linked. Every Albert form (dimension six form of discriminant −1) is isotropic. The quaternion algebras form a subgroup of the Brauer group of F. Every dimension five form over F is a Pfister neighbour. No biquaternion algebra over F is a division algebra. A nonreal linked field has u-invariant equal to 1,2,4 or 8.
https://en.wikipedia.org/wiki/Mosher%27s%20acid
Mosher's acid, or α-methoxy-α-trifluoromethylphenylacetic acid (MTPA) is a carboxylic acid which was first used by Harry Stone Mosher as a chiral derivatizing agent. It is a chiral molecule, consisting of R and S enantiomers. Applications As a chiral derivatizing agent, it reacts with an alcohol or amine of unknown stereochemistry to form an ester or amide. The absolute configuration of the ester or amide is then determined by proton and/or 19F NMR spectroscopy. Mosher's acid chloride, the acid chloride form, is sometimes used because it has better reactivity. See also Pirkle's alcohol
https://en.wikipedia.org/wiki/Secondary%20market%20annuity
The term secondary market annuity is a misnomer. Certain salespeople use it as a term to describe an investment in factored structured settlement payment rights. The financial instrument that certain salespeople call a "secondary market annuity" comes about when a structured settlement factoring company buys certain structured settlement payment rights. The originating structured settlement factoring company may then further assign the rights (or subsets of thoe rights to investors for monetary consideration. The factoring company is not an insurance company and the investors not paying a premium to an insurance company. The investment that some salespeople call a secondary market annuity does not meet the definition of annuity under the insurance law for many states and does not enjoy the statutory protections. The National Association of Insurance Commissioners, an association of state regulators of insurance (including annuities) issued its Statutory Issue Paper No. 160 which makes clear that the acquisition of structured settlement payment rights is a not an annuity or an insurance product. The term "structured settlement payment rights" means rights to receive payments under a structured settlement. When the ownership of structured settlement payment rights is transferred by a Qualified Order, the ownership of the actual insurance product, the structured settlement annuity, does not change hands. The ownership of the annuity stays the same, as it was when the structured settlement was established. It is in most cases owned by a qualified assignment company, a subsidiary of the structured settlement annuity issuer. The structured settlement payment streams being marketed to investors as "secondary market annuities" do not have the same statutory protections as legitimate annuities. As of March 11, 2022, 34 states had adopted the 2017 Revisions to the Life and Health Guaranty Association Model Act #520) which expressly exclude factored structured settlement paym
https://en.wikipedia.org/wiki/Liver%20kidney%20microsomal%20type%201%20antibody
Liver kidney microsomal type 1 antibody (anti-LKM1) is an autoantibody associated with autoimmune hepatitis (AIH). Specifically, its presence in AIH defines type 2 AIH, although it has been proposed that anti-liver cytosol type 1 autoantibody without detectable anti-LKM1 can be seen in type 2 AIH. It is one of the several subtypes of anti–liver-kidney microsome antibodies that are known. The frequent association of anti-LKM-1 antibodies and hepatitis C virus (HCV) infections and the probable existence of an infection-associated autoimmune form of anti-LKM-1-associated hepatitis, requiring a different therapeutic strategy, necessitate the exact determination of anti-LKM-1 specificities. See also CYP2D6 – The target of anti-LKM-1
https://en.wikipedia.org/wiki/Scavenger
Scavengers are animals that consume dead organisms that have died from causes other than predation or have been killed by other predators. While scavenging generally refers to carnivores feeding on carrion, it is also a herbivorous feeding behavior. Scavengers play an important role in the ecosystem by consuming dead animal and plant material. Decomposers and detritivores complete this process, by consuming the remains left by scavengers. Scavengers aid in overcoming fluctuations of food resources in the environment. The process and rate of scavenging is affected by both biotic and abiotic factors, such as carcass size, habitat, temperature, and seasons. Etymology Scavenger is an alteration of scavager, from Middle English skawager meaning "customs collector", from skawage meaning "customs", from Old North French escauwage meaning "inspection", from schauwer meaning "to inspect", of Germanic origin; akin to Old English scēawian and German schauen meaning "to look at", and modern English "show" (with semantic drift). Types of scavengers (animals) Obligate scavenging (subsisting entirely or mainly on dead animals) is rare among vertebrates, due to the difficulty of finding enough carrion without expending too much energy. Well-known invertebrate scavengers of animal material include burying beetles and blowflies, which are obligate scavengers, and yellowjackets. Fly larvae are also common scavengers for organic materials at the bottom of freshwater bodies. For example, Tokunagayusurika akamusi is a species of midge fly whose larvae live as obligate scavengers at the bottom of lakes and whose adults almost never feed and only live up to a few weeks. Most scavenging animals are facultative scavengers that gain most of their food through other methods, especially predation. Many large carnivores that hunt regularly, such as hyenas and jackals, but also animals rarely thought of as scavengers, such as African lions, leopards, and wolves will scavenge if given the
https://en.wikipedia.org/wiki/BattleHack
BattleHack (or Battle Hack in the 2013 series) was a series of global hackathon contests organised by PayPal. Competitors were required to solve a local problem by coding. Winners of the first prize of each contest got an axe as the trophy, and admission to the world finals where competitors competed for the $100,000 grand prize. Competitors retain the ownership of their applications made in the contests. In 2016 PayPal/Braintree shut down the developer relations program along with all related programs such as BattleHack. It was announced that BattleHack would be returning however as of Jan 16, 2018 the domain name lapsed and has since moved into new ownership. See also Hackathon
https://en.wikipedia.org/wiki/Sleep%20hygiene
Sleep hygiene is a behavioral and environmental practice developed in the late 1970s as a method to help people with mild to moderate insomnia. Clinicians assess the sleep hygiene of people with insomnia and other conditions, such as depression, and offer recommendations based on the assessment. Sleep hygiene recommendations include: establishing a regular sleep schedule; using naps with care; not exercising physically or mentally too close to bedtime; limiting worry; limiting exposure to light in the hours before sleep; getting out of bed if sleep does not come; not using bed for anything but sleep and sex; avoiding alcohol as well as nicotine, caffeine, and other stimulants in the hours before bedtime; and having a peaceful, comfortable and dark sleep environment. However, , the empirical evidence for the effectiveness of sleep hygiene is "limited and inconclusive" for the general population and for the treatment of insomnia, despite being the oldest treatment for insomnia. A systematic review by the AASM concluded that clinicians should not prescribe sleep hygiene for insomnia due to the evidence of absence of its efficacy and potential delaying of adequate treatment, recommending instead that effective therapies such as CBT-i should be preferred. Assessment Assessing sleep hygiene is important to determine whether an individual has inadequate sleep hygiene disorder. The diagnostic assessment is usually conducted using clinical interview and supplemented by self-report questionnaires and sleep diaries, which are typically kept from one to two weeks, to record a representative sample data. There are also computerized assessments such as the Sleep-EVAL system, which can be employed in the diagnostic process. It features 1,543 possible questions automatically selected according to the individual's previous answers. Practice of sleep hygiene and knowledge of sleep hygiene practices can be assessed with measures such as the Sleep Hygiene Index, Sleep Hygiene Awarene
https://en.wikipedia.org/wiki/OSEK
OSEK (Offene Systeme und deren Schnittstellen für die Elektronik in Kraftfahrzeugen; English: "Open Systems and their Interfaces for the Electronics in Motor Vehicles") is a standards body that has produced specifications for an embedded operating system, a communications stack, and a network management protocol for automotive embedded systems. It has produced related specifications, namely AUTOSAR. OSEK was designed to provide a reliable standard software architecture for the various electronic control units (ECUs) throughout a car. OSEK was founded in 1993 by a German automotive company consortium (BMW, Robert Bosch GmbH, DaimlerChrysler, Opel, Siemens, and Volkswagen Group) and the University of Karlsruhe. In 1994, the French cars manufacturers Renault and PSA Peugeot Citroën, which had a similar project called VDX (Vehicle Distributed eXecutive), joined the consortium. Therefore, the official name was OSEK/VDX and OSEK was registered trademark of Continental Automotive GmbH (until 2007: Siemens AG). Standards OSEK is an open standard, published by a consortium founded by the automobile industry. Some parts of OSEK are standardized in ISO 17356. ISO 17356-1:2005 Road vehicles—Open interface for embedded automotive applications—Part 1: General structure and terms, definitions and abbreviated terms ISO 17356-2:2005 Road vehicles—Open interface for embedded automotive applications—Part 2: OSEK/VDX specifications for binding OS, COM and NM ISO 17356-3:2005 Road vehicles—Open interface for embedded automotive applications—Part 3: OSEK/VDX Operating System (OS) ISO 17356-4:2005 Road vehicles—Open interface for embedded automotive applications—Part 4: OSEK/VDX Communication (COM) ISO 17356-5:2006 Road vehicles—Open interface for embedded automotive applications—Part 5: OSEK/VDX Network Management (NM) ISO 17356-6:2006 Road vehicles—Open interface for embedded automotive applications—Part 6: OSEK/VDX Implementation Language (OIL) before ISO OSEK VDX Portal
https://en.wikipedia.org/wiki/Criticisms%20of%20econometrics
There have been many criticisms of econometrics' usefulness as a discipline and perceived widespread methodological shortcomings in econometric modelling practices. Difficulties in model specification Like other forms of statistical analysis, badly specified econometric models may show a spurious correlation where two variables are correlated but causally unrelated. Economist Ronald Coase is widely reported to have said "if you torture the data long enough it will confess". McCloskey argues that in published econometric work, economists often fail to use economic reasoning for including or excluding variables, equivocate statistical significance with substantial significance, and fail to report the power of their findings. Economic variables are not readily isolated for experimental testing, but Edward Leamer argues that there is no essential difference between econometric analysis and randomized trials or controlled trials provided the use of statistical techniques reduces the specification bias, the effects of collinearity between the variables, to the same order as the uncertainty due to the sample size. Economists are often faced with a high number of often highly collinear potential explanatory variables, leaving researcher bias to play an important role in their selection. Leamer argues that economists can mitigate this by running statistical tests with different specified models and discarding any inferences which prove to be "fragile", concluding that "professionals ... properly withhold belief until an inference can be shown to be adequately insensitive to the choice of assumptions". However, as Sala-I-Martin showed, it is often the case that you can specify two models suggesting contrary relation between two variables. The phenomenon was labeled 'emerging recalcitrant result' phenomenon by Robert Goldfarb. Lucas critique Robert Lucas criticised the use of overly simplistic econometric models of the macroeconomy to predict the implications of econo
https://en.wikipedia.org/wiki/CADES
CADES (Computer Aided Design and Evaluation System) was a software engineering system produced to support the design and development of the VME/B Operating System for the ICL New Range - subsequently 2900 - computers. From its earliest days, VME/B was developed with the aid of CADES, which was built for the purpose using an underlying IDMS database (latterly upgraded to IDMS(X)). CADES was not merely a version control system for code modules: it was intended to manage all aspects of the software lifecycle from requirements capture through to field maintenance. It was the design of CADES that paved the way for the Alvey Project in IPSE (Integrated Project Support Environments) and Process Control Engines. Because CADES was used for more than 20 years throughout the development of a large software engineering project, the data collected has been used as input to a number of studies of software evolution. Early history CADES was conceived in 1970 by David Pearson and Brian Warboys when working for ICL's New Range Operating System Technology Centre, OSTECH, in Kidsgrove. Pearson, a theoretical physicist by training, had become a computer simulation specialist and joined ICL in 1968 after working in finite-element modelling at Cambridge and simulation research at Imperial College. Warboys had been chief architect for the ICL System 4 multi-access operating system, Multijob. ICL's commitment to large scale software development for the 2900 Series of computers provided the basis for the Pearson and Warboys early work on a new software development environment which would address the issues of designer/programmer productivity, design integrity, evaluation and testing, version control and systems regression. In designing the initial architecture of the CADES environment, Pearson in particular looked to parallels with the leading hardware computer-aided design systems of the time, even attempting the use of graphics in the design process. The CADES design approach, calle
https://en.wikipedia.org/wiki/Protein%20structure%20prediction
Protein structure prediction is the inference of the three-dimensional structure of a protein from its amino acid sequence—that is, the prediction of its secondary and tertiary structure from primary structure. Structure prediction is different from the inverse problem of protein design. Protein structure prediction is one of the most important goals pursued by computational biology; and it is important in medicine (for example, in drug design) and biotechnology (for example, in the design of novel enzymes). Starting in 1994, the performance of current methods is assessed biannually in the CASP experiment (Critical Assessment of Techniques for Protein Structure Prediction). A continuous evaluation of protein structure prediction web servers is performed by the community project CAMEO3D. Protein structure and terminology Proteins are chains of amino acids joined together by peptide bonds. Many conformations of this chain are possible due to the rotation of the main chain about the two torsion angles φ and ψ at the Cα atom (see figure). This conformational flexibility is responsible for differences in the three-dimensional structure of proteins. The peptide bonds in the chain are polar, i.e. they have separated positive and negative charges (partial charges) in the carbonyl group, which can act as hydrogen bond acceptor and in the NH group, which can act as hydrogen bond donor. These groups can therefore interact in the protein structure. Proteins consist mostly of 20 different types of L-α-amino acids (the proteinogenic amino acids). These can be classified according to the chemistry of the side chain, which also plays an important structural role. Glycine takes on a special position, as it has the smallest side chain, only one hydrogen atom, and therefore can increase the local flexibility in the protein structure. Cysteine on the other hand can react with another cysteine residue to form one cystine and thereby form a cross link stabilizing the whole structure.
https://en.wikipedia.org/wiki/Power%20harassment
Power harassment is a form of harassment and workplace bullying in which someone in a position of greater power uses that power to harass or bully a lower-ranking person. It includes a range of behavior from mild irritation and annoyances to serious abuses which can even involve forced activity beyond the boundaries of the job description. Prohibited in some countries, power harassment is considered a form of illegal discrimination and political and psychological abuse. Types of power harassment include physical or psychological attacks, segregation, excessive or demeaning work assignments, and intrusion upon the victim's personal life. Power harassment may combine with other forms of bias and harassment, including sexual harassment. In the context of sexual harassment, power harassment is distinguished from contra power harassment, in which the harasser is of lower rank than that of the victim, and peer harassment, in which the victim and harasser are of the same rank. The term "political power harassment" was coined by Ramona Rush in a 1993 paper on sexual harassment in academia. Because it operates to reinforce and justify an existing hierarchy, political power harassment can be difficult to assess. By country Japan Although power harassment is not unique to Japan, it has received significant attention in Japan as a policy and legal problem since the 1990s. A government survey in 2016 found that more than 30% of workers had experienced power harassment in the preceding three years. The Japanese term "power harassment" () was independently coined by Yasuko Okada of Tokoha Gakuen Junior College in 2002. The Japanese courts have applied the general compensation principle of Article 709 of the Civil Code of Japan to compensate victims of workplace bullying and power harassment. In 2019, the National Diet adopted the Power Harassment Prevention Act, which amends the Labor Policy Comprehensive Promotion Act to require employers to address power harassment. The 2
https://en.wikipedia.org/wiki/Susanna%20S.%20Epp
Susanna Samuels Epp (born 1943) is an author, mathematician, and professor. Her interests include discrete mathematics, mathematical logic, cognitive psychology, and mathematics education, and she has written numerous articles, publications, and textbooks. She is currently professor emerita at DePaul University, where she chaired the Department of Mathematical Sciences and was Vincent de Paul Professor in Mathematics. Education and career Epp holds degrees in mathematics from Northwestern University and the University of Chicago, where she completed her doctorate in 1968 under the supervision of Irving Kaplansky. She taught at Boston University and at the University of Illinois at Chicago before becoming a professor at DePaul University. Contributions Initially researching commutative algebra, Epp became interested by cognitive psychology, especially in education of Mathematics, Logic, Proof, and the Language of mathematics. She wrote several articles about teaching logic and proof in American Mathematical Monthly, and the Mathematics Teacher, a Journal by the National Council of Teachers of Mathematics. She is the author of several books including Discrete Mathematics with Applications (4th ed., Brooks/Cole, 2011), the third edition of which earned a Textbook Excellence Award from the Textbook and Academic Authors Association. "By combining discussion of theory and practice, I have tried to show that mathematics has engaging and important applications as well as being interesting and beautiful in its own right" - Susanna S. Epp wrote in the Preface of the 4th Edition of Discrete Mathematics. Recognition In 2005, she received the Louise Hay Award from the Association for Women in Mathematics in recognition for her contributions to mathematics education. Selected publications Epp, S.S., Variables in Mathematics Education. In Tools for Teaching Logic. Blackburn, P., van Ditmarsch, H., et al., eds. Springer Publishing, 2011. (Reprinted in Best Writing on
https://en.wikipedia.org/wiki/Geniculate%20ganglion
The geniculate ganglion (from Latin genu, for "knee") is a collection of pseudounipolar sensory neurons of the facial nerve located in the facial canal of the head. It receives fibers from the facial nerve. It sends fibers that supply the lacrimal glands, submandibular glands, sublingual glands, tongue, palate, pharynx, external auditory meatus, stapedius muscle, posterior belly of the digastric muscle, stylohyoid muscle, and muscles of facial expression. The geniculate ganglion is one of several ganglia of the head and neck. Like the others, it is a bilaterally distributed structure, with each side of the face having a geniculate ganglion. Structure The geniculate ganglion is located close to the internal auditory meatus. It is covered superiorly by the petrous part of the temporal bone (which is sometimes absent over the ganglion). The geniculate ganglion receives fibers from the motor, sensory, and parasympathetic components of the facial nerve. It contains special sensory neuronal cell bodies for taste, from fibers coming up from the tongue through the chorda tympani and from fibers coming up from the roof of the palate through the greater petrosal nerve. Sensory and parasympathetic inputs are carried into the geniculate ganglion via the nervus intermedius. Motor fibers are carried via the facial nerve proper. The greater petrosal nerve, which carries preganglionic parasympathetic fibers, emerges from the anterior aspect of the ganglion. The motor fibers of the facial nerve proper and parasympathetic fibers to the submandibular and pterygopalatine ganglia do not synapse in the geniculate ganglion. The afferent fibers carrying pain, temperature, and touch from the posterior auricular nerve, as well as those carrying special sensory (taste) fibers from the tongue (via the chorda tympani), do not synapse in the geniculate ganglion. Instead, the cells of the geniculate ganglion relay the signal to the appropriate brainstem nucleus, much like the Dorsal root gan
https://en.wikipedia.org/wiki/Georges%20Matheron%20Lectureship
The Georges Matheron Lecture Series is sponsored by the International Association for Mathematical Geosciences (IAMG) to honor the legacy of the French engineer Georges François Paul Marie Matheron, known as the founder of geostatistics and a co-founder (together with Jean Serra) of mathematical morphology. The Georges Matheron Lecture is given by a scientist with proven research ability in the field of spatial statistics or mathematical morphology. It is presented annually if an eligible and worthy nominee is found. The first recipient of the award was Jean Serra, for a long time a scientists with the Centre of Mathematical Morphology, Fontainebleau. Serra delivered the first lecture at the IAMG conference in Liège, Belgium in 2006. The IAMG Lectures Committee seeks nominations and makes the selection. Awardees See also IAMG Distinguished Lectureship Bullerwell Lecture List of geology awards List of geophysics awards List of mathematics awards
https://en.wikipedia.org/wiki/Photoimageable%20thick-film%20technology
Photoimageable thick-film technology is a combination of conventional thick film technology with elements of thin film technology, and it provides a low cost solution to producing high quality microwave circuits. The ability to directly photoimage the printed layers means that the technology can provide the high line and gap resolution required by high frequency planar components. It provides a feasible fabrication process to produce circuits operating at microwave and millimetre-wave frequencies. Circuits made using this technology meet the modern requirements for high density packaging, whilst yielding the high quality components required for very high frequency applications, including wireless communication, radar and measurement systems. This technology also enables both single-layer and multi-layer filters to be produced conveniently. Recent research work has investigated the combination of conventional thick film and fine line photoimageable technologies, in order to accommodate fine pitch and high density applications on the ceramic substrates. Furthermore, previous work has shown that this technology is capable of realizing the circuit quality necessary for high performance microwave components. Benefits of choosing this structure Edge coupled band-pass filters were chosen for this study as they are one of the most common and useful microwave and millimetre-wave planar components. The filter performance is based on the coupling between the resonant sections and controlled by the size of the gap. This characteristic makes edge coupled band-pass filters very sensitive to fabrication errors. Another reason of choosing this structure in a multi-layer form is due to limitation on the structure when it has been fabricated on a single layer. The gap between the two resonant structures becomes very small and cannot easily be fabricated due to the limitations of low-cost fabrication technologies. In multi-layer circuits, the coupling between resonant sections is
https://en.wikipedia.org/wiki/Maxillary%20artery
The maxillary artery supplies deep structures of the face. It branches from the external carotid artery just deep to the neck of the mandible. Structure The maxillary artery, the larger of the two terminal branches of the external carotid artery, arises behind the neck of the mandible, and is at first imbedded in the substance of the parotid gland; it passes forward between the ramus of the mandible and the sphenomandibular ligament, and then runs, either superficial or deep to the lateral pterygoid muscle, to the pterygopalatine fossa. It supplies the deep structures of the face, and may be divided into mandibular, pterygoid, and pterygopalatine portions. First portion The first or mandibular or bony portion passes horizontally forward, between the neck of the mandible and the sphenomandibular ligament, where it lies parallel to and a little below the auriculotemporal nerve; it crosses the inferior alveolar nerve, and runs along the lower border of the lateral pterygoid muscle. Branches include: Deep auricular artery Anterior tympanic artery Middle meningeal artery Inferior alveolar artery which gives off its mylohyoid branch just prior to entering the mandibular foramen Accessory meningeal artery Second portion The second or pterygoid or muscular portion runs obliquely forward and upward under cover of the ramus of the mandible and insertion of the temporalis, on the superficial (very frequently on the deep) surface of the lateral pterygoid muscle; it then passes between the two heads of origin of this muscle and enters the fossa. Branches include: Masseteric artery Pterygoid branches Deep temporal arteries (anterior and posterior) Buccal artery Third portion The third or pterygopalatine or pterygomaxillary portion lies in the pterygopalatine fossa in relation with the pterygopalatine ganglion. This is considered the terminal branch of the maxillary artery. Branches include: Sphenopalatine artery (nasopalatine artery) is the terminal branch of
https://en.wikipedia.org/wiki/Line%20segment
In geometry, a line segment is a part of a straight line that is bounded by two distinct end points, and contains every point on the line that is between its endpoints. It is a special case of an arc, with zero curvature. The length of a line segment is given by the Euclidean distance between its endpoints. A closed line segment includes both endpoints, while an open line segment excludes both endpoints; a half-open line segment includes exactly one of the endpoints. In geometry, a line segment is often denoted using a line above the symbols for the two endpoints (such as ). Examples of line segments include the sides of a triangle or square. More generally, when both of the segment's end points are vertices of a polygon or polyhedron, the line segment is either an edge (of that polygon or polyhedron) if they are adjacent vertices, or a diagonal. When the end points both lie on a curve (such as a circle), a line segment is called a chord (of that curve). In real or complex vector spaces If is a vector space over or and is a subset of , then is a line segment if can be parameterized as for some vectors where is nonzero. The endpoints of are then the vectors and . Sometimes, one needs to distinguish between "open" and "closed" line segments. In this case, one would define a closed line segment as above, and an open line segment as a subset that can be parametrized as for some vectors Equivalently, a line segment is the convex hull of two points. Thus, the line segment can be expressed as a convex combination of the segment's two end points. In geometry, one might define point to be between two other points and , if the distance added to the distance is equal to the distance . Thus in the line segment with endpoints and is the following collection of points: Properties A line segment is a connected, non-empty set. If is a topological vector space, then a closed line segment is a closed set in . However, an open line segment is an open set i
https://en.wikipedia.org/wiki/Common%20Industrial%20Protocol
The Common Industrial Protocol (CIP) is an industrial protocol for industrial automation applications. It is supported by ODVA. Previously known as Control and Information Protocol, CIP encompasses a comprehensive suite of messages and services for the collection of manufacturing automation applications – control, safety, synchronization, motion, configuration and information. It allows users to integrate these manufacturing applications with enterprise-level Ethernet networks and the Internet. It is supported by hundreds of vendors around the world, and is media-independent. CIP provides a unified communication architecture throughout the manufacturing enterprise. It is used in EtherNet/IP, DeviceNet, CompoNet and ControlNet. ODVA is the organization that supports network technologies built on the Common Industrial Protocol (CIP). These also currently include application extensions to CIP: CIP Safety, CIP Motion and CIP Sync.
https://en.wikipedia.org/wiki/Cherenkov%20detector
A Cherenkov detector (pronunciation: /tʃɛrɛnˈkɔv/; Russian: Черенко́в) is a particle detector using the speed threshold for light production, the speed-dependent light output or the speed-dependent light direction of Cherenkov radiation. Fundamental A particle passing through a material at a velocity greater than that at which light can travel through the material emits light. This is similar to the production of a sonic boom when an airplane is traveling through the air faster than sound waves can move through the air. The direction this light is emitted is on a cone with angle θc about the direction in which the particle is moving, with cos(θc) =  (c = the vacuum speed of light, n = the refractive index of the medium, and v is the speed of the particle). The angle of the cone θc thus is a direct measure of the particle's speed. The Frank–Tamm formula  = sin2θc gives the number of photons produced. Aspects Most Cherenkov detectors aim at recording the Cherenkov light produced by a primary charged particle. Some sensor technologies explicitly aim at Cherenkov light produced (also) by secondary particles, be it incoherent emission as occurring in an electromagnetic particle shower or by coherent emission, for example Askaryan effect. Cherenkov radiation is not only present in the range of visible light or UV light but also in any frequency range where the emission condition can be met i.e. in the radiofrequency range. Different levels of information can be used. Binary information can be based on the absence or presence of detected Cherenkov radiation. The amount or the direction of Cherenkov light can be used. In contrast to a scintillation counter the light production is instantaneous. Detector types In the simple case of a threshold detector the mass-dependent threshold energy allows the discrimination between a lighter particle (which does radiate) and a heavier particle (which does not radiate) of the same energy or momentum. Several threshold stages can
https://en.wikipedia.org/wiki/Phragmites%20australis
Phragmites australis, known as the common reed, is a species of flowering plant in the grass family Poaceae. It is a wetland grass that can grow up to tall and has a cosmopolitan distribution worldwide. Description Phragmites australis commonly forms extensive stands (known as reed beds), which may be as much as or more in extent. Where conditions are suitable it can also spread at or more per year by horizontal runners, which put down roots at regular intervals. It can grow in damp ground, in standing water up to or so deep, or even as a floating mat. The erect stems grow to tall, with the tallest plants growing in areas with hot summers and fertile growing conditions. The leaves are long and broad. The flowers are produced in late summer in a dense, dark purple panicle, about long. Later the numerous long, narrow, sharp pointed spikelets appear greyer due to the growth of long, silky hairs. These eventually help disperse the minute seeds. Taxonomy Recent studies have characterized morphological distinctions between the introduced and native stands of Phragmites australis in North America. The Eurasian phenotype can be distinguished from the North American phenotype by its shorter ligules of up to as opposed to over 1 mm, shorter glumes of under against over 3.2 mm (although there is some overlap in this character), and in culm characteristics. Phragmites australis subsp. americanus – the North American genotype has been described as a distinct species, Phragmites americanus Phragmites australis subsp. australis – the Eurasian genotype Phragmites australis subsp. berlandieri (E.Fourn.) Saltonst. & Hauber Phragmites australis subsp. isiacus (Arcang.) ined. Ecology It is a helophyte (aquatic plant), especially common in alkaline habitats, and it also tolerates brackish water, and so is often found at the upper edges of estuaries and on other wetlands (such as grazing marsh) which are occasionally inundated by the sea. A study demonstrated that P. a
https://en.wikipedia.org/wiki/Star%20domain
In geometry, a set in the Euclidean space is called a star domain (or star-convex set, star-shaped set or radially convex set) if there exists an such that for all the line segment from to lies in This definition is immediately generalizable to any real, or complex, vector space. Intuitively, if one thinks of as a region surrounded by a wall, is a star domain if one can find a vantage point in from which any point in is within line-of-sight. A similar, but distinct, concept is that of a radial set. Definition Given two points and in a vector space (such as Euclidean space ), the convex hull of is called the and it is denoted by where for every vector A subset of a vector space is said to be if for every the closed interval A set is and is called a if there exists some point such that is star-shaped at A set that is star-shaped at the origin is sometimes called a . Such sets are closed related to Minkowski functionals. Examples Any line or plane in is a star domain. A line or a plane with a single point removed is not a star domain. If is a set in the set obtained by connecting all points in to the origin is a star domain. Any non-empty convex set is a star domain. A set is convex if and only if it is a star domain with respect to any point in that set. A cross-shaped figure is a star domain but is not convex. A star-shaped polygon is a star domain whose boundary is a sequence of connected line segments. Properties The closure of a star domain is a star domain, but the interior of a star domain is not necessarily a star domain. Every star domain is a contractible set, via a straight-line homotopy. In particular, any star domain is a simply connected set. Every star domain, and only a star domain, can be "shrunken into itself"; that is, for every dilation ratio the star domain can be dilated by a ratio such that the dilated star domain is contained in the original star domain. The union and intersection of
https://en.wikipedia.org/wiki/Black%20Ribbon%20Movement%20Myanmar
The Black Ribbon Movement Myanmar was a movement of medical professions and medical students against the appointment of military officers to positions within Ministry of Health in Myanmar (Burma) in August 2015. Background Successive Burmese governments since the military took over in the 60s have transferred retired military officers into management positions in various civilian departments and army appointees across various sectors have often stifled the progress of experienced civilian staff. Pe Thet Khin, a paediatrician and Minister for Health, was forced into retirement on 29 July 2014, and was replaced by Than Aung, a former military doctor. On reform, the department of health under Ministry of Health was split into two departments; department of public health and department of medical services, and as a result, about 330 new vacancies appeared. There had been widespread speculation within the Ministry that the minister planned to appoint 350 military officers. On 28 and 30 July, Ministry of Health announced that 14 officers from Ministry of Defence would be transferred to the department. Five of the appointment notices were leaked online in early October. According to the appointments, not only military doctors but also other military officers graduated from Defence Services Academy were appointed to senior positions. These military officers will become assistant and deputy directors of Ministry of Health. As of 2015, more than 100 ex-military officers are serving in various positions within the Ministry appointed before reforms in Myanmar. Protests On 10 August, doctors from Mandalay Orthopaedic Hospital and Taungoo Hospital created a Facebook page named Black Ribbon Movement Myanmar 2015. The page reached 40000 likes within three days after creation. The campaign encouraged anyone, including medical professionals, to make a black ribbon, take a photo of themselves wearing it on their chest, and then share the image on Facebook. People who lead the c
https://en.wikipedia.org/wiki/Amblyanthus%20%28plant%29
Amblyanthus is the name or synonym of a plant genus that may refer to: Amblyanthus A.DC., a genus in the family Primulaceae Amblyanthus (Schltr.) Brieger, a synonym of Dendrobium
https://en.wikipedia.org/wiki/Cox%27s%20theorem
Cox's theorem, named after the physicist Richard Threlkeld Cox, is a derivation of the laws of probability theory from a certain set of postulates. This derivation justifies the so-called "logical" interpretation of probability, as the laws of probability derived by Cox's theorem are applicable to any proposition. Logical (also known as objective Bayesian) probability is a type of Bayesian probability. Other forms of Bayesianism, such as the subjective interpretation, are given other justifications. Cox's assumptions Cox wanted his system to satisfy the following conditions: Divisibility and comparability – The plausibility of a proposition is a real number and is dependent on information we have related to the proposition. Common sense – Plausibilities should vary sensibly with the assessment of plausibilities in the model. Consistency – If the plausibility of a proposition can be derived in many ways, all the results must be equal. The postulates as stated here are taken from Arnborg and Sjödin. "Common sense" includes consistency with Aristotelian logic in the sense that logically equivalent propositions shall have the same plausibility. The postulates as originally stated by Cox were not mathematically rigorous (although more so than the informal description above), as noted by Halpern. However it appears to be possible to augment them with various mathematical assumptions made either implicitly or explicitly by Cox to produce a valid proof. Cox's notation: The plausibility of a proposition given some related information is denoted by . Cox's postulates and functional equations are: The plausibility of the conjunction of two propositions , , given some related information , is determined by the plausibility of given and that of given . In form of a functional equation Because of the associative nature of the conjunction in propositional logic, the consistency with logic gives a functional equation saying that the function is an associative binary
https://en.wikipedia.org/wiki/Interclavicular%20ligament
The interclavicular ligament is a flattened band, which varies considerably in form and size in different individuals, it passes in a curved direction from the upper part of the sternal end of one clavicle to that of the other, and is also attached to the upper margin of the sternum. It is in relation, in front, with the integument and Sternocleidomastoidei; behind, with the Sternothyreoidei.
https://en.wikipedia.org/wiki/Domino%20logic
Domino logic is a CMOS-based evolution of the dynamic logic techniques based on either PMOS or NMOS transistors. It allows a rail-to-rail logic swing. It was developed to speed up circuits, solving the premature cascade problem, typically by inserting small and fast pFETs between domino stages to constrain the interstage cascade velocity to a curtailed maximum—a curtailed deterministic maximum—without requiring other circuit design interlocks. Terminology The term derives from the fact that in domino logic (cascade structure consisting of several stages), each stage ripples the next stage for evaluation, similar to dominoes falling one after the other. Dynamic logic drawbacks In dynamic logic, a problem arises when cascading one gate to the next. The precharge "1" state of the first gate may cause the second gate to discharge prematurely, before the first gate has reached its correct state. This uses up the "precharge" of the second gate, which cannot be restored until the next clock cycle, so there is no recovery from this error. In order to cascade dynamic logic gates, one solution is domino logic, which inserts an ordinary static inverter between stages. While this might seem to defeat the point of dynamic logic, since the inverter has a pFET (one of the main goals of dynamic logic is to avoid pFETs where possible, due to speed), there are two reasons it works well. First, there is no fan-out to multiple pFETs; the dynamic gate connects to exactly one inverter, so the gate is still very fast. Furthermore, since the inverter connects to only nFETs in dynamic logic gates, it too is very fast. Second, the pFET in an inverter can be made smaller than in some types of logic gates. In domino logic cascade structure of several stages, the evaluation of each stage ripples the next stage evaluation, similar to dominoes falling one after the other. Once fallen, the node states cannot return to "1" (until the next clock cycle) just as dominoes, once fallen, cannot s
https://en.wikipedia.org/wiki/Ovarian%20plexus
The ovarian plexus arises from the renal plexus, and is distributed to the ovary, and fundus of the uterus. It is carried in the suspensory ligament of the ovary.
https://en.wikipedia.org/wiki/5-Hydroxyuracil
5-Hydroxyuracil is an oxidized form of cytosine that is produced by the oxidative deamination of cytosines by reactive oxygen species. It does not distort the DNA molecule and is bypassed by replicative DNA polymerases. It can miscode for adenine and is potentially mutagenic.