source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/Activated%20protein%20C%E2%80%93protein%20C%20inhibitor
Activated protein C–protein C inhibitor (APC-PCI) is a complex of activated protein C (APC) and protein C inhibitor (PCI). It has been measured in coagulation testing to evaluate coagulation, thrombosis, and other cardiovascular complications. It is a marker of thrombin generation and indicates hypercoagulability. The half-life of APC-PCI is either 40 minutes or 140minutes. Ethinylestradiol-containing birth control pills have been found to increase levels of APC-PCI to a similar degree as thrombin–antithrombin complex (TAT) and to a greater extent than D-dimer. However, only APC-PCI was able to differentiate between a second- and third-generation birth control pill. Another complex related to APC-PCI is the activated protein C–α1-antitrypsin (APCAT) complex.
https://en.wikipedia.org/wiki/Street%20Fighter%20V
Street Fighter V is a 2.5D fighting game developed by Capcom and Dimps and published by Capcom for PlayStation 4 and Windows in 2016. An arcade version exclusive to Japan was released by Taito in 2019. Similar to the previous games in the Street Fighter series, Street Fighter V features a side-scrolling fighting gameplay system. The game also introduces the "V-Gauge", which builds as the player receives attacks and adds three new skills. The game featured 16 characters at launch, with four of them being new to the series. A main story mode and 30 additional characters were added through updates and downloadable content. According to Capcom, the game is a PlayStation 4 console-exclusive due to both Sony and Capcom having "the same vision for the growth potential in the fighting game space". The game is powered by Unreal Engine 4, and had a beta test prior to its launch. Upon release, the game received mixed reviews, with critics praising the game's graphics and gameplay, but criticizing its lack of content and characters, as well as its technical issues such as broken servers at launch and software bugs. Capcom was expecting the game to sell at least 2 million units by the end of their fiscal year 2016, a target that was missed, selling only 1.4 million units as of March 31, 2016. As of March 2023, the game has sold 7.2 million units, making it the best-selling Street Fighter game on home consoles, surpassing Street Fighter II. Its update, Street Fighter V: Arcade Edition, was released on January 16, 2018, and was received more positively, with improvements to the user interface and content, in particular its single-player modes and the addition of the much-requested Arcade Mode. A second update, titled Street Fighter V: Champion Edition, includes all fighters, stages, and most of the costumes released through its first four seasons. It was released on February 14, 2020, receiving a physical release in the form of the original Street Fighter V disc, containing a v
https://en.wikipedia.org/wiki/Orbitalis%20muscle
The orbitalis muscle is a vestigial or rudimentary nonstriated muscle (smooth muscle) that crosses from the infraorbital groove and sphenomaxillary fissure and is intimately united with the periosteum of the orbit. It was described by Heinrich Müller and is often called Müller's muscle. It lies at the back of the orbit and spans the infraorbital fissure. It is a thin layer of smooth muscle that bridges the inferior orbital fissure. It is supplied by sympathetic nerves, and its function is unknown. Function The muscle forms an important part of the lateral orbital wall in some animals and can act to change the wall's volume in lower mammals, while in humans it is not known to have any significant function, but its contraction may possibly produce a slight forward protrusion of the eyeball. Several sources have suggested a role in the autonomic regulation of the vascular system due to the pattern of innervation of the orbitalis. Pathology Horner's syndrome causes paralysis of the structures of the eye and orbit that receive sympathetic innervation. The signs of Horner's syndrome are ptosis, miosis, anhydrosis, and (apparent) enophthalmos. While some attribute the enophthalmos of Horner's Syndrome to paralysis of the orbitalis muscle, this is inaccurate. Enophthalmos in Horner's syndrome is an illusion created by the subtle ptosis of the upper eyelid caused by paralysis of the superior tarsal muscle. Sinking in of the eye (true enophthalmos) is possibly caused by paralysis of the smooth (orbitalis) muscle in the floor of the orbit. Eponym While the orbitalis muscle is also known as Müller's muscle, the use of this term should be discouraged to avoid confusion with the superior tarsal muscle and the circular fibres of the ciliary muscle.
https://en.wikipedia.org/wiki/Monolinuron
Monolinuron is a pesticide, more specifically a selective systemic herbicide and an algaecide. As an herbicide, it is used to control broad-leaved weeds and annual grasses in vegetable crops such as leeks, potatoes, and dwarf French beans. Monolinuron affects the photosynthesis in weeds. Following uptake of monolinuron through roots and leaves of weeds, monolinuron causes early symptoms of yellowing and die-back of the leaves, eventually resulting in weed death. In fishkeeping, it is used to control blanket weed and hair algae. As a herbicide, it is used in a range of food and non-food crops, e.g. beans and other vegetables, onions, fruits, cereals, potatoes, vines and ornamental plants. It was introduced to use in circa 1965. It is an inhibitor of photosystem II of photosynthesis. It is a derivative of urea and related to paraquat. It is available under various trade names e.g. Monamex, Gramonol, Aresin. It is soluble in water (0.735 g/L), and very soluble in organic solvents (200 g/L in acetone, methanol, and toluene, 3.9 g/L in hexane). Its volatility is low and its leachability is high. It is moderately persistent and moderately mobile in soil, stable in water, and fast degraded in water sediments (DT50 of 22 days). It has low toxicity for mammals (LD50 oral for a rat is 2100 mg/kg).
https://en.wikipedia.org/wiki/Boss%20%28architecture%29
In architecture, a boss is a decorative knob on a ceiling, wall or sculpture. Bosses can often be found in the ceilings of buildings, particularly at the keystones at the intersections of a rib vault. In Gothic architecture, such roof bosses (or ceiling bosses) are often intricately carved with foliage, heraldic devices or other decorations. Many feature animals, birds, or human figures or faces, sometimes realistic, but often Grotesque: the Green Man is a frequent subject. The Romanesque Norwich Cathedral in Norfolk, United Kingdom, has the largest number of painted carved stone bosses in the world; an extensive and varied collection of over one thousand individual pieces. Many of these decorated bosses still bear the original gilt and pigments from the time of their creation. Gallery See also Bossage Lifting boss Three hares
https://en.wikipedia.org/wiki/Breeding%20in%20the%20wild
Breeding in the wild is the natural process of animal reproduction occurring in the natural habitat of a given species. This terminology is distinct from animal husbandry or breeding of species in captivity. Breeding locations are often chosen for very specific requirements of shelter and proximity to food; moreover, the breeding season is a particular time window that has evolved for each species to suit species anatomical, mating-ritual, or climatic and other ecological factors. Many species migrate considerable distances to reach the requisite breeding locations. Certain common characteristics apply to various taxa within the animal kingdom, which traits are often sorted among amphibians, reptiles, mammals, avafauna, arthropods and lower life forms. Amphibians For many amphibians, an annual breeding cycle applies, typically regulated by ambient temperature, precipitation, availability of surface water and food supply. This breeding season is accentuated in temperate regions, where prolonged aestivation or hibernation renders many amphibian species inactive for prolonged periods. Breeding habitats are typically ponds and streams. Mammals Annual breeding cycles sometimes apply to mammals, with regulating environmental effects including seasonal temperature variation and food availability. Migration patterns of a mammal may sometimes govern breeding times. Mammal breeding in the wild sometimes involves the use of maternity dens for birthing and protection of the young. The polar bear is an example of a mammal who uses a maternity den, whose locations are influenced by migration movements of this species to the seasonal Arctic pack ices. In particular, the polar bears who breed in Wapusk National Park need to migrate to the Hudson Bay pack ice. Effects of inbreeding in wild populations Keller and Waller reviewed the effects of inbreeding in wild-populations. Evidence from mammalian and bird populations indicated that inbreeding depression often significantly ad
https://en.wikipedia.org/wiki/Ictal-Interictal%20SPECT%20Analysis%20by%20SPM
ISAS (Ictal-Interictal SPECT Analysis by SPM) is an objective tool for analyzing ictal vs. interictal SPECT scans. The goal of ictal SPECT is to localize the region of seizure onset for epilepsy surgery planning. ISAS was introduced and validated in two recent studies (Chang et al., 2002; McNally et al., 2005). This site is a technical supplement to (McNally et al., 2005), which should enable ISAS to be implemented at any center for further study and analysis. Analysis The basic idea of ISAS is to compute the difference between an ictal and interictal SPECT scan for a single patient. The differences of the ictal/inter-ictal comparison are checked against a healthy normal database to determine the normal expected variation. Significant increases and decreases in CBF (cerebral blood flow) between the interictal and ictal SPECT can then be detected. The analysis is conducted using SPM (statistical parametric mapping). Cerebral blood flow is known to increase during seizures at the site of seizure onset. Since SPECT is an indicator of CBF, increases in the SPECT during seizures can be useful for seizure localization. CBF decreases are more complicated, and occur both during and following seizures in multiple locations. Details of ISAS interpretation can be found in McNally et al., 2005, but in summary: 1. For true ictal SPECT (patient injected before end of seizure), seizure onset can be reliably localized based on SPECT increases. 2. For post-ictal SPECT, seizure onset cannot be reliably localized to a single lobe based on SPECT increase or decreases. However, the side (L or R) of seizure onset can be reliably determined based on which hemisphere has greater overall SPECT decreases (hypoperfusion asymmetry index). The requirements for implementing ISAS are relatively simple. All that is needed is a computer running MATLAB and an operator with sufficient imaging experience to download and implement the SPM (statistical parametric mapping) analysis. NOTE: ISAS
https://en.wikipedia.org/wiki/Direct%20simulation%20Monte%20Carlo
Direct simulation Monte Carlo (DSMC) method uses probabilistic Monte Carlo simulation to solve the Boltzmann equation for finite Knudsen number fluid flows. The DSMC method was proposed by Graeme Bird, emeritus professor of aeronautics, University of Sydney. DSMC is a numerical method for modeling rarefied gas flows, in which the mean free path of a molecule is of the same order (or greater) than a representative physical length scale (i.e. the Knudsen number Kn is greater than 1). In supersonic and hypersonic flows rarefaction is characterized by Tsien's parameter, which is equivalent to the product of Knudsen number and Mach number (KnM) or M/Re, where Re is the Reynolds number. In these rarefied flows, the Navier-Stokes equations can be inaccurate. The DSMC method has been extended to model continuum flows (Kn < 1) and the results can be compared with Navier Stokes solutions. The DSMC method models fluid flows using probabilistic simulation molecules to solve the Boltzmann equation. Molecules are moved through a simulation of physical space in a realistic manner that is directly coupled to physical time such that unsteady flow characteristics can be modeled. Intermolecular collisions and molecule-surface collisions are calculated using probabilistic, phenomenological models. Common molecular models include the hard sphere model, the variable hard sphere (VHS) model, and the variable soft sphere (VSS) model. Various collision models are presented in. Currently, the DSMC method has been applied to the solution of flows ranging from estimation of the Space Shuttle re-entry aerodynamics to the modeling of microelectromechanical systems (MEMS). DSMC Algorithm The direct simulation Monte Carlo algorithm is like molecular dynamics in that the state of the system is given by the positions and velocities of the particles, , for . Unlike molecular dynamics, each particle in a DSMC simulation represents molecules in the physical system that have roughly at the same
https://en.wikipedia.org/wiki/Ageing%20studies
Ageing studies (or age studies) is a field of theoretically, politically, and empirically engaged cultural analysis that was developed by scholars from many different disciplines. Over the past fifteen years the field of ageing studies has flourished, with a growing number of scholars paying attention to the cultural implications of population ageing. History and description of the field Ageing studies break from the traditional field of gerontology by highlighting how biological ageing is mediated by cultural construction, and by emphasising the self-representation of the elderly. The first issue of the academic journal Age, Culture, Humanities includes many essays that address the "coming of age" of this field, in which Stephen Katz draws a comparison between the current state of age studies and gender studies. Ageing studies is an interdisciplinary field, which can be affiliated with the wider approaches found in cultural studies, gender studies, media and film studies, consumer culture, etc. Researchers working in this field interrogate the cultural discourses and practices that construct the meaning of ageing. For example, they invoke calculations of age, asking at what age someone is considered "old" (fifty? sixty?) and referring to the social practice of trying to figure out someone's age. They investigate how public expressions of ageing in the West, such as representations of old age in the news, films and television, create limited views of old age, therefore leading to ageism, a general lack of awareness of diversity, and intergenerational misunderstandings and divisions. Within this perspective, ageing is understood not merely as a biological state, but as a lived experience, embodied and mediated, occurring within specific material and social circumstances. Research in this field is primarily conducted through methodologies associated with the social sciences and the humanities. Notable books in the field Margaret M. Gullette was the first to cal
https://en.wikipedia.org/wiki/Proof%20of%20stake
Proof-of-stake (PoS) protocols are a class of consensus mechanisms for blockchains that work by selecting validators in proportion to their quantity of holdings in the associated cryptocurrency. This is done to avoid the computational cost of proof-of-work (POW) schemes. The first functioning use of PoS for cryptocurrency was Peercoin in 2012, although the scheme, on the surface, still resembled a POW. Description For a blockchain transaction to be recognized, it must be appended to the blockchain. In the proof of stake blockchain the appending entities are named minters or (in the proof of work blockchains this task is carried out by the miners); in most protocols, the validators receive a reward for doing so. For the blockchain to remain secure, it must have a mechanism to prevent a malicious user or group from taking over a majority of validation. PoS accomplishes this by requiring that validators have some quantity of blockchain tokens, requiring potential attackers to acquire a large fraction of the tokens on the blockchain to mount an attack. Proof of work (PoW), another commonly used consensus mechanism, uses a validation of computational prowess to verify transactions, requiring a potential attacker to acquire a large fraction of the computational power of the validator network. This incentivizes consuming huge quantities of energy. PoS is more energy-efficient. Early PoS implementations were plagued by a number of new attacks that exploited the unique vulnerabilities of the PoS protocols. Eventually two dominant designs emerged: so called Byzantine Fault Tolerance-based and chain-based approaches. Bashir identifies three more types of PoS: committee-based PoS (a.k.a. nominated PoS, NPoS); delegated proof of stake (DPoS); liquid proof of stake (LPoS). Attacks The additional vulnerabilities of the PoS schemes are directly related to their advantage, a relatively low amount of calculations to be performed while constructing a blockchain. Long-rang
https://en.wikipedia.org/wiki/Engineering%20diffraction
Engineering diffraction refers to a sub-field of neutron scattering which investigates microstructural features that influence the mechanical properties of materials. These include: lattice strain, a measure of distortion in crystals texture, a measure of grain orientations dislocation density, a measure of the microstructure grain morphology
https://en.wikipedia.org/wiki/List%20of%20mathematical%20identities
This article lists mathematical identities, that is, identically true relations holding in mathematics. Bézout's identity (despite its usual name, it is not, properly speaking, an identity) Binomial inverse theorem Binomial identity Brahmagupta–Fibonacci two-square identity Candido's identity Cassini and Catalan identities Degen's eight-square identity Difference of two squares Euler's four-square identity Euler's identity Fibonacci's identity see Brahmagupta–Fibonacci identity or Cassini and Catalan identities Heine's identity Hermite's identity Lagrange's identity Lagrange's trigonometric identities MacWilliams identity Matrix determinant lemma Newton's identity Parseval's identity Pfister's sixteen-square identity Sherman–Morrison formula Sophie Germain identity Sun's curious identity Sylvester's determinant identity Vandermonde's identity Woodbury matrix identity Identities for classes of functions Exterior calculus identities Fibonacci identities: Combinatorial Fibonacci identities and Other Fibonacci identities Hypergeometric function identities List of integrals of logarithmic functions List of topics related to List of trigonometric identities Inverse trigonometric functions Logarithmic identities Summation identities Vector calculus identities See also External links A Collection of Algebraic Identities Matrix Identities Identities
https://en.wikipedia.org/wiki/Copy%20protection
Copy protection, also known as content protection, copy prevention and copy restriction, describes measures to enforce copyright by preventing the reproduction of software, films, music, and other media. Copy protection is most commonly found on videotapes, DVDs, Blu-ray discs, HD-DVDs, computer software discs, video game discs and cartridges, audio CDs and some VCDs. Some methods of copy protection have also led to criticism because it caused inconvenience for paying consumers or secretly installed additional or unwanted software to detect copying activities on the consumer's computer. Making copy protection effective while protecting consumer rights remains a problem with media publication. Terminology Media corporations have always used the term copy protection, but critics argue that the term tends to sway the public into identifying with the publishers, who favor restriction technologies, rather than with the users. Copy prevention and copy control may be more neutral terms. "Copy protection" is a misnomer for some systems, because any number of copies can be made from an original and all of these copies will work, but only in one computer, or only with one dongle, or only with another device that cannot be easily copied. The term is also often related to, and confused with, the concept of digital restrictions management. Digital restrictions management is a more general term because it includes all sorts of management of works, including copy restrictions. Copy restriction may include measures that are not digital. A more appropriate term may be "technological protection measures" (TPMs), which is often defined as the use of technological tools in order to restrict the use or access to a work. Business rationale Unauthorized copying and distribution accounted for $2.4 billion per year in lost revenue in the United States alone in 1990, and is assumed to be causing impact on revenues in the music and the video game industry, leading to proposal of stri
https://en.wikipedia.org/wiki/Frog%20battery
A frog battery is an electrochemical battery consisting of a number of dead frogs (or sometimes live ones), which form the cells of the battery connected in a series arrangement. It is a kind of biobattery. It was used in early scientific investigations of electricity and academic demonstrations. The principle behind the battery is the injury potential created in a muscle when it is damaged, although this was not fully understood in the 18th and 19th centuries; the potential being caused incidentally due to the dissection of the frog's muscles. The frog battery is an example of a class of biobatteries which can be made from any number of animals. The general term for an example of this class is the muscular pile. The first well-known frog battery was created by Carlo Matteucci in 1845, but there had been others before him. Matteucci also created batteries out of other animals, and Giovanni Aldini created a battery from ox heads. Background In the early days of electrical research, a common method of detecting electric current was by means of a frog's leg galvanoscope. A good supply of live frogs was kept to hand by the researcher ready to have their legs prepared for the galvanoscope. Frogs were therefore a convenient material to use in other experiments. They were small, easily handled, the legs were especially sensitive to electric current, and they carried on responding longer than other animal candidates for this role. Preparation It was usual to use the thighs of frogs for the battery construction. The legs of the frog were first skinned, then the lower leg was cut off at the knee joint and discarded. Damaging the muscle during this procedure would detract from the results. The thigh muscle was then cut in two transversely to produce two half-thighs. Only the lower, conical shaped piece was kept. The half-thighs were then laid on an insulator of varnished wood so arranged that the inside surface of one was in contact with the outside surface o
https://en.wikipedia.org/wiki/AMD%20PowerTune
AMD PowerTune is a series of dynamic frequency scaling technologies built into some AMD GPUs and APUs that allow the clock speed of the processor to be dynamically changed (to different P-states) by software. This allows the processor to meet the instantaneous performance needs of the operation being performed, while minimizing power draw, heat generation and noise avoidance. AMD PowerTune aims to solve thermal design power and performance constraints. Besides the reduced energy consumption, AMD PowerTune helps to lower the noise levels created by the cooling in desktop computers, and extends battery life in mobile devices. AMD PowerTune is the successor to AMD PowerPlay. Support for "PowerPlay" was added to the Linux kernel driver "amdgpu" on November 11, 2015. As a lecture from CCC in 2014 shows, AMD's x86-64 SMU firmware is executed on some LatticeMico32 and PowerTune was modeled using Matlab. This is similar to Nvidia's PDAEMON, the RTOS responsible for power on their GPUs. Overview AMD PowerTune was introduced in the TeraScale 3 (VLIW4) with Radeon HD 6900 on 15 December 2010 and has been available in different development stages on Radeon- and AMD FirePro-branded products ever since. Over the years, reviews which document the development of AMD PowerTune have been published by AnandTech. An additional technology named AMD ZeroCore Power has been available since the Radeon HD 7000 Series, implementing the Graphics Core Next microarchitecture. The pointlessness of a fixed clock frequency was accredited in January 2014 by SemiAccurate. Operating system support AMD Catalyst is available for Microsoft Windows and Linux and supports AMD PowerTune since version. The free and open-source "Radeon" graphics device driver has some support for AMD PowerTune, see "Enduro". Feature overview for AMD APUs Feature overview for AMD graphics cards See also AMD Cool'n'Quiet (for desktop CPUs) AMD PowerNow! (for laptop CPUs) AMD Turbo Core (for CPUs) AMD PowerX
https://en.wikipedia.org/wiki/Signature%20%28topology%29
In the field of topology, the signature is an integer invariant which is defined for an oriented manifold M of dimension divisible by four. This invariant of a manifold has been studied in detail, starting with Rokhlin's theorem for 4-manifolds, and Hirzebruch signature theorem. Definition Given a connected and oriented manifold M of dimension 4k, the cup product gives rise to a quadratic form Q on the 'middle' real cohomology group . The basic identity for the cup product shows that with p = q = 2k the product is symmetric. It takes values in . If we assume also that M is compact, Poincaré duality identifies this with which can be identified with . Therefore the cup product, under these hypotheses, does give rise to a symmetric bilinear form on H2k(M,R); and therefore to a quadratic form Q. The form Q is non-degenerate due to Poincaré duality, as it pairs non-degenerately with itself. More generally, the signature can be defined in this way for any general compact polyhedron with 4n-dimensional Poincaré duality. The signature of M is by definition the signature of Q, that is, where any diagonal matrix defining Q has positive entries and negative entries. If M is not connected, its signature is defined to be the sum of the signatures of its connected components. Other dimensions If M has dimension not divisible by 4, its signature is usually defined to be 0. There are alternative generalization in L-theory: the signature can be interpreted as the 4k-dimensional (simply connected) symmetric L-group or as the 4k-dimensional quadratic L-group and these invariants do not always vanish for other dimensions. The Kervaire invariant is a mod 2 (i.e., an element of ) for framed manifolds of dimension 4k+2 (the quadratic L-group ), while the de Rham invariant is a mod 2 invariant of manifolds of dimension 4k+1 (the symmetric L-group ); the other dimensional L-groups vanish. Kervaire invariant When is twice an odd integer (singly even), the same constru
https://en.wikipedia.org/wiki/Melatonin%20receptor%201A
Melatonin receptor type 1A is a protein that in humans is encoded by the MTNR1A gene. Function This gene encodes the MT1 protein, one of two high-affinity forms of a receptor for melatonin, the primary hormone secreted by the pineal gland. This receptor is a G protein-coupled, 7-transmembrane receptor that is responsible for melatonin effects on mammalian circadian rhythm and reproductive alterations affected by day length. The receptor is an integral membrane protein that is readily detectable and localized to two specific regions of the brain. The hypothalamic suprachiasmatic nucleus appears to be involved in circadian rhythm while the hypophysial pars tuberalis may be responsible for the reproductive effects of melatonin. Ligands Melatonin – full agonist Afobazole – agonist Agomelatine – agonist See also Melatonin receptor Discovery and development of melatonin receptor agonists
https://en.wikipedia.org/wiki/Mobile%20operating%20system
A mobile operating system is an operating system for smartphones, tablets, smartwatches, smartglasses, or other non-laptop personal mobile computing devices. While computers such as typical/mobile laptops are "mobile", the operating systems used on them are generally not considered mobile ones, as they were originally designed for desktop computers that historically did not have or need specific mobile features. This line distinguishing mobile and other forms has become blurred in recent years, due to the fact that newer devices have become smaller and more mobile unlike hardware of the past. Key notabilities blurring this line are the introduction of tablet computers and light-weight laptops and the hybridization of the two in 2-in-1 PCs. Mobile operating systems combine features of a desktop computer operating system with other features useful for mobile or handheld use, and usually including a wireless inbuilt modem and SIM tray for telephony and data connection. In Q1 2018, over 123 million smartphones were sold (highest ever recorded) with 60.2 percent running Android and 20.9 percent running iOS. Nonetheless, although not as many as 2018 (1.56 billion), 2021 still had soaring sales, 1.43 billion to be exact with 53.32 percent being Android. Android alone is more popular than the popular desktop operating system Microsoft Windows, and in general smartphone use (even without tablets) outnumbers desktop use. Mobile devices, with mobile communications abilities (e.g., smartphones), contain two mobile operating systemsthe main user-facing software platform is supplemented by a second low-level proprietary real-time operating system which operates the radio and other hardware. Research has shown that these low-level systems may contain a range of security vulnerabilities permitting malicious base stations to gain high levels of control over the mobile device. Mobile operating systems have majority use since 2017 (measured by web use); with even only the smartph
https://en.wikipedia.org/wiki/Oxford%20Centre%20for%20Gene%20Function
The Oxford Centre for Gene Function is a multidisciplinary research institute in the University of Oxford, England. It is directed by Frances Ashcroft, Kay Davies and Peter Donnelly. It involves the departments of Human anatomy and genetics, Physiology, and Statistics. External links Oxford Centre for Gene Function website Wellcome Trust Centre for Human Genetics Departments of the University of Oxford Genetics in the United Kingdom Human genetics Research institutes in Oxford
https://en.wikipedia.org/wiki/Hafner%E2%80%93Sarnak%E2%80%93McCurley%20constant
The Hafner–Sarnak–McCurley constant is a mathematical constant representing the probability that the determinants of two randomly chosen square integer matrices will be relatively prime. The probability depends on the matrix size, n, in accordance with the formula where pk is the kth prime number. The constant is the limit of this expression as n approaches infinity. Its value is roughly 0.3532363719... .
https://en.wikipedia.org/wiki/Nebular%20hypothesis
The nebular hypothesis is the most widely accepted model in the field of cosmogony to explain the formation and evolution of the Solar System (as well as other planetary systems). It suggests the Solar System is formed from gas and dust orbiting the Sun which clumped up together to form the planets. The theory was developed by Immanuel Kant and published in his Universal Natural History and Theory of the Heavens (1755) and then modified in 1796 by Pierre Laplace. Originally applied to the Solar System, the process of planetary system formation is now thought to be at work throughout the universe. The widely accepted modern variant of the nebular theory is the solar nebular disk model (SNDM) or solar nebular model. It offered explanations for a variety of properties of the Solar System, including the nearly circular and coplanar orbits of the planets, and their motion in the same direction as the Sun's rotation. Some elements of the original nebular theory are echoed in modern theories of planetary formation, but most elements have been superseded. According to the nebular theory, stars form in massive and dense clouds of molecular hydrogen—giant molecular clouds (GMC). These clouds are gravitationally unstable, and matter coalesces within them to smaller denser clumps, which then rotate, collapse, and form stars. Star formation is a complex process, which always produces a gaseous protoplanetary disk (proplyd) around the young star. This may give birth to planets in certain circumstances, which are not well known. Thus the formation of planetary systems is thought to be a natural result of star formation. A Sun-like star usually takes approximately 1 million years to form, with the protoplanetary disk evolving into a planetary system over the next 10–100 million years. The protoplanetary disk is an accretion disk that feeds the central star. Initially very hot, the disk later cools in what is known as the T Tauri star stage; here, formation of small dust grains ma
https://en.wikipedia.org/wiki/Normal%20morphism
In category theory and its applications to mathematics, a normal monomorphism or conormal epimorphism is a particularly well-behaved type of morphism. A normal category is a category in which every monomorphism is normal. A conormal category is one in which every epimorphism is conormal. Definition A monomorphism is normal if it is the kernel of some morphism, and an epimorphism is conormal if it is the cokernel of some morphism. A category C is binormal if it's both normal and conormal. But note that some authors will use the word "normal" only to indicate that C is binormal. Examples In the category of groups, a monomorphism f from H to G is normal if and only if its image is a normal subgroup of G. In particular, if H is a subgroup of G, then the inclusion map i from H to G is a monomorphism, and will be normal if and only if H is a normal subgroup of G. In fact, this is the origin of the term "normal" for monomorphisms. On the other hand, every epimorphism in the category of groups is conormal (since it is the cokernel of its own kernel), so this category is conormal. In an abelian category, every monomorphism is the kernel of its cokernel, and every epimorphism is the cokernel of its kernel. Thus, abelian categories are always binormal. The category of abelian groups is the fundamental example of an abelian category, and accordingly every subgroup of an abelian group is a normal subgroup.
https://en.wikipedia.org/wiki/Stipe%20%28botany%29
In botany, a stipe is a stalk that supports some other structure. The precise meaning is different depending on which taxonomic group is being described. In the case of ferns, the stipe is only the petiole from the rootstock to the beginning of the leaf tissue, or lamina. The continuation of the structure within the lamina is then termed a rachis. In flowering plants, the term is often used in reference to a stalk that sometimes supports a flower's ovary. In orchids, the stipe or caudicle is the stalk-like support of the pollinia. It is a non-viscid band or strap connecting the pollinia with the viscidium (the viscid part of the rostellum or beak). A stipe is also a structure found in organisms that are studied by botanists but that are no longer classified as plants. It may be the stem-like part of the thallus of a mushroom or a seaweed, and is particularly common among brown algae such as kelp. The stipe of a kelp often contains a central region of cells that, like the phloem of vascular plants, serves to transport nutrients within the alga. See also Rachis Stipe (mycology), the stalk supporting the fruiting body of some fungi.
https://en.wikipedia.org/wiki/Eugene%20Isaacson
Eugene Isaacson (1919–2008), was a US mathematician who pioneered modern numerical analysis. He was a mathematics and physics graduate of City College in New York, he then entered the graduate program in mathematics at New York University under the supervision of Kurt Otto Friedrichs, gaining a PhD on water waves on sloping beaches in 1949. His academic career was then spent at the Courant Institute until his retirement. Despite an initial interest in topology, Isaacson worked for most of his career in applied and computational mathematics and is best known for his work on the numerical solution of differential equations. His book with Herbert Keller Analysis of Numerical Methods was a leading textbook at the time. He served as editor for Mathematics of Computation and the SIAM Journal on Numerical Analysis. A special issue of Mathematics of Computation was published in 1989 to celebrate his 70th birthday. The contributors included his sons David and Eli Isaacson, also mathematicians, who wrote a paper together for this volume. The paper is dedicated to their father and their mother Muriel "for making this collaboration possible".
https://en.wikipedia.org/wiki/Clustering%20coefficient
In graph theory, a clustering coefficient is a measure of the degree to which nodes in a graph tend to cluster together. Evidence suggests that in most real-world networks, and in particular social networks, nodes tend to create tightly knit groups characterised by a relatively high density of ties; this likelihood tends to be greater than the average probability of a tie randomly established between two nodes (Holland and Leinhardt, 1971; Watts and Strogatz, 1998). Two versions of this measure exist: the global and the local. The global version was designed to give an overall indication of the clustering in the network, whereas the local gives an indication of the embeddedness of single nodes. Local clustering coefficient The local clustering coefficient of a vertex (node) in a graph quantifies how close its neighbours are to being a clique (complete graph). Duncan J. Watts and Steven Strogatz introduced the measure in 1998 to determine whether a graph is a small-world network. A graph formally consists of a set of vertices and a set of edges between them. An edge connects vertex with vertex . The neighbourhood for a vertex is defined as its immediately connected neighbours as follows: We define as the number of vertices, , in the neighbourhood, , of a vertex. The local clustering coefficient for a vertex is then given by a proportion of the number of links between the vertices within its neighbourhood divided by the number of links that could possibly exist between them. For a directed graph, is distinct from , and therefore for each neighbourhood there are links that could exist among the vertices within the neighbourhood ( is the number of neighbours of a vertex). Thus, the local clustering coefficient for directed graphs is given as An undirected graph has the property that and are considered identical. Therefore, if a vertex has neighbours, edges could exist among the vertices within the neighbourhood. Thus, the local cluste
https://en.wikipedia.org/wiki/Volumetric%20display
A volumetric display device is a display device that forms a visual representation of an object in three physical dimensions, as opposed to the planar image of traditional screens that simulate depth through a number of different visual effects. One definition offered by pioneers in the field is that volumetric displays create 3D imagery via the emission, scattering, or relaying of illumination from well-defined regions in (x,y,z) space. A true volumetric display produces in the observer a visual experience of a material object in three-dimensional space, even though no such object is present. The perceived object displays characteristics similar to an actual material object by allowing the observer to view it from any direction, to focus a camera on a specific detail, and to see perspective – meaning that the parts of the image closer to the viewer appear larger than those further away. Volumetric 3D displays are technically not autostereoscopic, even though they create three-dimensional imagery visible to the unaided eye. This is because the displays do not generate stereoscopic images; They naturally provide focally-accurate holographic wavefronts to the eyes. Due to this, they have accurate characteristics of material objects such as focal depth, motion parallax, and vergence. Volumetric displays are one of several kinds of 3D displays. Other types are stereoscopes, view-sequential displays, electro-holographic displays, "two view" displays, and panoramagrams. Although first postulated in 1912, and a staple of science fiction, volumetric displays are still not widely used in everyday life. There are numerous potential markets for volumetric displays with use cases including medical imaging, mining, education, advertising, simulation, video games, communication and geophysical visualisation. When compared to other 3D visualisation tools such as virtual reality, volumetric displays offer an inherently different mode of interaction, providing the opport
https://en.wikipedia.org/wiki/Furazolidone
Furazolidone is a nitrofuran antibacterial agent and monoamine oxidase inhibitor (MAOI). It is marketed by Roberts Laboratories under the brand name Furoxone and by GlaxoSmithKline as Dependal-M. Medical uses Furazolidone has been used in human and veterinary medicine. It has a broad spectrum of activity being active against Gram positivei Clostridium perfringens Corynebacterium pyogenes Streptococci Staphylococci Gram negative Escherichia coli Salmonella dublin Salmonella typhimurium Shigella Protozoa Giardia lamblia Eimeria species Histomonas meleagridis Use in humans In humans it has been used to treat diarrhoea and enteritis caused by bacteria or protozoan infections, including traveler's diarrhoea, cholera and bacteremic salmonellosis. From the early 1970's it has been used in China to treat peptic ulcers, where the mechanism is treatment of the causative Helicobacter pylori infection. In 2002, a journal article suggested it's use in treatment of helicobacter pylori infections in children. Furazolidone has also been used for giardiasis (due to Giardia lamblia), amoebiasis and shigellosis also though it is not a first line treatment. Use in animals As a veterinary medicine, furazolidone has been used with some success to treat salmonids for Myxobolus cerebralis infections. It has also been used in aquaculture. Since furazolidone is a nitrofuran antibiotic, its use in food animals is currently prohibited by the FDA under the Animal Medicinal Drug Use Clarification Act, 1994. Furazolidone is no longer available in the US. Use in laboratory It is used to differentiate micrococci and staphylococci. Mechanism of action It is believed to work by crosslinking of DNA. Side effects Though an effective antibiotic when all others fail, against extremely drug resistant infections, it has many side effects. including inhibition of monoamine oxidase, and as with other nitrofurans generally, minimum inhibitory concentrations also produce systemic toxic
https://en.wikipedia.org/wiki/Paleoethnobotany
Paleoethnobotany (also spelled palaeoethnobotany), or archaeobotany, is the study of past human-plant interactions through the recovery and analysis of ancient plant remains. Both terms are synonymous, though paleoethnobotany (from the Greek words palaios [παλαιός] meaning ancient, ethnos [έθνος] meaning race or ethnicity, and votano [βότανο] meaning plants) is generally used in North America and acknowledges the contribution that ethnographic studies have made towards our current understanding of ancient plant exploitation practices, while the term archaeobotany (from the Greek words archaios [αρχαίος] meaning ancient and votano) is preferred in Europe and emphasizes the discipline's role within archaeology. As a field of study, paleoethnobotany is a subfield of environmental archaeology. It involves the investigation of both ancient environments and human activities related to those environments, as well as an understanding of how the two co-evolved. Plant remains recovered from ancient sediments within the landscape or at archaeological sites serve as the primary evidence for various research avenues within paleoethnobotany, such as the origins of plant domestication, the development of agriculture, paleoenvironmental reconstructions, subsistence strategies, paleodiets, economic structures, and more. Paleoethnobotanical studies are divided into two categories: those concerning the Old World (Eurasia and Africa) and those that pertain to the New World (the Americas). While this division has an inherent geographical distinction to it, it also reflects the differences in the flora of the two separate areas. For example, maize only occurs in the New World, while olives only occur in the Old World. Within this broad division, paleoethnobotanists tend to further focus their studies on specific regions, such as the Near East or the Mediterranean, since regional differences in the types of recovered plant remains also exist. Macrobotanical vs. microbotanical remains
https://en.wikipedia.org/wiki/Open-access%20repository
An open repository or open-access repository is a digital platform that holds research output and provides free, immediate and permanent access to research results for anyone to use, download and distribute. To facilitate open access such repositories must be interoperable according to the Open Archives Initiative Protocol for Metadata Harvesting (OAI-PMH). Search engines harvest the content of open access repositories, constructing a database of worldwide, free of charge available research. Open-access repositories, such as an institutional repository or disciplinary repository, provide free access to research for users outside the institutional community and are one of the recommended ways to achieve the open access vision described in the Budapest Open Access Initiative definition of open access. This is sometimes referred to as the self-archiving or "green" route to open access. Benefits The benefits of open-access repositories are: Opening up outputs of the institution to a worldwide audience; Maximizing the visibility and impact of these outputs as a result; Showcasing the institution to interested constituencies – prospective staff, prospective students and other stakeholders; Collecting and curating digital output; Managing and measuring research and teaching activities; Providing a workspace for work-in-progress, and for collaborative or large-scale projects; Enabling and encouraging interdisciplinary approaches to research; Facilitating the development and sharing of digital teaching materials and aids, and Supporting student endeavours, providing access to theses and dissertations and a location for the development of e-portfolios. Software The most frequently used repository software for open repositories according to OpenDOAR are Digital Commons, DSpace and EPrints. Other examples are arXiv, bioRxiv, Dryad, Figshare, Open Science Framework, Samvera, Ubiquity Repositories and invenio (solution used by Zenodo). See also CORE (research ser
https://en.wikipedia.org/wiki/Treiber%20stack
The Treiber stack algorithm is a scalable lock-free stack utilizing the fine-grained concurrency primitive compare-and-swap. It is believed that R. Kent Treiber was the first to publish it in his 1986 article "Systems Programming: Coping with Parallelism". Basic principle The basic principle for the algorithm is to only add something new to the stack once you know the item you are trying to add is the only thing that has been added since you began the operation. This is done by using compare-and-swap. Pushing an item to the stack is done by first taking the top of the stack (old head) and placing it after your new item to create a new head. You then compare the old head to the current head. If the two are matching then you can swap old head to the new one, if not then it means another thread has added an item to the stack, in which case you must try again. When popping an item from the stack, before returning the item you must check that another thread has not added a new item since the operation began. Correctness In some languages—particularly, those without garbage collection—the Treiber stack can be at risk for the ABA problem. When a process is about to remove an element from the stack (just before the compare and set in the pop routine below) another process can change the stack such that the head is the same, but the second element is different. The compare and swap will set the head of the stack to the old second element in the stack mixing up the complete data structure. However, the Java version on this page is not subject to this problem, because of the stronger guarantees offered by the Java runtime (it is impossible for a newly created, unaliased object reference to be reference-equal to any other reachable object.) Testing for failures such as ABA can be exceedingly difficult, because the problematic sequence of events is very rare. Model checking is an excellent way to uncover such problems. See for instance exercise 7.3.3 in "Modeling and an
https://en.wikipedia.org/wiki/Ore%20algebra
In computer algebra, an Ore algebra is a special kind of iterated Ore extension that can be used to represent linear functional operators, including linear differential and/or recurrence operators. The concept is named after Øystein Ore. Definition Let be a (commutative) field and be a commutative polynomial ring (with when ). The iterated skew polynomial ring is called an Ore algebra when the and commute for , and satisfy , for . Properties Ore algebras satisfy the Ore condition, and thus can be embedded in a (skew) field of fractions. The constraint of commutation in the definition makes Ore algebras have a non-commutative generalization theory of Gröbner basis for their left ideals.
https://en.wikipedia.org/wiki/Factory%20Interface%20Network%20Service
FINS, Factory Interface Network Service, is a network protocol used by Omron PLCs, over different physical networks like Ethernet, Controller Link, DeviceNet and RS-232C. The FINS communications service was developed by Omron to provide a consistent way for PLCs and computers on various networks to communicate. Compatible network types include Ethernet, Host Link, Controller Link, SYSMAC LINK, SYSMAC WAY, and Toolbus. FINS allows communications between nodes up to three network levels. A direct connection between a computer and a PLC via Host Link is not considered a network level.
https://en.wikipedia.org/wiki/Carboxypeptidase%20E
Carboxypeptidase E (CPE), also known as carboxypeptidase H (CPH) and enkephalin convertase, is an enzyme that in humans is encoded by the CPE gene. This enzyme catalyzes the release of C-terminal arginine or lysine residues from polypeptides. CPE is involved in the biosynthesis of most neuropeptides and peptide hormones. The production of neuropeptides and peptide hormones typically requires two sets of enzymes that cleave the peptide precursors, which are small proteins. First, proprotein convertases cut the precursor at specific sites to generate intermediates containing C-terminal basic residues (lysine and/or arginine). These intermediates are then cleaved by CPE to remove the basic residues. For some peptides, additional processing steps, such as C-terminal amidation, are subsequently required to generate the bioactive peptide, although for many peptides the action of the proprotein convertases and CPE is sufficient to produce the bioactive peptide. Tissue distribution Carboxypeptidase E is found in brain and throughout the neuroendocrine system, including the endocrine pancreas, pituitary, and adrenal gland chromaffin cells. Within cells, carboxypeptidase E is present in the secretory granules along with its peptide substrates and products. Carboxypeptidase E is a glycoprotein that exists in both membrane-associated and soluble forms. The membrane-binding is due to an amphiphilic α-helix within the C-terminal region of the protein. Species distribution Carboxypeptidase E is found in all species of vertebrates that have been examined, and is also present in many other organisms that have been studied (nematode, sea slug). Carboxypeptidase E is not found in the fruit fly (Drosophila), and another enzyme (presumably carboxypeptidase D) fills in for carboxypeptidase E in this organism. In humans, CPE is encoded by the CPE gene. Function Carboxypeptidase E functions in the production of nearly all neuropeptides and peptide hormones. The enzyme acts as a
https://en.wikipedia.org/wiki/Lumican
Lumican, also known as LUM, is an extracellular matrix protein that, in humans, is encoded by the LUM gene on chromosome 12. Structure Lumican is a proteoglycan Class II member of the small leucine-rich proteoglycan (SLRP) family that includes decorin, biglycan, fibromodulin, keratocan, epiphycan, and osteoglycin. Like the other SLRPs, lumican has a molecular weight of about 40 kiloDaltons and has four major intramolecular domains: a signal peptide of 16 amino acid residues; a negatively-charged N-terminal domain containing sulfated tyrosine and disulfide bond(s); ten tandem leucine-rich repeats allowing lumican to bind to other extracellular components such as collagen; a carboxyl terminal domain of 50 amino acid residues containing two conserved cysteines 32 residues apart. There are four N-linked sites within the leucine-rich repeat domain of the protein core that can be substituted with keratan sulfate. The core protein of lumican (like decorin and fibromodulin) is horseshoe shaped. This enables it bind to collagen molecules within a collagen fibril, thus helping keep adjacent fibrils apart. Function Lumican is a major keratan sulfate proteoglycan of the cornea but is ubiquitously distributed in most mesenchymal tissues throughout the body. Lumican is involved in collagen fibril organization and circumferential growth, corneal transparency, and epithelial cell migration and tissue repair. Corneal transparency is possible due to the exact alignment of collagen fibers by lumican (and keratocan) in the intrafibrillar space. Clinical significance Mice that have the lumican gene knocked out (Lum-/-) develop opacities of the cornea in both eyes and fragile skin. The lumican (LUM) gene was thought to be a candidate susceptibility gene for high myopia; however, a meta-analysis showed no association between LUM polymorphism and high myopia susceptibility in all genetic models studied. Lum knockout mice also have abnormal collagen in their heart tissue, wi
https://en.wikipedia.org/wiki/Two%20capacitor%20paradox
The two capacitor paradox or capacitor paradox is a paradox, or counterintuitive thought experiment, in electric circuit theory. The thought experiment is usually described as follows: Two identical capacitors are connected in parallel with an open switch between them. One of the capacitors is charged with a voltage of , the other is uncharged. When the switch is closed, some of the charge on the first capacitor flows into the second, reducing the voltage on the first and increasing the voltage on the second. When a steady state is reached and the current goes to zero, the voltage on the two capacitors must be equal since they are connected together. Since they both have the same capacitance the charge will be divided equally between the capacitors so each capacitor will have a charge of and a voltage of . At the beginning of the experiment the total initial energy in the circuit is the energy stored in the charged capacitor: At the end of the experiment the final energy is equal to the sum of the energy in the two capacitors Thus the final energy is equal to half of the initial energy . Where did the other half of the initial energy go? Solutions This problem has been discussed in electronics literature at least as far back as 1955. Unlike some other paradoxes in science, this paradox is not due to the underlying physics, but to the limitations of the 'ideal circuit' conventions used in circuit theory. The description specified above is not physically realizable if the circuit is assumed to be made of ideal circuit elements, as is usual in circuit theory. If the wires connecting the two capacitors, the switch, and the capacitors themselves are idealized as having no electrical resistance or inductance as is usual, then closing the switch would connect points at different voltage with a perfect conductor, causing an infinite current to flow. Therefore a solution requires that one or more of the 'ideal' characteristics of the elements in t
https://en.wikipedia.org/wiki/PKCS%201
In cryptography, PKCS #1 is the first of a family of standards called Public-Key Cryptography Standards (PKCS), published by RSA Laboratories. It provides the basic definitions of and recommendations for implementing the RSA algorithm for public-key cryptography. It defines the mathematical properties of public and private keys, primitive operations for encryption and signatures, secure cryptographic schemes, and related ASN.1 syntax representations. The current version is 2.2 (2012-10-27). Compared to 2.1 (2002-06-14), which was republished as RFC 3447, version 2.2 updates the list of allowed hashing algorithms to align them with FIPS 180-4, therefore adding SHA-224, SHA-512/224 and SHA-512/256. Keys The PKCS #1 standard defines the mathematical definitions and properties that RSA public and private keys must have. The traditional key pair is based on a modulus, , that is the product of two distinct large prime numbers, and , such that . Starting with version 2.1, this definition was generalized to allow for multi-prime keys, where the number of distinct primes may be two or more. When dealing with multi-prime keys, the prime factors are all generally labeled as for some , such that: for As a notational convenience, and . The RSA public key is represented as the tuple , where the integer is the public exponent. The RSA private key may have two representations. The first compact form is the tuple , where is the private exponent. The second form has at least five terms , or more for multi-prime keys. Although mathematically redundant to the compact form, the additional terms allow for certain computational optimizations when using the key. In particular, the second format allows to derive the public key. Primitives The standard defines several basic primitives. The primitive operations provide the fundamental instructions for turning the raw mathematical formulas into computable algorithms. I2OSP - Integer to Octet String Primitive - Conve
https://en.wikipedia.org/wiki/Acoustic%20tubercle
The acoustic tubercle is a nucleus on the end of the cochlear nerve. The cochlear nerve is lateral to the root of the vestibular nerve. Its fibers end in two nuclei: one, the accessory nucleus, lies immediately in front of the inferior peduncle; the other, the acoustic tubercle, somewhat lateral to it.
https://en.wikipedia.org/wiki/Wilfred%20Cockcroft
Sir Wilfred Cockcroft (7 June 1923 – 1999) was an eminent mathematics educator from the University of Hull. In 1978 he was commissioned by the then Labour government to chair a comprehensive inquiry into the teaching of mathematics in primary and secondary schools in England and Wales. The committee of inquiry produced its report in 1982, published as Mathematics Counts but widely known as "the Cockcroft report". Cockcroft was knighted in 1983, and in May 1984 was awarded an honorary degree from the Open University as Doctor of the University. External links Tony Crilly (2001), Memories of Sir Wilfred Cockcroft 1923–1999, The Mathematical Gazette 85 (502), 72. Obituary, Bulletin of the London Mathematical Society (2005), 37, 149-155 Short biography, and account of archives at Hull University, Access to Archives 1923 births 1999 deaths 20th-century English mathematicians
https://en.wikipedia.org/wiki/Evolutionarily%20significant%20unit
An evolutionarily significant unit (ESU) is a population of organisms that is considered distinct for purposes of conservation. Delineating ESUs is important when considering conservation action. This term can apply to any species, subspecies, geographic race, or population. Often the term "species" is used rather than ESU, even when an ESU is more technically considered a subspecies or variety rather than a biological species proper. In marine animals the term "stock" is often used as well. Definition Definitions of an ESU generally include at least one of the following criteria: Current geographic separation, Genetic differentiation at neutral markers among related ESUs caused by past restriction of gene flow, or Locally adapted phenotypic traits caused by differences in selection. Criterion 2 considers the gene flow between populations, measured by FST. A high degree of differentiation between two populations among genes that provide no adaptive advantage to either population (known as neutral markers) implies a lack of gene flow, showing that random drift has occurred in isolation from other populations. Very few migrants per generation are needed to prevent strong differentiation of neutral markers. Even a single migrant per generation may be enough for neutral markers to show gene flow between populations, making it difficult to differentiate the populations through neutral markers. Criterion 3 does not consider neutral genetic markers, instead looking at locally adapted traits of the population. Local adaptations may be present even with some gene flow from other populations, and even when there is little differentiation at neutral markers among ESUs. Reciprocal transplantation experiments are necessary to test for genetic differentiation for phenotypic traits, and differences in selection gradients across habitats. Such experiments are generally more difficult than the fixation index tests of criterion 2, and may be impossible for very rare or endange
https://en.wikipedia.org/wiki/Vicine
Vicine is an alkaloid glycoside found mainly in fava beans, which are also called broad beans (Vicia faba). Vicine is toxic in individuals who have a hereditary loss of the enzyme glucose-6-phosphate dehydrogenase. It causes haemolytic anaemia, called favism. The formation of vicine in Vicia faba has been studied, but this natural formation has not yet been found. History Vicine was initially isolated in 1870 from the seeds of Vicia sativa by a method of extraction with sulfuric acid and subsequent precipitation with mercury sulfate (HgSO4). Later vicine was also found in other Vicia species, namely Vicia faba, beet juice and peas. The chemical structure of the compound was built gradually. First the glycosidic nature of the compound was recognized in 1896. The same year the aglycone of vicine, divicine, was isolated. In the beginning of the 20th century the pyrimidine structure was recognized. Despite these initial successes, the correct formula of vicine was determined only in 1953 and it is 2,4-diamino-6-oxypyrimidine-5-(ß-d-glucopyranoside). Metabolism Vicine is an inactive compound in the body. When vicine enters the body through food, it is hydrolysed by the intestinal microflora to a highly reactive free radical generating compound, the aglycone divicine. Upon hydrolysis, the glucose part of the molecule is split off and that results in the reduced divicine. Divicine is then taken up in the blood through the intestinal epithelium. Adverse effects Adverse effects almost solely occur in humans that suffer from glucose-6-phosphate dehydrogenase deficiency. This deficiency causes a shortage of glutathione in erythrocytes and glutathione is needed for the neutralization of ROS (reactive oxygen species) created by the strongly oxidizing agent divicine. Indications Persons with G6PD deficiency are asymptomatic. An attack of acute haemolytic anaemia can appear out of nowhere and can be very severe and life-threatening. Indications of such a sudden attack of f
https://en.wikipedia.org/wiki/Dan%20Grimaldi
Dan Grimaldi (born March 7, 1946) is an American actor and mathematics professor who is known for his roles as twins Philly and Patsy Parisi on the HBO television series The Sopranos, various characters on Law & Order (1991-2001), Don't Go in the House (1979), The Junkman (1983), Men of Respect (1990), and The Yards (2000). Education Grimaldi has a bachelor's degree in mathematics from Fordham University, a master's degree in operations research from New York University, and a PhD in data processing from the City University of New York, and teaches in the Department of Mathematics and Computer Science at Kingsborough Community College in Brooklyn, New York. Career In addition to his role on The Sopranos, he has also had some minor film credits, most notably as mother-fixated pyromaniac Donny Kohler in the 1980 slasher film Don't Go in the House, and some guest TV appearances, including several episodes on Law & Order as well as appearing in 2011 as Tommy Barrone Sr. in "Moonlighting", the 9th episode of the 2nd season of the CBS show Blue Bloods. He appeared as an executive in the 2000 film The Yards and Grimaldi also voices "Frank" for the video game Mafia. Filmography Film Television Video games
https://en.wikipedia.org/wiki/Malgrange%E2%80%93Ehrenpreis%20theorem
In mathematics, the Malgrange–Ehrenpreis theorem states that every non-zero linear differential operator with constant coefficients has a Green's function. It was first proved independently by and . This means that the differential equation where P is a polynomial in several variables and δ is the Dirac delta function, has a distributional solution u. It can be used to show that has a solution for any compactly supported distribution f. The solution is not unique in general. The analogue for differential operators whose coefficients are polynomials (rather than constants) is false: see Lewy's example. Proofs The original proofs of Malgrange and Ehrenpreis were non-constructive as they used the Hahn–Banach theorem. Since then several constructive proofs have been found. There is a very short proof using the Fourier transform and the Bernstein–Sato polynomial, as follows. By taking Fourier transforms the Malgrange–Ehrenpreis theorem is equivalent to the fact that every non-zero polynomial P has a distributional inverse. By replacing P by the product with its complex conjugate, one can also assume that P is non-negative. For non-negative polynomials P the existence of a distributional inverse follows from the existence of the Bernstein–Sato polynomial, which implies that Ps can be analytically continued as a meromorphic distribution-valued function of the complex variable s; the constant term of the Laurent expansion of Ps at s = −1 is then a distributional inverse of P. Other proofs, often giving better bounds on the growth of a solution, are given in , and . gives a detailed discussion of the regularity properties of the fundamental solutions. A short constructive proof was presented in : is a fundamental solution of P(∂), i.e., P(∂)E = δ, if Pm is the principal part of P, η ∈ Rn with Pm(η) ≠ 0, the real numbers λ0, ..., λm are pairwise different, and
https://en.wikipedia.org/wiki/List%20of%20botanists%20by%20author%20abbreviation%20%28D%29
A–C To find entries for A–C, use the table of contents above. D D.A.Cooke – David Alan Cooke (born 1949) D.A.Herb. – Desmond Andrew Herbert (1898–1976) Dahl – Anders Dahl (1751–1789) Dahlgren – Bror Eric Dahlgren (1877–1961) Dahlst. – Gustav Adolf Hugo Dahlstedt (1856–1934) D.A.Keith – David A. Keith (fl. 2001) Daléchamps – Jacques Daléchamps (also as Jacobus Dale Champius) (1513–1588) Dallim. – William Dallimore (1871–1959) Dalpé – Yolande Dalpé (born 1948) Dalström – Stig Dalström (fl. 1983) Daly – Douglas C. Daly (born 1953) Dalzell – Nicol Alexander Dalzell (1817–1877) Dalziel – (1872–1948) D'Amato – Giovanni Frederico D'Amato (born 1941) (also Federico) Da M.Li – Da Ming Li (fl. 2007) Dammer – Carl Lebrecht Udo Dammer (1860–1920) D.A.Morrison – David A. Morrison (born 1958) Dana – James Dwight Dana (1813–1895) Dandy – James Edgar Dandy (1903–1976) Danert – Siegfried Danert (1926–1973) Danesi – L. Danesi (1851–1915) Danet – Frédéric Danet (born 1968) Danguy – Paul Auguste Danguy (1862–1942) Daniell – William Freeman Daniell (1818–1865) Daniels – Francis Potter Daniels (1869–1947) Danielsen – Anders Danielsen (1919–2006) Danihelka – Jiří Danihelka (born 1968) Däniker – Albert Ulrich Däniker (1894–1957) Danin – (born 1939) Dans. – Pierre Mackay Dansereau (1911–2011) Danser – Benedictus Hubertus Danser (1891–1943) Danthoine – D. (Étienne) Danthoine (1739–1794) Danzé-Corsin – Paule Danzé-Corsin (fl. 1956) Darby – John M. Darby (1804–1877) Darbysh. – S.J. Darbyshire (born 1953) Darl. – William Darlington (1782–1863) Darling – Samuel Taylor Darling (1872–1925) D.Arora – David Arora (born 1957) Darwin – Charles Darwin (1809–1882) Das – Atulananda Das (1879–1952) D.A.Simpson – David Alan Simpson (born 1955) D.A.Sutton – David A. Sutton (born 1952) Daubs – Edwin Horace Daubs (fl. 1965) Dauncey – Elizabeth Anne Dauncey (born 1965) Daveau – Jules Alexandre Daveau (1852–1929) Davenp. – George Edward Davenport (1833–19
https://en.wikipedia.org/wiki/North%20Carolina%20statistical%20areas
The U.S. currently has 44 statistical areas that have been delineated by the Office of Management and Budget (OMB). On July 21, 2023, the OMB delineated 7 combined statistical areas, 16 metropolitan statistical areas, and 21 micropolitan statistical areas in . Statistical areas The Office of Management and Budget (OMB) has designated more than 1,000 statistical areas for the United States and Puerto Rico. These statistical areas are important geographic delineations of population clusters used by the OMB, the United States Census Bureau, planning organizations, and federal, state, and local government entities. The OMB defines a core-based statistical area (commonly referred to as a CBSA) as "a statistical geographic entity consisting of the county or counties (or county-equivalents) associated with at least one core of at least 10,000 population, plus adjacent counties having a high degree of social and economic integration with the core as measured through commuting ties with the counties containing the core." The OMB further divides core-based statistical areas into metropolitan statistical areas (MSAs) that have "a population of at least 50,000" and micropolitan statistical areas (μSAs) that have "a population of at least 10,000, but less than 50,000." The OMB defines a combined statistical area (CSA) as "a geographic entity consisting of two or more adjacent core-based statistical areas with employment interchange measures of at least 15%." The primary statistical areas (PSAs) include all combined statistical areas and any core-based statistical area that is not a constituent of a combined statistical area. Table The table below describes the 49 United States statistical areas and 100 counties of the State of North Carolina with the following information: The combined statistical area (CSA) as designated by the OMB as of 2023. The CSA population according to 2023 US Census Bureau population estimates. The core based statistical area (CBSA) as designated b
https://en.wikipedia.org/wiki/Dimensional%20transmutation
In particle physics, dimensional transmutation is a physical mechanism providing a linkage between a dimensionless parameter and a dimensionful parameter. In classical field theory, such as gauge theory in four-dimensional spacetime, the coupling constant is a dimensionless constant. However, upon quantization, logarithmic divergences in one-loop diagrams of perturbation theory imply that this "constant" actually depends on the typical energy scale of the processes under considerations, called the renormalization group (RG) scale. This "running" of the coupling is specified by the beta-function of the renormalization group. Consequently, the interaction may be characterised by a dimensionful parameter , namely the value of the RG scale at which the coupling constant diverges. In the case of quantum chromodynamics, this energy scale is called the QCD scale, and its value 220 MeV supplants the role of the original dimensionless coupling constant in the form of the logarithm (at one-loop) of the ratio and . Perturbation theory, which produced this type of running formula, is only valid for a (dimensionless) coupling ≪ 1. In the case of QCD, the energy scale is an infrared cutoff, such that implies , with the RG scale. On the other hand, in the case of theories such as QED, is an ultraviolet cutoff, such that implies . This is also a way of saying that the conformal symmetry of the classical theory is anomalously broken upon quantization, thereby setting up a mass scale. See conformal anomaly. Quantum field theory Renormalization group
https://en.wikipedia.org/wiki/Vein%20of%20Galen%20aneurysmal%20malformations
Vein of Galen aneurysmal malformations (VGAMs) and Vein of Galen aneurysmal dilations (VGADs) are the most frequent arteriovenous malformations in infants and fetuses. A VGAM consists of a tangled mass of dilated vessels supplied by an enlarged artery. The malformation increases greatly in size with age, although the mechanism of the increase is unknown. Dilation of the great cerebral vein of Galen is a secondary result of the force of arterial blood either directly from an artery via an arteriovenous fistula or by way of a tributary vein that receives the blood directly from an artery. There is usually a venous anomaly downstream from the draining vein that, together with the high blood flow into the great cerebral vein of Galen causes its dilation. The right sided cardiac chambers and pulmonary arteries also develop mild to severe dilation. Signs and symptoms Malformations often lead to cardiac failure, cranial bruits (pattern 1), hydrocephaly, and subarachnoid hemorrhage in neonates. The heart failure is due to the size of the arteriovenous shunt that can steal 80% or more of the cardiac output, with large volumes of blood under high pressure returning to the right heart and pulmonary circulation and sinus venosus atrial septal defects. It is also the most common cause of death in such patients. Associated conditions Non-developmental syndromes also directly or indirectly affect the Great Cerebral Vein of Galen, although they are extremely rare. These include superior vena cava syndrome (SVCS), and thrombosis of the lateral sinus, superior sagittal sinus, internal jugular vein, or of the Great Cerebral Vein of Galen itself. Genetics 10% of vein of Galen aneurysmal malformations are associated with deleterious heterozygous mutations of EPHB4 Another study found that 30% of cases were associated with mutations in EPH receptor B4 (EPHB4) gene. Diagnosis Testing for a malformed vein of Galen is indicated when a patient has heart failure which has no obvious c
https://en.wikipedia.org/wiki/Maximum%20residue%20limit
The maximum residue limit (also maximum residue level, MRL) is the maximum amount of pesticide residue that is expected to remain on food products when a pesticide is used according to label directions, that will not be a concern to human health. Determination The MRL is usually determined by repeated (on the order of 10) field trials, where the crop has been treated according to good agricultural practice (GAP) and an appropriate pre harvest interval or withholding period has elapsed. For many pesticides this is set at the limit of determination (LOD) – since only major pesticides have been evaluated and understanding of acceptable daily intake (ADI) is incomplete (i.e. producers or public bodies have not submitted MRL data – often because these were not required in the past). LOD can be considered a measure of presence/absence, but certain residues may not be quantifiable at very low levels. For this reason the limit of quantification (LOQ) is often used instead of the LOD. As a rule of thumb the LOQ is approximately two times the LOD. For substances that are not included in any of the annexes in EU regulations, a default MRL of 0.01 mg/kg normally applies. It follows that adoption of GAP at the farm level must be a priority, and includes the withdrawal of obsolete pesticides. With increasingly sensitive detection equipment, a certain amount of pesticide residue will often be measured following field use. In the current regulatory environment, it would be wise for cocoa producers to focus only on pest control agents that are permitted for use in the EU and US. It should be stressed that MRLs are set on the basis of observations and not on ADIs. In medicinal plants If MRL of some medicinal plant is not known it is calculated by the formula: where SF is the safety factor MDI is the mean daily intake W is the body weight ADI is the acceptable daily intake Ornamental crops In some cases in the EU MRL's are also used for ornamental produce, and checked again
https://en.wikipedia.org/wiki/Sashimono
Sashimono (指物, 差物, 挿物) were small banners historically worn by soldiers in feudal Japan, for identification during battles. Description Sashimono poles were attached to the backs of the dō "cuirass" by special fittings. Sashimono were worn both by foot soldiers, including the common soldiers known as ashigaru, as well as by the elite samurai and members of the shogunate, and in special holders on the horses of some cavalry. The banners, resembling small flags and bearing clan symbols, were most prominent during the Sengoku period, a long period of civil war in Japan from the middle 15th to early 17th century. Variety Given the great variety in Japanese armour, sashimono were used to provide a kind of "uniform" to armies. Sashimono typically came in either square or short rectangular forms, although many variations existed. A variation that is often bigger and coloured is the uma-jirushi, which were large, personalized, sashimono-like flags worn by commanders. Similar to this were the very large and narrow nobori banners, which commonly took two or three men to hold erect and were used to control the direction of fighting during large battles. (Uma-jirushi and nobori are still used today at sports events, as Japanese versions of the banners common among Western sports audiences.) The banner hung from an L-shaped frame, which was attached to the chest armour dō or dou by a socket machi-uke or uketsubo near the waistline and hinged at shoulder level with a ring gattari or sashimono-gane. While this arrangement was perhaps one of the most common, there were other variations. Silk and leather were the most common materials used. Design The designs on sashimono were usually very simple geometric shapes, sometimes accompanied by Japanese characters providing the name of the leader or clan, the clan's mon, or a clan's slogan. Often, the background colour of the flag indicated which army unit the wearer belonged to, while different divisions in these armies emblazoned th
https://en.wikipedia.org/wiki/Uridine%20diphosphate%20N-acetylgalactosamine
Uridine diphosphate N-acetylgalactosamine or UDP-GalNAc is a nucleotide sugar composed of uridine diphosphate (UDP) and N-acetyl galactosamine (GalNAc). It is used by glycosyltransferases to transfer N-acetylgalactosamine residues to substrates. UDP-GalNAc is an important building block for the production of glycoproteins and glycolipids in the body. It also serves as a precursor for the synthesis of mucin-type O-glycans, which are important components of mucus and play important roles in biological processes such as cell signaling, immune defense, and lubrication of the digestive tract. See also Galactosamine Globoside N-Acetylgalactosamine (N-Acetylglucosamine) GlcNAc
https://en.wikipedia.org/wiki/Volume%20solid
Volume solid is the volume of paint after it has dried. This is different than the weight solid. Paint may contain solvent, resin, pigments, and additives. Many paints do not contain any solvent. After applying the paint, the solid portion will be left on the substrate. Volume solid is the term that indicates the solid proportion of the paint on a volume basis. For example, if the paint is applied in a wet film at a 100 μm thickness and the volume solid of paint is 50%, then the dry film thickness (DFT) will be 50 μm as 50% of the wet paint has evaporated. Suppose the volume solid is 100%, and the wet film thickness is also 100 μm. Then after complete drying of the paint, the DFT will be 100 μm because no solvent will be evaporated. This is an important concept when using paint industrially to calculate the cost of painting. It can be said that it is the real volume of paint. Here is the formula by which one can calculate the volume solid of paint, (Total sum by volume of each solid ingredient in paint x 100%)/ Total sum by volume of each ingredient in paint. A simple method that anyone can do to determine volume solids empirically is to apply paint to a steel surface with an application knife and measure the wet film thickness. Then cure the paint and measure the dry film thickness. The percentage of dry to wet represents the percentage of volume solids. In earlier days, the volume solid was measured by a disc method but now a sophisticated instrument is also available which takes only a drop of paint to check the volume solid. Understanding 'volume solids' allows knowing the true cost of different coatings and how much paint is used to perform its function. Generally, more expensive paints have a higher volume of solids and provide better coverage.
https://en.wikipedia.org/wiki/PyMOL
PyMOL is an open source but proprietary molecular visualization system created by Warren Lyford DeLano. It was commercialized initially by DeLano Scientific LLC, which was a private software company dedicated to creating useful tools that become universally accessible to scientific and educational communities. It is currently commercialized by Schrödinger, Inc. As the original software license was a permissive licence, they were able to remove it; new versions are no longer released under the Python license, but under a custom license (granting broad use, redistribution, and modification rights, but assigning copyright to any version to Schrodinger, LLC.), and some of the source code is no longer released. PyMOL can produce high-quality 3D images of small molecules and biological macromolecules, such as proteins. According to the original author, by 2009, almost a quarter of all published images of 3D protein structures in the scientific literature were made using PyMOL. PyMOL is one of the few mostly open-source model visualization tools available for use in structural biology. The Py part of the software's name refers to the program having been written in the programming language Python. PyMOL uses OpenGL Extension Wrangler Library (GLEW) and FreeGLUT, and can solve Poisson–Boltzmann equations using the Adaptive Poisson Boltzmann Solver. PyMOL used Tk for the GUI widgets and had native Aqua binaries for macOS through Schrödinger, which were replaced with a PyQt user interface on all platforms with the release of version 2.0. History and commercialization Early versions of PyMol were released under the Python License. On 1 August 2006, DeLano Scientific adopted a controlled-access download system for precompiled PyMOL builds (including betas) distributed by the company. Access to these executables is now limited to registered users who are paying customers; educational builds are available free to students and teachers. However, most of the current source code
https://en.wikipedia.org/wiki/Aluminium%20nitride
Aluminium nitride (AlN) is a solid nitride of aluminium. It has a high thermal conductivity of up to 321 W/(m·K) and is an electrical insulator. Its wurtzite phase (w-AlN) has a band gap of ~6 eV at room temperature and has a potential application in optoelectronics operating at deep ultraviolet frequencies. History and physical properties AlN was first synthesized in 1862 by F. Briegleb and A. Geuther. AlN, in the pure (undoped) state has an electrical conductivity of 10−11–10−13 Ω−1⋅cm−1, rising to 10−5–10−6 Ω−1⋅cm−1 when doped. Electrical breakdown occurs at a field of 1.2–1.8 V/mm (dielectric strength). The material exists primarily in the hexagonal wurtzite crystal structure, but also has a metastable cubic zincblende phase, which is synthesized primarily in the form of thin films. It is predicted that the cubic phase of AlN (zb-AlN) can exhibit superconductivity at high pressures. In AlN wurtzite crystal structure, Al and N alternate along the c-axis, and each bond is tetrahedrally coordinated with four atoms per unit cell. One of the unique intrinsic properties of wurtzite AlN is its spontaneous polarization. The origin of spontaneous polarization is the strong ionic character of chemical bonds in wurtzite AlN due to the large difference in electronegativity between aluminium and nitrogen atoms. Furthermore, the non-centrosymmetric wurtzite crystal structure gives rise to a net polarization along the c-axis. Compared with other III-nitride materials, AlN has a larger spontaneous polarization due to the higher nonideality of its crystal structure (Psp: AlN 0.081 C/m2 > InN 0.032 C/m2 > GaN 0.029 C/m2). Moreover, the piezoelectric nature of AlN gives rise to internal piezoelectric polarization charges under strain. These polarization effects can be utilized to induce a high density of free carriers at III-nitride semiconductor heterostructure interfaces completely dispensing with the need of intentional doping. Owing to the broken inversion symmetry along
https://en.wikipedia.org/wiki/Leaf%20window
Leaf window, also known as epidermal window, and fenestration, is a specialized leaf structure consisting of a translucent area through which light can enter the interior surfaces of the leaf where photosynthesis can occur. The translucent structure may include epidermal tissue, and in some succulent plants it consists of several cell layers of parenchyma, which may also function as water-storage tissue. It can appear as a large continuous patch, a variegated or reticulated region, or as numerous small spots. It is found in some succulent plants native to arid climates, allowing much of the plant to remain beneath the soil surface where it is protected from desiccation by winds and heat while optimizing light absorption. Many species featuring leaf windows are native to Southern Africa. Functions The primary function of the translucent windows is to increase the absorption of radiant energy, and thereby the rate of photosynthesis. Epidermal windows are commonly situated at the apex of leaves, allowing light to be captured and utilized even when the plant is almost entirely below the soil surface, minimizing exposure of leaf surface area to desiccation by intense heat. The windows focus and channel diffused light to the green photosynthesizing surface hidden inside the underground part of the plant. An absence of stomata in the translucent tissue prevents water loss. It has been found in the Lithops species of succulents that leaf window size correlates inversely with habitat solar irradiance; plants growing in regions of high irradiance and low rainfall have smaller windows than those growing in cloudy, high-rainfall regions. The carnivorous cobra lily (Darlingtonia californica) uses fenestration to confuse insects and increase the effectiveness of its trap. Species Genera of plants with numerous species having translucent epidermal windows include: Bulbine spp. (B. haworthioides, B. mesembryanthemoides, etc.) Conophytum spp. Callisia spp. (C. navicularis, etc.)
https://en.wikipedia.org/wiki/Pregnanediol
Pregnanediol, or 5β-pregnane-3α,20α-diol, is an inactive metabolic product of progesterone. A test can be done to measure the amount of pregnanediol in urine, which offers an indirect way to measure progesterone levels in the body. From the urine of pregnant women from London clinics, Guy Frederic Marrian isolated a substance that contained two hydroxyl groups and could be converted into a diacetate with acetic anhydride. However, the formula had not been clearly clarified. Almost at the same time, Adolf Butenandt at the Chemical University Laboratory in Göttingen investigated the constituents of pregnant urine and clarified the structure of the diol. The name pregnandiol, coined by Butenandt, is derived from the Latin verb praegnans (pregnant) or the English pregnant and pregnancy. This gave rise to the name pregnane for the underlying parent hydrocarbon. In 1936, Venning and Browne demonstrated the presence of pregnanediol, specifically the glucuronide of pregnanediol in pregnancy urine. Their study extracted pregnanediol from pregnancy urine and revealed that pregnanediol concentration in urine indicates the amount of progesterone excreted. Since progesterone levels indicate the functionality of a corpus luteum, and pregnanediol concentration represents 40-45% of the progesterone excreted, estimations of pregnanediol reveal the functionality of a corpus luteum. However, pregnanediol concentrations vary with menstrual cycle phases, so it is essential to consider the menstrual cycle phase when examining them. Furthermore, current research has demonstrated that pregnanediol concentration in urine is also a measure of ovarian activity. Chemistry See also Pregnanedione Progesterone (pregn-4-ene-3,20-dione)
https://en.wikipedia.org/wiki/FreeOTP
FreeOTP is a free and open-source software token that can be used for two-factor authentication. It provides implementations of HOTP and TOTP. Tokens can be added by scanning a QR code or by manually entering the token configuration. It is maintained by Red Hat under the Apache 2.0 license, and supports Android and iOS. FreeOTP Plus (aka FreeOTP+) is a fork of FreeOTP with enhancements including exporting and importing settings. See also Google Authenticator LinOTP Security token Comparison of TOTP applications
https://en.wikipedia.org/wiki/Stem%20cell%20research%20policy
Stem cell research policy varies significantly throughout the world. There are overlapping jurisdictions of international organizations, nations, and states or provinces. Some government policies determine what is allowed versus prohibited, whereas others outline what research can be publicly financed. Of course, all practices not prohibited are implicitly permitted. Some organizations have issued recommended guidelines for how stem cell research is to be conducted. International bodies The United Nations adopted a declaration on human cloning that can be interpreted as calling on member states to prohibit somatic cell nuclear transfer, or therapeutic cloning. In 2005, in a divided vote, "Member States were called on to adopt all measures necessary to prohibit all forms of human cloning in as much as they are incompatible with human dignity and the protection of human life." The World Health Organization has opposed a ban on cloning techniques in stem cell research. The Council of Europe's Convention on Human Rights and Biomedicine seems to ban the creation of embryos solely for research purposes. It has been signed by 31 countries and ratified by 19: Bulgaria, Croatia, Cyprus, the Czech Republic, Denmark, Estonia, Georgia, Greece, Hungary, Iceland, Lithuania, Moldova, Portugal, Romania, San Marino, Slovakia, Slovenia, Spain, and Turkey. The Hinxton Group Researchers, ethicists and assorted spokespersons from 14 different countries have published a set of legal and ethical guidelines relating to stem cell research, in an effort to address conflicting international laws in this area. The ‘Hinxton Group’ met recently for the first time, in Cambridge, and published a consensus statement calling for a ‘flexible’ regulatory framework, which can simultaneously accommodate rapid scientific advance and at the same time accommodate the diversity of international approaches towards stem cell science. It also recommends that, in countries which oppose embryonic stem cell
https://en.wikipedia.org/wiki/Opto-electronic%20oscillator
In optoelectronics, an opto-electronic oscillator (OEO) is a circuit that produces a repetitive electronic sine wave and/or modulated optical continuous wave signals. An opto-electronic oscillator is based on converting the continuous light energy from a pump laser to radio frequency (RF), microwave or mm-wave signals. The OEO is characterized by having very high quality factor (Q) and stability, as well as other functional characteristics that are not readily achieved with electronic oscillators. Its unique behavior results from the use of electro-optical (E/O) and photonic components, which are generally characterized with high efficiency, high speed, and low dispersion in the microwave frequency regime. In an OEO, the phase noise of the oscillator does not increase with the frequency that is subject to other implementations by electronic oscillators such as quartz crystal oscillators, dielectric resonators, sapphire resonators or air-dielectric resonators. History OEO was introduced in the early 1990s. Since then, the device’s key properties have been continuously improved. Operation Most OEOs utilize the transmission characteristics of an optical modulator together with a fiber-optic delay line to convert light energy into stable, spectrally pure RF/microwave reference signals. Light from a laser is introduced into an electro-optic (E/O) modulator, the output of which is passed through a long optical fiber and detected with a photodetector. The output of the photodetector is amplified and filtered and fed back to the electric port of the modulator. This configuration supports self-sustained oscillations, at a frequency determined by the fiber delay length, the bias setting of the modulator, and the band pass characteristics of the filter. It also provides for both electric and optical outputs. The conditions for self-sustained oscillations include coherent addition of partial waves each way around the loop and a loop gain exceeding losses for the circul
https://en.wikipedia.org/wiki/Double%20extension%20set%20theory
In mathematics, the Double extension set theory (DEST) is an axiomatic set theory proposed by Andrzej Kisielewicz consisting of two separate membership relations on the universe of sets, denoted here by and , and a set of axioms relating the two. The intention behind defining the two membership relations is to avoid the usual paradoxes of set theory, without substantially weakening the axiom of unrestricted comprehension. Intuitively, in DEST, comprehension is used to define the elements of a set under one membership relation using formulas that involve only the other membership relation. Let be a first-order formula with free variable in the language of DEST not involving the membership relation . Then, the axioms of DEST posit a set such that . For instance, is a formula involving only , and thus DEST posits the Russell set , where . Observe that for , we obtain . Since the membership relations are different, we thus avoid the Russell's paradox. The focus in DEST is on regular sets, which are sets whose extensions under the two membership relations coincide, i.e., sets for which it holds that . The preceding discussion suggests that the Russell set cannot be regular, as otherwise it leads to the Russell's paradox.
https://en.wikipedia.org/wiki/Poly-MVA
Poly-MVA (or Lipoic Acid Mineral Complex) is a dietary supplement created by Merrill Garnett (1931–), a former dentist turned biochemist. Poly-MVA is an ineffective alternative cancer treatment. Description The "MVA" in "Poly-MVA" means "minerals vitamins and amino acids". Poly-MVA contains lipoic acid, acetylcysteine, palladium, B vitamins, and other ingredients. The substance is red-brown liquid that is taken by mouth. In 2004, a year's supply of Poly-MVA was reported as costing US$19,800. As of 2019, the cost appears to fluctuate according to an individual's situation and dosage. Alternative medicine Poly-MVA is promoted with claims that it can treat a variety of human diseases, including cancer and HIV/AIDS. The promotional effort is supported by customer testimonials, but there is no medical evidence that Poly-MVA confers any health benefit and some concern it may inhibit the effectiveness of mainstream cancer treatments if used at the same time. In 2005, Poly-MVA was listed as one of the ineffective alternative cancer treatments being sold by the clinics clustered in and around Tijuana, Mexico. None of the information referenced in this review is specific to Poly-MVA. See also Antioxidant List of unproven and disproven cancer treatments
https://en.wikipedia.org/wiki/Nested%20dissection
In numerical analysis, nested dissection is a divide and conquer heuristic for the solution of sparse symmetric systems of linear equations based on graph partitioning. Nested dissection was introduced by ; the name was suggested by Garrett Birkhoff. Nested dissection consists of the following steps: Form an undirected graph in which the vertices represent rows and columns of the system of linear equations, and an edge represents a nonzero entry in the sparse matrix representing the system. Recursively partition the graph into subgraphs using separators, small subsets of vertices the removal of which allows the graph to be partitioned into subgraphs with at most a constant fraction of the number of vertices. Perform Cholesky decomposition (a variant of Gaussian elimination for symmetric matrices), ordering the elimination of the variables by the recursive structure of the partition: each of the two subgraphs formed by removing the separator is eliminated first, and then the separator vertices are eliminated. As a consequence of this algorithm, the fill-in (the set of nonzero matrix entries created in the Cholesky decomposition that are not part of the input matrix structure) is limited to at most the square of the separator size at each level of the recursive partition. In particular, for planar graphs (frequently arising in the solution of sparse linear systems derived from two-dimensional finite element method meshes) the resulting matrix has O(n log n) nonzeros, due to the planar separator theorem guaranteeing separators of size O(). For arbitrary graphs there is a nested dissection that guarantees fill-in within a factor of optimal, where d is the maximum degree and m is the number of non-zeros. See also Cycle rank of a graph, or a symmetric Boolean matrix, measures the minimum parallel time needed to perform Cholesky decomposition Vertex separator Notes
https://en.wikipedia.org/wiki/List%20of%20Unix%20daemons
This is a list of Unix daemons that are found on various Unix-like operating systems. Unix daemons typically have a name ending with a d. See also List of Unix commands
https://en.wikipedia.org/wiki/Psammon
Psammon (from Greek "psammos", "sand") is a group of organisms inhabiting coastal sand moist — biota buried in sediments. Psammon is a part of water fauna, along with periphyton, plankton, nekton, and benthos. Psammon is also sometimes considered a part of benthos due to its near-bottom distribution. Psammon term is commonly used to refer to freshwater reservoirs such as lakes.
https://en.wikipedia.org/wiki/Symphiles
Symphiles are insects or other organisms which live as welcome guests in the nest of a social insect (such as the ant, myrmecophily, or termite, termitophily) by which they are fed and guarded. The relationship between the symphile and host may be symbiotic, inquiline or parasitic. Symphile taxa This is a selection of taxa exhibiting symphilia, not a complete list. Fibularhizoctonia Fibularhizoctonia, sometimes referred to as cuckoo fungus due to their adaptation to mimic termite eggs, employ chemical and morphological mimicry to benefit from the defense termites provide their brood. If termite workers are present to care for a brood which contains cuckoo fungus, the sclerotia, or "termite balls", are unlikely to germinate and their presence will increase the survival rate of the termite eggs. When worker termites were experimentally removed from brood that contained slerotia, the fungus germinated by exploiting the termite eggs. This means the termitophilic relationship between termites and Fibularhizoctonia can be parasitic or mutualistic. Phengaris arion The large blue butterfly, Phengaris arion (formerly Maculinea arion), exhibits a unique parasitic relationship with a single species of red ant, Myrmica sabuleti. Psithyrus Cuckoo bumblebees, members of the subgenus Psithyrus in the genus Bombus, are obligate brood parasites; they must use colonies of true bumblebees to rear their young. A Psithyrus female will kill or subdue the host colony's queen and then use pheromones and/or physical attacks to force the host colony to feed her and raise her brood. Staphylinidae Many species of Staphylinidae (commonly known as “Rove Beetles”) have developed complex interspecies relationships with ants. Ant associations range from near free-living species which prey only on ants, to obligate inquilines of ants, which exhibit extreme morphological and chemical adaptations to the harsh environments of ant nests. Some species are fully integrated into the host colony, and a
https://en.wikipedia.org/wiki/Cottonseed%20oil
Cottonseed oil is cooking oil from the seeds of cotton plants of various species, mainly Gossypium hirsutum and Gossypium herbaceum, that are grown for cotton fiber, animal feed, and oil. Cotton seed has a similar structure to other oilseeds such as sunflower seed, having an oil-bearing kernel surrounded by a hard outer hull; in processing, the oil is extracted from the kernel. Cottonseed oil is used for salad oil, mayonnaise, salad dressing, and similar products because of its flavor stability. Composition Its fatty acid profile generally consists of 70% unsaturated fatty acids (18% monounsaturated, and 52% polyunsaturated), 26% saturated fatty acids. When it is fully hydrogenated, its profile is 94% saturated fat and 2% unsaturated fatty acids (1.5% monounsaturated, and 0.5% polyunsaturated). According to the National Cottonseed Products Association, cottonseed oil does not need to be hydrogenated as much as other polyunsaturated oils to achieve similar results. Gossypol is a toxic, yellow, polyphenolic compound produced by cotton and other members of the order Malvaceae, such as okra. This naturally occurring colored compound is found in tiny glands in the seed, leaf, stem, tap root bark, and root of the cotton plant. The adaptive function of the compound facilitates natural insect resistance. The three key steps of refining, bleaching, and deodorization in producing finished oil act to eliminate the gossypol level. Ferric chloride is often used to decolorize cotton seed oil. Comparison to other vegetable oils Physical properties Once processed, cottonseed oil has a mild taste and appears generally clear with a light golden color, the amount of color depending on the amount of refining. It has a relatively high smoke point as a frying medium. Density ranges from . Like other long-chain fatty acid oils, cottonseed oil has a smoke point of about , and is high in tocopherols, which also contribute its stability, giving products that contain it a long shelf lif
https://en.wikipedia.org/wiki/Supervisory%20attentional%20system
Executive functions are a cognitive apparatus that controls and manages cognitive processes. Norman and Shallice (1980) proposed a model on executive functioning of attentional control that specifies how thought and action schemata become activated or suppressed for routine and non-routine circumstances. Schemas, or scripts, specify an individual's series of actions or thoughts under the influence of environmental conditions. Every stimulus condition turns on the activation of a response or schema. The initiation of appropriate schema under routine, well-learned situations is monitored by contention scheduling which laterally inhibits competing schemas for the control of cognitive apparatus. Under unique, non-routine procedures controls schema activation. The SAS is an executive monitoring system that oversees and controls contention scheduling by influencing schema activation probabilities and allowing for general strategies to be applied to novel problems or situations during automatic attentional processes. Basic background Executive functions Executive functions are cognitive processes that control other brain activities and are predominantly functioning in the prefrontal areas of the frontal lobe. Executive functions are limited in capacity and accountable for the initiation, consolidation, regulation, and inhibition of cognitive, language, motor and emotional processes. These processes underlie such functions as self-evaluation, planning, problem solving, controlling impulses and attention, and strategic selection or sequencing of behaviour to reach desired goals. Measuring executive functions is often less accurate than measuring non-executive tasks because of the interconnectedness and multi-determined complexity of the brain. Executive functions are hard to measure independently of all other cognitive functions and are often influenced by non-executive factors. Consequently, understanding the relationship between behaviours and cognitive processes c
https://en.wikipedia.org/wiki/Inoculation
Inoculation is the act of implanting a pathogen or other microbe or virus into a person or other organism. It is a method of artificially inducing immunity against various infectious diseases. The term "inoculation" is also used more generally to refer to intentionally depositing microbes into any growth medium, as into a Petri dish used to culture the microbe, or into food ingredients for making cultured foods such as yoghurt and fermented beverages such as beer and wine. This article is primarily about the use of inoculation for producing immunity against infection. Inoculation has been used to eradicate smallpox and to markedly reduce other infectious diseases such as polio. Although the terms "inoculation", "vaccination", and "immunization" are often used interchangeably, there are important differences. Inoculation is the act of implanting a pathogen or microbe into a person or other recipient; vaccination is the act of implanting or giving someone a vaccine specifically; and immunization is the development of disease resistance that results from the immune system's response to a vaccine or natural infection. Terminology Until the early 1800s inoculation referred only to variolation (from the Latin word variola = smallpox), the predecessor to the smallpox vaccine. The smallpox vaccine, introduced by Edward Jenner in 1796, was called cowpox inoculation or vaccine inoculation (from Latin vacca = cow). Smallpox inoculation continued to be called variolation, whereas cowpox inoculation was called vaccination (from Jenner's term variolae vaccinae = smallpox of the cow). Louis Pasteur proposed in 1861 to extend the terms vaccine and vaccination to include the new protective procedures being developed. Immunization refers to the use of vaccines as well as the use of antitoxin, which contains pre-formed antibodies such as to diphtheria or tetanus exotoxins. In nontechnical usage inoculation is now more or less synonymous with protective injections and other methods of
https://en.wikipedia.org/wiki/Euclidean%20random%20matrix
Within mathematics, an N×N Euclidean random matrix  is defined with the help of an arbitrary deterministic function f(r, r′) and of N points {ri} randomly distributed in a region V of d-dimensional Euclidean space. The element Aij of the matrix is equal to f(ri, rj): Aij = f(ri, rj). History Euclidean random matrices were first introduced in 1999. They studied a special case of functions f that depend only on the distances between the pairs of points: f(r, r′) = f(r - r′) and imposed an additional condition on the diagonal elements Aii, Aij = f(ri - rj) - u δijΣkf(ri - rk), motivated by the physical context in which they studied the matrix. A Euclidean distance matrix is a particular example of Euclidean random matrix with either f(ri - rj) = |ri - rj|2 or f(ri - rj) = |ri - rj|. For example, in many biological networks, the strength of interaction between two nodes depends on the physical proximity of those nodes. Spatial interactions between nodes can be modelled as a Euclidean random matrix, if nodes are placed randomly in space. Properties Because the positions of the points {ri} are random, the matrix elements Aij are random too. Moreover, because the N×N elements are completely determined by only N points and, typically, one is interested in N≫d, strong correlations exist between different elements. Hermitian Euclidean random matrices Hermitian Euclidean random matrices appear in various physical contexts, including supercooled liquids, phonons in disordered systems, and waves in random media. Example 1: Consider the matrix  generated by the function f(r, r′) = sin(k0|r-r′|)/(k0|r-r′|), with k0 = 2π/λ0. This matrix is Hermitian and its eigenvalues Λ are real. For N points distributed randomly in a cube of side L and volume V = L3, one can show that the probability distribution of Λ is approximately given by the Marchenko-Pastur law, if the density of points ρ = N/V obeys ρλ03 ≤ 1 and 2.8N/(k0 L)2 < 1 (see figure). Non-Hermitian Euclidean random
https://en.wikipedia.org/wiki/Bredig%27s%20arc%20method
Bredig's arc method or electrical disintegration is a method of preparation of colloidal solution, of metals such as gold, silver or platinum. This method consists of both dispersion and condensation. An arc is struck between electrodes of the desired metal, under the surface of water containing some stabilizing agent such as traces of potassium hydroxide. The intense heat of the arc vaporizes some of the metal which then condenses under cold water. The water is kept cold with an ice bath. This method is not suitable when the dispersion medium is an organic liquid as considerable charring occurs.
https://en.wikipedia.org/wiki/Human%20Microbiome%20Project
The Human Microbiome Project (HMP) was a United States National Institutes of Health (NIH) research initiative to improve understanding of the microbiota involved in human health and disease. Launched in 2007, the first phase (HMP1) focused on identifying and characterizing human microbiota. The second phase, known as the Integrative Human Microbiome Project (iHMP) launched in 2014 with the aim of generating resources to characterize the microbiome and elucidating the roles of microbes in health and disease states. The program received $170 million in funding by the NIH Common Fund from 2007 to 2016. Important components of the HMP were culture-independent methods of microbial community characterization, such as metagenomics (which provides a broad genetic perspective on a single microbial community), as well as extensive whole genome sequencing (which provides a "deep" genetic perspective on certain aspects of a given microbial community, i.e. of individual bacterial species). The latter served as reference genomic sequences — 3000 such sequences of individual bacterial isolates are currently planned — for comparison purposes during subsequent metagenomic analysis. The project also financed deep sequencing of bacterial 16S rRNA sequences amplified by polymerase chain reaction from human subjects. Introduction Prior to the HMP launch, it was often reported in popular media and scientific literature that there are about 10 times as many microbial cells and 100 times as many microbial genes in the human body as there are human cells; this figure was based on estimates that the human microbiome includes around 100 trillion bacterial cells and an adult human typically has around 10 trillion human cells. In 2014 the American Academy of Microbiology published a FAQ that emphasized that the number of microbial cells and the number of human cells are both estimates, and noted that recent research had arrived at a new estimate of the number of human cells at around 37 tr
https://en.wikipedia.org/wiki/Nucleotidase
A nucleotidase is a hydrolytic enzyme that catalyzes the hydrolysis of a nucleotide into a nucleoside and a phosphate. A nucleotide + H2O = a nucleoside + phosphate For example, it converts adenosine monophosphate to adenosine, and guanosine monophosphate to guanosine. Nucleotidases have an important function in digestion in that they break down consumed nucleic acids. They can be divided into two categories, based upon the end that is hydrolyzed: : 5'-nucleotidase - NT5C, NT5C1A, NT5C1B, NT5C2, NT5C3 : 3'-nucleotidase - NT3 5'-Nucleotidases cleave off the phosphate from the 5' end of the sugar moiety. They can be classified into various kinds depending on their substrate preferences and subcellular localization. Membrane-bound 5'-nucleotidases display specificity toward adenosine monophosphates and are involved predominantly in the salvage of preformed nucleotides and in signal transduction cascades involving purinergic receptors. Soluble 5'-nucleotidases are all known to belong to the haloacid dehalogenase superfamily of enzymes, which are two domain proteins characterised by a modified Rossman fold as the core and variable cap or hood. The soluble forms are further subclassified based on the criterion mentioned above. mdN and cdN are mitochondrial and cytosolic 5'-3'-pyrimidine nucleotidases. cN-I is a cytosolic nucleotidase(cN) characterized by its affinity toward AMP as its substrate. cN-II is identified by its affinity toward either IMP or GMP or both. cN-III is a pyrimidine 5'-nucleotidase. A new class of nucleotidases called IMP-specific 5'-nucleotidase has been recently defined. 5'-Nucleotidases are involved in varied functions like cell–cell communication, nucleic acid repair, purine salvage pathway for the synthesis of nucleotides, signal transduction, membrane transport, etc.
https://en.wikipedia.org/wiki/UltraMon
UltraMon is a commercial application for Microsoft Windows users who use multiple displays. UltraMon is developed by Realtime Soft, a small software development company based in Bern, Switzerland. UltraMon currently contains the following features: Two additional title bar buttons for managing windows among the monitors Customizable button location A taskbar on each additional monitor that displays tasks on that monitor Pre-defined application window placement Display profiles for multiple pre-defined display settings Spannable wallpaper option Different wallpapers for different monitors Advanced multiple-monitor screensaver management Display mirroring (Forces to software rendering) Overcome Windows' limit of 10 displays UltraMon is distributed as trialware, requiring the user to purchase the software after a trial period (30 days). UltraMon 3.3.0 is available with full Windows XP, Vista, 7, and 8 support. See also Multi-monitor
https://en.wikipedia.org/wiki/National%20Institute%20of%20Neurosciences%20%26%20Hospital
National Institute of Neurosciences & Hospital (NINS) is a state run tertiary care institute in Bangladesh focused on neurological disorders. It is located at Agargaon in Dhaka. Bangladeshi neurologist Prof Quazi Deen Mohammad is the founding director of this institute. History The institute started its journey in September 2012.This Institute was inaugurated by Prime Minister Sheikh Hasina. Administration Director of NINS is the administrative chief of the institute. He is assisted by Joint Director, Deputy Director, Assistant Directors (Admin, Finance & Store). Administrative officers assist them in this task. Departments Department of Neurology Department of Neurosurgery Department of Pediatric Neurology Department of Pediatric Neurosurgery Department of Neurophysiology Department of Neurointervention Department of Neurorehabilitation (Physical Medicine) Department of Neuro Radiology (Radiology & Imaging) Department of Critical Care Medicine Department of Cardiology Department of Neuro-anesthesia Department of Laboratory Services - Department of Neuropathology - Department of Biochemistry - Department of Microbiology - Department of Transfusion Medicine Hospital facilities It's a 450-bed hospital. In its ten-storey building, it has six modern operation theaters, a 16-bed ICU, a 12-bed HDU, a six-bed Recovery Unit and a 12-bed post-operative rooms. The hospital is equipped with most modern radiology and imaging department. It has a 24x7 emergency unit. Hospital Services There are a lot of Medical Technologist ,So 24x7 diagnostic services (Blood, urine, ECG, CT Scan Etc) are available. Publications Journal of National Institute of Neurosciences Bangladesh is the official journal of NINS. It's launched in 2015 and publisbed biannually.
https://en.wikipedia.org/wiki/Sintran%20III
Sintran III is a real-time, multitasking, multi-user operating system used with Norsk Data minicomputers from 1974. Unlike its predecessors Sintran I and II, it was written entirely by Norsk Data, in Nord Programming Language (Nord PL, NPL), an intermediate language for Norsk Data computers. Overview Sintran was mainly a command-line interface based operating system, though there were several shells which could be installed to control the user environment more strictly, by far the most popular of which was USER-ENVIRONMENT. One of the clever features was to be able to abbreviate commands and file names between hyphens. For example, typing LIST-FILES would give users several prompts, including for print, paging etc. Users could override this using the following LI-FI ,,n, which would abbreviate the LIST-FILES command prompt and bypass any of the prompts. One could also refer to files in this way, for example, with PED H-W: which would refer to HELLO-WORLD:SYMB if this was the only file having H, any number of characters, a hyphen -, a W, any number of characters, and any file ending. This saved many keystrokes and would allow users a very nice learning experience, from complete and self-explanatory commands like LIST-ALL-FILES to L-A-F for an advanced user. (The hyphen key on Norwegian keyboards resides where the slash key does on U.S. ones.) Now that Sintran has mostly disappeared as an operating system, there are few references to it. However a job control or batch processing language was available named JEC, believed to be named Job Execution Controller, this could be used to set up batch jobs to compile COBOL programs, etc.
https://en.wikipedia.org/wiki/Chiral%20Lewis%20acid
Chiral Lewis acids (CLAs) are a type of Lewis acid catalyst. These acids affect the chirality of the substrate as they react with it. In such reactions, synthesis favors the formation of a specific enantiomer or diastereomer. The method is an enantioselective asymmetric synthesis reaction. Since they affect chirality, they produce optically active products from optically inactive or mixed starting materials. This type of preferential formation of one enantiomer or diastereomer over the other is formally known as asymmetric induction. In this kind of Lewis acid, the electron-accepting atom is typically a metal, such as indium, zinc, lithium, aluminium, titanium, or boron. The chiral-altering ligands employed for synthesizing these acids often have multiple Lewis basic sites (often a diol or a dinitrogen structure) that allow the formation of a ring structure involving the metal atom. Achiral Lewis acids have been used for decades to promote the synthesis of racemic mixtures in myriad different reactions. Since the 1960s, chemists have used Chiral Lewis acids to induce enantioselective reactions. This is useful when the desired product is a specific enantiomer, as is common in drug synthesis. Common reaction types include Diels–Alder reactions, the ene reaction, [2+2] cycloaddition reactions, hydrocyanation of aldehydes, and most notably, Sharpless epoxidations. Theory The enantioselectivity of CLAs derives from their ability to perturb the free energy barrier along with the reaction coordinate pathway that leads to either the R- or S- enantiomer. Ground state diastereomers and enantiomers are of equal energy in the ground state, and when reacted with an achiral Lewis acid, their diastereomeric intermediates, transition states, and products are also of equal energy. This leads to the production of racemic mixtures. However, when a CLA is used in the same reaction, the energetic barrier of formation of one diastereomer is less than that of another; the reaction is u
https://en.wikipedia.org/wiki/Mandibular%20fossa
The mandibular fossa, also known as the glenoid fossa in some dental literature, is the depression in the temporal bone that articulates with the mandible. Structure In the temporal bone, the mandibular fossa is bounded anteriorly by the articular tubercle and posteriorly by the tympanic portion of the temporal bone, which separates it from the external acoustic meatus. The fossa is divided into two parts by a narrow slit, the petrotympanic fissure (Glaserian fissure). It is concave in shape to receive the condyloid process of the mandible. Development The mandibular fossa develops from condylar cartilage. This may be stimulated by SOX9 or ALK2, as has been seen in mouse models. Function The condyloid process of the mandible articulates with the temporal bone of the skull at the mandibular fossa. Clinical significance Problems with morphogenesis during embryonic development can lead to the mandibular fossa not forming. This may be caused by mutations to SOX9 or ALK2. If the mandibular fossa is very shallow, this can cause problems with the strength of the temporomandibular joint. This can lead to easy subluxation of the joint and trismus (lock jaw). Deformation of the mandibular fossa, often part of temporomandibular dysplasia, causes similar problems in dogs. This may resolve spontaneously, or require surgery. History The mandibular fossa is also known as the glenoid fossa in some dental literature. Other animals The mandibular fossa is a feature of the skulls of various other animals, including dogs. See also Temporomandibular joint
https://en.wikipedia.org/wiki/Futurama%3A%20Bender%27s%20Game
Futurama: Bender's Game is a 2008 American direct-to-video adult animated science fantasy comedy film and the third of the four Futurama films that make up the show's fifth season. It was released on DVD and Blu-ray on November 4, 2008. According to the Beast with a Billion Backs DVD commentary, the film, which spoofs Dungeons & Dragons, was in production when Dungeons & Dragons creator, Gary Gygax, died. The film contains a post-credits tribute to Gygax in the form of a title card and a clip of him from the episode "Anthology of Interest I". Elements of J. R. R. Tolkien's The Lord of the Rings and George Lucas' Star Wars are also parodied. The title of the film is a pun on the book Ender's Game, by Orson Scott Card, though the Futurama film has "very little to do with the subject material" of the book. Conversely, the 1985 book also used "Bender" as a mocking pun for "Ender", but Matt Groening stated this is not the original inspiration for Bender's name. Plot Defying Professor Farnsworth, Leela borrows the Planet Express Ship to enter a demolition derby after insults from rednecks anger her. Bender finds Cubert and Dwight playing Dungeons & Dragons, but cannot join in since, as a robot, he has no imagination. Bender manages to imagine himself as a medieval knight named "Titanius Anglesmith, fancy man of Cornwood" and enters the game. He becomes lost in his fantasy and goes on a rampage, resulting in his commitment to the Hal Institute for Criminally Insane Robots. The crew learns that Mom, who controls the world's only dark matter mine, is restricting the supply to drive up profits. The Professor reveals that while working for Mom, he discovered how to turn dark matter into fuel. The process created two crystals, with Mom keeping one for herself and Farnsworth hiding the other, "anti-backwards" crystal. If the crystals are brought together, they will render dark matter useless. Farnsworth, Fry, and Leela fly to Mom's mine with the crystal to neutralize the dark
https://en.wikipedia.org/wiki/Television%20lines
Television lines (TVL) is a specification of an analog camera or monitor's horizontal image resolution. The TVL is one of the most important resolution measures in a video system. The TVL can be measured with the standard EIA 1956 resolution chart. Definition TVL is defined as the maximum number of alternating light and dark vertical lines that can be resolved per picture height. A resolution of 400 TVL means that 200 distinct dark vertical lines and 200 distinct white vertical lines can be counted over a horizontal span equal to the height of the picture. For example, on monitor with 400 TVL, 200 vertical dark lines can be counted over width on monitor (Note that the of monitor height is used rather than the of whole monitor width). TVL is an inherent quality of a camera or monitor, influenced by the visual bandwidth of the transmission system used. It should not be confused with the number of horizontal scanning lines of such systems, which e.g. 625 lines for the PAL system, 525 lines for the NTSC system. Limitations Since analog transmission of video is scan line-based, the same number of horizontal lines is always transmitted. However, several factors impede the ability to display fine detail within a line: The camera or other source of material. The storage and processing of the picture. The transmission of the TV signal e.g. broadcast by radio or by cable. The reception and reproduction of the picture on a TV set. See also Kell factor, which limits the vertical resolution in analog television, and both horizontal and vertical resolution in digital television, to a fraction of the number of scan lines or pixels
https://en.wikipedia.org/wiki/Global%20microbial%20identifier
The genomic epidemiological database for global identification of microorganisms or global microbial identifier is a platform for storing whole genome sequencing data of microorganisms, for the identification of relevant genes and for the comparison of genomes to detect and track-and-trace infectious disease outbreaks and emerging pathogens. The database holds two types of information: 1) genomic information of microorganisms, linked to, 2) metadata of those microorganism such as epidemiological details. The database includes all genera of microorganisms: bacteria, viruses, parasites and fungi. Technology For genotyping of microorganisms for medical diagnosis, or other purposes, scientists may use a wide variety of DNA profiling techniques, such as polymerase chain reaction, pulsed-field gel electrophoresis or multilocus sequence typing. A complication of this broad variety of techniques is the difficulty to standardize between techniques, laboratories and microorganisms, which may be overcome using the complete DNA code of the genome generated by whole genome sequencing. For straightforward diagnostic identification, the whole genome sequencing information of a microbiological sample is fed into a global genomic database and compared using BLAST procedures to the genomes already present in the database. In addition, whole genome sequencing data may be used to back calculate to the different pre-whole genome sequencing genotyping methods, so previous collected valuable information is not lost. For the global microbial identifier the genomic information is coupled to a wide spectrum of metadata about the specific microbial clone and includes important clinical and epidemiological information such as the global finding places, treatment options and antimicrobial resistance, making it a general microbiological identification tool. This makes personalized treatment of microbial disease possible as well as real-time tracing systems for global surveillance of infectiou
https://en.wikipedia.org/wiki/Decay%20chain
In nuclear science, the decay chain refers to a series of radioactive decays of different radioactive decay products as a sequential series of transformations. It is also known as a "radioactive cascade". The typical radioisotope does not decay directly to a stable state, but rather it decays to another radioisotope. Thus there is usually a series of decays until the atom has become a stable isotope, meaning that the nucleus of the atom has reached a stable state. Decay stages are referred to by their relationship to previous or subsequent stages. A parent isotope is one that undergoes decay to form a daughter isotope. One example of this is uranium (atomic number 92) decaying into thorium (atomic number 90). The daughter isotope may be stable or it may decay to form a daughter isotope of its own. The daughter of a daughter isotope is sometimes called a granddaughter isotope. Note that the parent isotope becomes the daughter isotope, unlike in the case of a biological parent and daughter. The time it takes for a single parent atom to decay to an atom of its daughter isotope can vary widely, not only between different parent-daughter pairs, but also randomly between identical pairings of parent and daughter isotopes. The decay of each single atom occurs spontaneously, and the decay of an initial population of identical atoms over time t, follows a decaying exponential distribution, e−λt, where λ is called a decay constant. One of the properties of an isotope is its half-life, the time by which half of an initial number of identical parent radioisotopes can be expected statistically to have decayed to their daughters, which is inversely related to λ. Half-lives have been determined in laboratories for many radioisotopes (or radionuclides). These can range from nearly instantaneous (less than 10−21 seconds) to more than 1019 years. The intermediate stages each emit the same amount of radioactivity as the original radioisotope (i.e., there is a one-to-one relationsh
https://en.wikipedia.org/wiki/Smart%20antenna
Smart antennas (also known as adaptive array antennas, digital antenna arrays, multiple antennas and, recently, MIMO) are antenna arrays with smart signal processing algorithms used to identify spatial signal signatures such as the direction of arrival (DOA) of the signal, and use them to calculate beamforming vectors which are used to track and locate the antenna beam on the mobile/target. Smart antennas should not be confused with reconfigurable antennas, which have similar capabilities but are single element antennas and not antenna arrays. Smart antenna techniques are used notably in acoustic signal processing, track and scan radar, radio astronomy and radio telescopes, and mostly in cellular systems like W-CDMA, UMTS, and LTE and 5G-NR. Smart antennas have many functions: DOA estimation, beamforming, interference nulling, and constant modulus preservation. Direction of arrival (DOA) estimation The smart antenna system estimates the direction of arrival of the signal, using techniques such as MUSIC (MUltiple SIgnal Classification), estimation of signal parameters via rotational invariance techniques (ESPRIT) algorithms, Matrix Pencil method or one of their derivatives. They involve finding a spatial spectrum of the antenna/sensor array, and calculating the DOA from the peaks of this spectrum. These calculations are computationally intensive. Matrix Pencil is very efficient in case of real time systems, and under the correlated sources. Beamforming Beamforming is the method used to create the radiation pattern of the antenna array by adding constructively the phases of the signals in the direction of the targets/mobiles desired, and nulling the pattern of the targets/mobiles that are undesired/interfering targets. This can be done with a simple Finite Impulse Response (FIR) tapped delay line filter. The weights of the FIR filter may also be changed adaptively, and used to provide optimal beamforming, in the sense that it reduces the Minimum Mean Square Er
https://en.wikipedia.org/wiki/Boolean%20hierarchy
The boolean hierarchy is the hierarchy of boolean combinations (intersection, union and complementation) of NP sets. Equivalently, the boolean hierarchy can be described as the class of boolean circuits over NP predicates. A collapse of the boolean hierarchy would imply a collapse of the polynomial hierarchy. Formal definition BH is defined as follows: BH1 is NP. BH2k is the class of languages which are the intersection of a language in BH2k-1 and a language in coNP. BH2k+1 is the class of languages which are the union of a language in BH2k and a language in NP. BH is the union of all the BHi classes. Derived classes DP (Difference Polynomial Time) is BH2. Equivalent definitions Defining the conjunction and the disjunction of classes as follows allows for more compact definitions. The conjunction of two classes contains the languages that are the intersection of a language of the first class and a language of the second class. Disjunction is defined in a similar way with the union in place of the intersection. C ∧ D = { A ∩ B | A ∈ C   B ∈ D } C ∨ D = { A ∪ B | A ∈ C   B ∈ D } According to this definition, DP = NP ∧ coNP. The other classes of the Boolean hierarchy can be defined as follows. The following equalities can be used as alternative definitions of the classes of the Boolean hierarchy: Alternatively, for every k ≥ 3: Hardness Hardness for classes of the Boolean hierarchy can be proved by showing a reduction from a number of instances of an arbitrary NP-complete problem A. In particular, given a sequence {x1, ... xm} of instances of A such that xi ∈ A implies xi-1 ∈ A, a reduction is required that produces an instance y such that y ∈ B if and only if the number of xi ∈ A is odd or even: BH2k-hardness is proved if and the number of xi ∈ A is odd BH2k+1-hardness is proved if and the number of xi ∈ A is even Such reductions work for every fixed . If such reductions exist for arbitrary , the problem is hard for PNP[O(log n)].
https://en.wikipedia.org/wiki/Stochastic%20grammar
A stochastic grammar (statistical grammar) is a grammar framework with a probabilistic notion of grammaticality: Stochastic context-free grammar Statistical parsing Data-oriented parsing Hidden Markov model Estimation theory The grammar is realized as a language model. Allowed sentences are stored in a database together with the frequency how common a sentence is. Statistical natural language processing uses stochastic, probabilistic and statistical methods, especially to resolve difficulties that arise because longer sentences are highly ambiguous when processed with realistic grammars, yielding thousands or millions of possible analyses. Methods for disambiguation often involve the use of corpora and Markov models. "A probabilistic model consists of a non-probabilistic model plus some numerical quantities; it is not true that probabilistic models are inherently simpler or less structural than non-probabilistic models." Examples A probabilistic method for rhyme detection is implemented by Hirjee & Brown in their study in 2013 to find internal and imperfect rhyme pairs in rap lyrics. The concept is adapted from a sequence alignment technique using BLOSUM (BLOcks SUbstitution Matrix). They were able to detect rhymes undetectable by non-probabilistic models. See also Colorless green ideas sleep furiously Computational linguistics L-system#Stochastic grammars Stochastic context-free grammar Statistical language acquisition
https://en.wikipedia.org/wiki/Heart%20in%20hand
The Heart in Hand or Heart-in-Hand is a symbol of a heart in an open palm, and is symbolic of charity, given from the heart. It is an easily recognizable symbol in the Northeastern United States and used by the Shakers as a pictorial reminder of the words of Mother Ann Lee, the founder of the Shaker sect, who promoted a simple life of hard work and spirituality, "Put your hands to work, and your hearts to God." The image is typical of the Shaker attitude, and also implies a loving welcome. A heart in hand has also been a symbol of the Independent Order of Odd Fellows, a fraternal organization derived from English Oddfellows orders of the mid-18th century. These commonly display three linked rings representing friendship, love, and truth. The symbol originated as the seal of the Protestant Reformer John Calvin, and thus appears in the official seal of Calvin College with Calvin's motto, "My heart I offer to you LORD, promptly and sincerely." The symbol is also frequently associated with Amish communities. The symbol was also used as a chop mark in the woodblock prints of Gustave Baumann (Germ./Amer. 1881-1971). See also Religious symbolism Notes Religious symbols Heart symbols
https://en.wikipedia.org/wiki/Optimal%20control
Optimal control theory is a branch of control theory that deals with finding a control for a dynamical system over a period of time such that an objective function is optimized. It has numerous applications in science, engineering and operations research. For example, the dynamical system might be a spacecraft with controls corresponding to rocket thrusters, and the objective might be to reach the Moon with minimum fuel expenditure. Or the dynamical system could be a nation's economy, with the objective to minimize unemployment; the controls in this case could be fiscal and monetary policy. A dynamical system may also be introduced to embed operations research problems within the framework of optimal control theory. Optimal control is an extension of the calculus of variations, and is a mathematical optimization method for deriving control policies. The method is largely due to the work of Lev Pontryagin and Richard Bellman in the 1950s, after contributions to calculus of variations by Edward J. McShane. Optimal control can be seen as a control strategy in control theory. General method Optimal control deals with the problem of finding a control law for a given system such that a certain optimality criterion is achieved. A control problem includes a cost functional that is a function of state and control variables. An optimal control is a set of differential equations describing the paths of the control variables that minimize the cost function. The optimal control can be derived using Pontryagin's maximum principle (a necessary condition also known as Pontryagin's minimum principle or simply Pontryagin's principle), or by solving the Hamilton–Jacobi–Bellman equation (a sufficient condition). We begin with a simple example. Consider a car traveling in a straight line on a hilly road. The question is, how should the driver press the accelerator pedal in order to minimize the total traveling time? In this example, the term control law refers specifically to the way
https://en.wikipedia.org/wiki/Thread%20%28computing%29
In computer science, a thread of execution is the smallest sequence of programmed instructions that can be managed independently by a scheduler, which is typically a part of the operating system. In many cases, a thread is a component of a process. The multiple threads of a given process may be executed concurrently (via multithreading capabilities), sharing resources such as memory, while different processes do not share these resources. In particular, the threads of a process share its executable code and the values of its dynamically allocated variables and non-thread-local global variables at any given time. The implementation of threads and processes differs between operating systems. In Modern Operating Systems, Tanenbaum shows that many distinct models of process organization are possible. History Threads made an early appearance under the name of "tasks" in OS/360 Multiprogramming with a Variable Number of Tasks (MVT) in 1967. Saltzer (1966) credits Victor A. Vyssotsky with the term "thread". The use of threads in software applications became more common in the early 2000s as CPUs began to utilize multiple cores. Applications wishing to take advantage of multiple cores for performance advantages were required to employ concurrency to utilize the multiple cores. Related concepts Scheduling can be done at the kernel level or user level, and multitasking can be done preemptively or cooperatively. This yields a variety of related concepts. Processes At the kernel level, a process contains one or more kernel threads, which share the process's resources, such as memory and file handles – a process is a unit of resources, while a thread is a unit of scheduling and execution. Kernel scheduling is typically uniformly done preemptively or, less commonly, cooperatively. At the user level a process such as a runtime system can itself schedule multiple threads of execution. If these do not share data, as in Erlang, they are usually analogously called proces
https://en.wikipedia.org/wiki/Terzaghi%27s%20principle
Terzaghi's Principle states that when stress is applied to a porous material, it is opposed by the fluid pressure filling the pores in the material. Karl von Terzaghi introduced the idea in a series of papers in the 1920s based on his examination of building consolidation on soil. The principle states that all quantifiable changes in stress to a porous medium are a direct result of a change in effective stress. The effective stress, , is related to total stress, , and the pore pressure, , by , where is the identity matrix. The negative sign is there because the pore pressure serves to lessen the volume-changing stress; physically this is because there is fluid in the pores which bears a part of the total stress, so partially unloading the solid matrix from normal stresses. Terzaghi's principle applies well to porous materials whose solid constituents are incompressible - soil, for example, is composed of grains of incompressible silica so that the volume change in soil during consolidation is due solely to the rearrangement of these constituents with respect to one another. Generalizing Terzaghi's principle to include compressible solid constituents was accomplished by Maurice Anthony Biot in the 1940s, giving birth to the theory of poroelasticity and poromechanics. Assumptions of Terzaghi's Principle The soil is homogenous (uniform in composition throughout) and isotropic (show same physical property in each direction). The soil is fully saturated (zero air voids due to water content being so high). The solid particles are incompressible. Compression and flow are one-dimensional (vertical axis being the one of interest). Strains in the soil are relatively small. Darcy's Law is valid for all hydraulic gradients. The coefficient of permeability and the coefficient of volume compressibility remain constant throughout the process. There is a unique relationship, independent of time, between the void ratio and effective stress Validity Though the first
https://en.wikipedia.org/wiki/Isotopomer
Isotopomers or isotopic isomers are isomers which differ by isotopic substitution, and which have the same number of atoms of each isotope but in a different arrangement. For example, CH3OD and CH2DOH are two isotopomers of monodeuterated methanol. The molecules may be either structural isomers (constitutional isomers) or stereoisomers depending on the location of the isotopes. Isotopomers have applications in areas including nuclear magnetic resonance spectroscopy, reaction kinetics, and biochemistry. Description Isotopomers or isotopic isomers are isomers with isotopic atoms, having the same number of each isotope of each element but differing in their positions in the molecule. The result is that the molecules are either constitutional isomers or stereoisomers solely based on isotopic location. The term isotopomer was first proposed by Seeman and Paine in 1992 to distinguish isotopic isomers from isotopologues (isotopic homologues). Examples CH3CHDCH3 and CH3CH2CH2D are a pair of structural isotopomers of propane. (R)- and (S)-CH3CHDOH are isotopic stereoisomers of ethanol. (Z)- and (E)-CH3CH=CHD are examples of isotopic stereoisomers of propene. Use 13C-NMR In nuclear magnetic resonance spectroscopy, the highly abundant 12C isotope does not produce any signal whereas the comparably rare 13C isotope is easily detected. As a result, carbon isotopomers of a compound can be studied by carbon-13 NMR to learn about the different carbon atoms in the structure. Each individual structure that contains a single 13C isotope provides data about the structure in its immediate vicinity. A large sample of a chemical contains a mixture of all such isotopomers, so a single spectrum of the sample contains data about all carbons in it. Nearly all of the carbon in normal samples of carbon-based chemicals is 12C, with only about 1% abundance of 13C, so there is only about a 1% abundance of the total of the singly-substituted isotopologues, and exponentially smaller amoun
https://en.wikipedia.org/wiki/Basic%20Battle%20Skills
Basic Battle Skills is a field training manual formerly issued to individual soldiers in the British Army. Editions were modified over a period of approximately 25 years before the book was withdrawn from issue. Format The manual was divided into seven chapters, each one an illustrated set of instructions on how to carry out the most basic subjects taught to soldiers during their first weeks of training. Individual Fieldcraft and Minor Tactics Map Reading Basic First aid Health and Hygiene NBC Defence (NBC=Nuclear, Biological and Chemical [warfare]) administration and morale Physical Fitness The opening chapter Individual Fieldcraft and Minor Tactics starts with the information: A sound knowledge of Fieldcraft and Minor Tactics will help you in combat to: Fight Live Protect yourself and your equipment Remain fit Be an asset to your commander In the first chapter the manual describes in more detail the skills needed in Fieldcraft and Minor Tactics. Commencing with "Judging Distance" and describing methods on how to ascertain how far a subject is from the soldier up to ranges of 400 metres and recommending that weapon sights be carefully adjusted for accuracy. The chapter also gives guidance on "Personal Camouflage", "Target Recognition", "Fire Control Orders", "Movement in the field", "Movement at night", "Night Noises", "Night vision", "Sentry duties", "Sentries at night in the field", "Rifle section formations", "Field signals (on foot)", "Advancing towards the enemy", "Weapon handling", "Ammunition" and "Trenches". Illustrations The book contains simple illustrations with vivid colours detailing in pictures what the wording of each section instructs a soldier to do in various situations. For example: in chapter one instructions are provided on the method and wording to be used when challenging movement of persons approaching the sentry position. The manual instructs the sentry to call out, "Halt, who goes there? Advance one and be recognised." When
https://en.wikipedia.org/wiki/Transit%20Wireless
Transit Wireless is an American telecommunication company founded in 2005, based in New York City. It was formed as a consortium of several entities, including Dianet Communications. It specializes in building wireless communication infrastructure using distributed antenna system networks to provide Wi-Fi and cellular phone coverage in the places that are unreachable by traditional cellular phone services such as in the underground portions of the New York City Subway. In 2010, the company was injected with financial support from infrastructure company BAI Communications for its first project with the New York City Transit Authority, which consisted of adding wireless access to subway stations. The architectural framework for the system and wireless engineering efforts were led by RF designer Mark Parr. The resulting wireless solution grew to provide coverage for hundreds of stations and serve well over a billion riders annually. The company is now a subsidiary of BAI Communications.
https://en.wikipedia.org/wiki/Google%20Pay%20%28payment%20method%29
Google Pay (formerly Android Pay) is a mobile payment service developed by Google to power in-app, online, and in-person contactless purchases on mobile devices, enabling users to make payments with Android phones, tablets, or watches. Users can authenticate via a PIN, passcode, or biometrics such as 3D face scanning or fingerprint recognition. , it is currently available in 71 countries. Service Google Pay uses near-field communication (NFC) to transmit card information facilitating funds transfer to the retailer. It replaces the credit or debit card chip and PIN or magnetic stripe transaction at point-of-sale terminals by allowing the user to upload these in Google Wallet. It is similar to contactless payments already used in many countries, with the addition of two-factor authentication. The service lets Android devices wirelessly communicate with point of sale systems using a near field communication (NFC) antenna and host-based card emulation (HCE). When the user makes a payment to a merchant, Google Pay does not send the actual payment card number. Instead, it generates a virtual account number representing the user's account information. Google Pay requires that a screen lock be set on the phone or watch. It has no card limit. Users can add payment cards to the service by taking a photo of the card, or by entering the card information manually. To pay at points of sale, users hold their authenticated device to the point of sale system. The service has smart-authentication, allowing the system to detect when the device is considered secure (for instance if unlocked in the last five minutes) and challenge if necessary for unlock information. Technology Google Pay uses the EMV Payment Tokenization Specification. The service keeps customer payment information private from the retailer by replacing the customer's credit or debit card Funding Primary Account Number (FPAN) with a tokenized Device Primary Account Number (DPAN) and creates a "dynamic security
https://en.wikipedia.org/wiki/WMYV
WMYV (channel 48) is a television station licensed to Greensboro, North Carolina, United States, serving the Piedmont Triad region as an affiliate of MyNetworkTV. It is owned by Sinclair Broadcast Group alongside Winston-Salem–licensed ABC affiliate WXLV-TV (channel 45). Both stations share studios on Myer Lee Drive (along US 421) in Winston-Salem, while WMYV's transmitter is located in Randleman (along I-73/US 220). History Prior history of UHF channel 48 in the Piedmont Triad The UHF channel 48 allocation in Greensboro was originally occupied by WUBC, an independent station that operated from 1967 until 1970. Early history The current licensed station on channel 48 first signed on the air on May 9, 1981, as WGGT, running a general entertainment format featuring cartoons, classic movies, classic sitcoms, religious programs, and CBS network shows that were preempted by WFMY-TV (channel 2), as well as business news programming from the Financial News Network. It was owned locally by Guilford Broadcasters. The station was initially a strong competitor with the area's other independent station, WJTM-TV (channel 45). However, it took a toll financially. The Piedmont Triad region was too small at the time to support two independent stations, and there was not enough programming to go around. In addition, WGGT's signal was considerably weaker than rival WJTM's. WGGT fell further behind when TVX Broadcast Group bought WJTM in 1983 and changed its call letters to WNRW. With stronger ownership, WNRW was able to pick clean most of the better programming that was available, and had also beat out WGGT for the Triad's Fox affiliation in 1986. In the mid-1980s, the station aired a program on Sunday mornings called Pet Pals, which showcased a variety of information for people and pets. Co-hosted by Greensboro native Jerry Cunningham, Pet Pals was North Carolina's first weekly program dedicated to pets and became a hit. The most visible on-air personality at WGGT was "Billy Bo
https://en.wikipedia.org/wiki/Rainbow%20gravity%20theory
Rainbow gravity (or "gravity's rainbow") is a theory that different wavelengths of light experience different gravity levels and are separated in the same way that a prism splits white light into the rainbow. This phenomenon would be imperceptible in areas of relatively low gravity, such as Earth, but would be significant in areas of extremely high gravity, such as a black hole. As such the theory claims to disprove that the universe has a beginning or Big Bang, as the big bang theory calls for all wavelengths of light to be impacted by gravity to the same extent. The theory was first proposed in 2003 by physicists Lee Smolin and João Magueijo, and claims to bridge the gap between general relativity and quantum mechanics. Scientists are currently attempting to detect rainbow gravity using the Large Hadron Collider. Background Rainbow gravity theory's origin is largely the product of the disparity between general relativity and quantum mechanics. More specifically, "locality," or the concept of cause and effect that drives the principles of general relativity, is mathematically irreconcilable with quantum mechanics. This issue is due to incompatible functions between the two fields; in particular, the fields apply radically different mathematical approaches in describing the concept of curvature in four-dimensional space-time. Historically, this mathematical split begins with the disparity between Einstein's theories of relativity, which saw physics through the lens of causality, and classical physics, which interpreted the structure of space-time to be random and inherent. The prevailing notion about cosmic change is that the universe is expanding at a constantly accelerating rate; moreover, it is understood that as one traces the universe's history backwards one finds that it was, at one point, far denser. If true, the Rainbow gravity theory prohibits a singularity such as that which is postulated in the Big Bang. This indicates that, when viewed in reverse, the
https://en.wikipedia.org/wiki/Tear%20resistance
Tear resistance (or tear strength) is a measure of how well a material can withstand the effects of tearing. It is a useful engineering measurement for a wide variety of materials by many different test methods. Discussion For example, with rubber, tear resistance measures how the test specimen resists the growth of any cuts when under tension, it is usually expressed in kN/m. Tear resistance can be gauged via the same ASTM D 412 apparatus used to measure tensile strength, modulus and elongation. ASTM D 624 can be applied to measure the resistance to the formation of a tear (tear initiation) and the resistance to the expansion of a tear (tear propagation). Regardless of which of these two is being measured, the sample is held between two holders and a uniform pulling force applied until the aforementioned deformation occurs. Tear resistance is then calculated by dividing the force applied by the thickness of the material. Materials with low tear resistance sometimes have poor resistance to abrasion and when damaged will quickly fail (this includes hard materials, since hardness is not related to tear resistance). Tear resistance of different materials Substances with high tear resistance include epichlorohydrin, natural rubber and polyurethane. In contrast, materials such as silicone and fluorosilicone have low tear resistance. Tear-yield ratio The ratio of tear resistance to the yield strength is called the tear-yield ratio. It is a measure of notch toughness. Standards by ASTM International C1681-14 Standard Test Method for Evaluating the Tear Resistance of a Sealant Under Constant Strain D1004-13 Standard Test Method for Tear Resistance (Graves Tear) of Plastic Film and Sheeting D1922-15 Standard Test Method for Propagation Tear Resistance of Plastic Film and Thin Sheeting by Pendulum Method D1938-14 Standard Test Method for Tear-Propagation Resistance (Trouser Tear) of Plastic Film and Thin Sheeting by a Single-Tear Method D2212-00(2015) Standard Test Meth
https://en.wikipedia.org/wiki/H4K12ac
H4K12ac is an epigenetic modification to the DNA packaging protein histone H4. It is a mark that indicates the acetylation at the 12th lysine residue of the histone H4 protein. H4K12ac is involved in learning and memory. It is possible that restoring this modification could reduce age-related decline in memory. Nomenclature H4K12ac indicates acetylation of lysine 12 on histone H4 protein subunit: Histone modifications The genomic DNA of eukaryotic cells is wrapped around special protein molecules known as histones. The complexes formed by the looping of the DNA are known as chromatin. The basic structural unit of chromatin is the nucleosome: this consists of the core octamer of histones (H2A, H2B, H3 and H4) as well as a linker histone and about 180 base pairs of DNA. These core histones are rich in lysine and arginine residues. The carboxyl (C) terminal end of these histones contribute to histone-histone interactions, as well as histone-DNA interactions. The amino (N) terminal charged tails are the site of the post-translational modifications, such as the one seen in H3K36me3. H4 histone H4 modifications are not as well known as H3's and H4 have fewer variations which might explain their important function. H4K12ac Acetylation of histone H4K5 and H4K12 is enriched at centromeres. H4K8ac and H4K12ac are associated with active promoters to form a backbone. H4 localizes more to gene bodies promoters than other acetylations so H4K8ac facilitates transcriptional elongation. H4K12ac is involved in learning and memory so it could help with reducing age-related decline in memory. Lysine acetylation and deacetylation Proteins are typically acetylated on lysine residues and this reaction relies on acetyl-coenzyme A as the acetyl group donor. In histone acetylation and deacetylation, histone proteins are acetylated and deacetylated on lysine residues in the N-terminal tail as part of gene regulation. Typically, these reactions are catalyzed by enzymes with hist
https://en.wikipedia.org/wiki/Conservation%20paleobiology
Conservation paleobiology is a field of paleontology that applies the knowledge of the geological and paleoecological record to the conservation and restoration of biodiversity and ecosystem services. Despite the influence of paleontology on ecological sciences can be traced back at least at the 18th century, the current field has been established by the work of K.W. Flessa and G.P. Dietl in the first decade of the 21st century. The discipline utilizes paleontological and geological data to understand how biotas respond to climate and other natural and anthropogenic environmental change. These information are then used to address the challenges faced by modern conservation biology, like understanding the extinction risk of endangered species, providing baselines for restoration and modelling future scenarios for species range's contraction or expansion. Description of the discipline The main strength of conservation paleobiology is the availability of long term data on species, communities and ecosystems that exceeds the timeframe of direct human experience.  The discipline takes one of two approaches: near-time and deep-time. Near-time conservation paleobiology The near-time approach uses the recent fossil record (usually from the Late Pleistocene or the Holocene) to provide a long-term context to extant ecosystems dynamics. The fossil record is, in many cases, the only source of information on conditions previous to human impacts. These records can be used as reference baselines for comparisons in order to identify targets for restoration ecology, to analyze species responses to perturbations (natural and anthropogenic), understand historical species distributions and their variability, discriminate the factors that distinguish natural from non-natural changes in biological populations and identify ecological legacies only explicable by referring to past events or conditions. Example - Conservation of the European bison The European bison or wisent (Bison
https://en.wikipedia.org/wiki/Wang%20Institute%20of%20Graduate%20Studies
The Wang Institute of Graduate Studies was an independent educational institution founded in 1979 by computer entrepreneur An Wang. Its purpose was to provide professional and continuing studies in the nascent field of software engineering. It was accredited by the New England Association of Schools and Colleges in 1983. Faculty members were recruited from industry and students were required to have a minimum of three years prior experience in industry as a condition of acceptance. The Institute acquired its campus from the Marist Brothers who had operated a seminary on the site since 1924. Located in Tyngsborough, Massachusetts, it housed two divisions: The School of Information Technology and a fellowship program in East Asian studies. The Institute never grew beyond a dozen or so faculty. As a result of declining business fortunes Dr. Wang closed the Institute, graduating the last class on August 27, 1988.McKeeman, William, "Graduation Talk at Wang Institute," Computer, vol. 22, no. 5, pp. 78-80 (1989) The campus was transferred to Boston University where it served as a corporate education center. Today, it is the location of the Innovation Academy Charter School. Software engineering curriculum The Institute graduated seven classes between 1982 and 1988 in its Master of Software Engineering program, requiring study in eleven three-credit courses. Two project courses involved students in team-based analysis, specification, design, implementation, testing, and integration of software products.Fairley, Richard and Martin, Nancy. "Software engineering programs at the Wang Institute of Graduate Studies," Proceedings of the 1983 annual conference on Computers (1983) The original six core courses were: The curriculum was later modified to include an optional operating systems course instead of the architecture course.Ardis, Mark. "The Evolution of Wang Institute's Master of Software Engineering Program," IEEE Transactions on Software Engineering 13(11),
https://en.wikipedia.org/wiki/Broadcast%20seeding
In agriculture, gardening, and forestry, broadcast seeding is a method of seeding that involves scattering seed, by hand or mechanically, over a relatively large area. This is in contrast to: precision seeding, where seed is placed at a precise spacing and depth; hydroseeding, where a slurry of seed, mulch and water is sprayed over prepared ground in a uniform layer. Broadcast seeding is of particular use in establishing dense plant spacing, as for cover crops and lawns. In comparison to traditional drill planting, broadcast seeding will require 10–20% more seed. It is simpler, faster, and easier than traditional row sowing. Broadcast seeding works best for plants that do not require singular spacing or that are more easily thinned later. After broadcasting, seed is often lightly buried with some type of raking action, often done using vertical tillage tools. Utilizing these tools increases the success rate of germination by increasing seed-to-soil contact. Seeds sown in this manner are distributed unevenly, which may result in overcrowding. This method may not ensure that all seeds are sown at the correct depth. Incorrect depth, if too deep, would result in germination that would not allow the young plant to break the surface of the soil and prevent sprouting. If they are not sown evenly then there would be a lack of various nutrients from sunlight, oxygen, et cetera, in many crops or plants. Not all seeds are good candidates for broadcast seeding. Often, only smaller seeds will sprout and continue to grow successfully when planted by way of broadcasting. In general, the larger the seed, the deeper it can be planted. See also Aerial seeding
https://en.wikipedia.org/wiki/Actuarial%20reserves
In insurance, an actuarial reserve is a reserve set aside for future insurance liabilities. It is generally equal to the actuarial present value of the future cash flows of a contingent event. In the insurance context an actuarial reserve is the present value of the future cash flows of an insurance policy and the total liability of the insurer is the sum of the actuarial reserves for every individual policy. Regulated insurers are required to keep offsetting assets to pay off this future liability. The loss random variable The loss random variable is the starting point in the determination of any type of actuarial reserve calculation. Define to be the future state lifetime random variable of a person aged x. Then, for a death benefit of one dollar and premium , the loss random variable, , can be written in actuarial notation as a function of From this we can see that the present value of the loss to the insurance company now if the person dies in t years, is equal to the present value of the death benefit minus the present value of the premiums. The loss random variable described above only defines the loss at issue. For K(x) > t, the loss random variable at time t can be defined as: Net level premium reserves Net level premium reserves, also called benefit reserves, only involve two cash flows and are used for some US GAAP reporting purposes. The valuation premium in an NLP reserve is a premium such that the value of the reserve at time zero is equal to zero. The net level premium reserve is found by taking the expected value of the loss random variable defined above. They can be formulated prospectively or retrospectively. The amount of prospective reserves at a point in time is derived by subtracting the actuarial present value of future valuation premiums from the actuarial present value of the future insurance benefits. Retrospective reserving subtracts accumulated value of benefits from accumulated value of valuation premiums as of a point in time. T
https://en.wikipedia.org/wiki/Anatomic%20space
In anatomy, a spatium or anatomic space is a space (cavity or gap). Anatomic spaces are often landmarks to find other important structures. When they fill with gases (such as air) or liquids (such as blood) in pathological ways, they can suffer conditions such as pneumothorax, edema, or pericardial effusion. Many anatomic spaces are potential spaces, which means that they are potential rather than realized (with their realization being dynamic according to physiologic or pathophysiologic events). In other words, they are like an empty plastic bag that has not been opened (two walls collapsed against each other; no interior volume until opened) or a balloon that has not been inflated. Examples of anatomic spaces (or potential spaces) include: Axillary space Buccal space Canine space Cystohepatic triangle Deep perineal space Deep temporal space Epidural space Extraperitoneal space Fascial spaces of the head and neck Infratemporal space Intercostal space Intermembrane space Interstitial spaces Mental space Pericardial space Intraperitoneal space Pleural space Potential space Pterygomandibular space Quadrangular space Retroperitoneal space Retropharyngeal space Retropubic space Subarachnoid space Subdural space Sublingual space Submandibular space Submasseteric space Traube's space See also Body cavity Anatomy
https://en.wikipedia.org/wiki/Selberg%20class
In mathematics, the Selberg class is an axiomatic definition of a class of L-functions. The members of the class are Dirichlet series which obey four axioms that seem to capture the essential properties satisfied by most functions that are commonly called L-functions or zeta functions. Although the exact nature of the class is conjectural, the hope is that the definition of the class will lead to a classification of its contents and an elucidation of its properties, including insight into their relationship to automorphic forms and the Riemann hypothesis. The class was defined by Atle Selberg in , who preferred not to use the word "axiom" that later authors have employed. Definition The formal definition of the class S is the set of all Dirichlet series absolutely convergent for Re(s) > 1 that satisfy four axioms (or assumptions as Selberg calls them): Comments on definition The condition that the real part of μi be non-negative is because there are known L-functions that do not satisfy the Riemann hypothesis when μi is negative. Specifically, there are Maass forms associated with exceptional eigenvalues, for which the Ramanujan–Peterssen conjecture holds, and have a functional equation, but do not satisfy the Riemann hypothesis. The condition that θ < 1/2 is important, as the θ = 1 case includes whose zeros are not on the critical line. Without the condition there would be which violates the Riemann hypothesis. It is a consequence of 4. that the an are multiplicative and that Examples The prototypical example of an element in S is the Riemann zeta function. Another example, is the L-function of the modular discriminant Δ where and τ(n) is the Ramanujan tau function. All known examples are automorphic L-functions, and the reciprocals of Fp(s) are polynomials in p−s of bounded degree. The best results on the structure of the Selberg class are due to Kaczorowski and Perelli, who show that the Dirichlet L-functions (including the Riemann zeta-function