source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/Citral
Citral is an acyclic monoterpene aldehyde. Being a monoterpene, it is made of two isoprene units. Citral is a collective term which covers two geometric isomers that have their own separate names; the E-isomer is named geranial (trans-citral; α-citral) or citral A. The Z-isomer is named neral (cis-citral; β-citral) or citral B. These stereoisomers occur as a mixture, not necessarily racemic; e.g. in essential oil of Australian ginger, the neral to geranial ratio is 0.61. Occurrence Citral is present in the volatile oils of several plants, including lemon myrtle (90–98%), Litsea citrata (90%), Litsea cubeba (70–85%), lemongrass (65–85%), lemon tea-tree (70–80%), Ocimum gratissimum (66.5%), Lindera citriodora (about 65%), Calypranthes parriculata (about 62%), petitgrain (36%), lemon verbena (30–35%), lemon ironbark (26%), lemon balm (11%), lime (6–9%), lemon (2–5%), and orange. Further, in the lipid fraction (essential oil) of Australian ginger (51–71%) Of the many sources of citral, the Australian myrtaceous tree, lemon myrtle, Backhousia citriodora F. Muell. (of the family Myrtaceae), is considered superior. Uses Citral has a strong lemon (citrus) scent and is used as an aroma compound in perfumery. It is used to fortify lemon oil. (Nerol, another perfumery compound, has a less intense but sweeter lemon note.) The aldehydes citronellal and citral are considered key components responsible for the lemon note with citral preferred. It also has pheromonal effects in acari and insects. Citral is used in the synthesis of vitamin A, lycopene, ionone and methylionone, and to mask the smell of smoke. The herb Cymbopogon citratus has shown promising insecticidal and antifungal activity against storage pests. Food additive Citral is commonly used as a food additive ingredient. It has been tested (2016) in vitro against the food-borne pathogen Cronobacter sakazakii. Medical exploration In a report (1997), citral is mentioned as cytotoxic to P(388) mouse leukaemia ce
https://en.wikipedia.org/wiki/Perfume%20intolerance
Perfume intolerance or perfume allergy is a condition wherein people exhibit sensitivity or allergic reactions to ingredients in some perfumes and some other fragrances. It is a form of multiple chemical sensitivity, a more general phenomenon for this diagnosis. Symptoms Symptoms depend on each person's allergies and each perfume's or fragrance's ingredients. Symptoms may include allergic contact dermatitis, asthma attacks, headaches, and others. The most common allergic reactions to perfume or fragrances added to products is contact dermatitis, though other symptoms may occur, including allergic conjunctivitis. The diagnosis of the causal allergen is made by patch testing with a mixture of fragrance ingredients, the fragrance mix. This gives a positive patch-test reaction in about 10% of tested patients with eczema, and the most recent estimates show that 1.7–4.1% of the general population are sensitized to ingredients of the fragrance mix. Two studies show that inhalant-like allergies and sensitivity/intolerances are experienced by a subset of the US population, in the form of asthma and chemical sensitivities. Results aggregated from both surveys found that 30.5% of the general population reported scented products on others irritating, 19% reported adverse health effects from air fresheners, and 10.9% reported irritation by scented laundry products vented outside. Household products, such as soaps and detergents, perfume products, cosmetics, and other consumer goods, are estimated to use 2,500 different fragrance ingredients. Of those, approximately 100 different substances are known to elicit responses in at least some individuals. An estimated 1.7–4.1% of the general population shows a contact allergic response to a mix of common perfume ingredients. Although products can be labeled "fragrance-free", many still contain lesser-known fragrance chemicals that consumers may not recognize. Cinnamaldehyde (cinnamic aldehyde) is a common fragrance allergen.
https://en.wikipedia.org/wiki/Singular%20boundary%20method
In numerical analysis, the singular boundary method (SBM) belongs to a family of meshless boundary collocation techniques which include the method of fundamental solutions (MFS), boundary knot method (BKM), regularized meshless method (RMM), boundary particle method (BPM), modified MFS, and so on. This family of strong-form collocation methods is designed to avoid singular numerical integration and mesh generation in the traditional boundary element method (BEM) in the numerical solution of boundary value problems with boundary nodes, in which a fundamental solution of the governing equation is explicitly known. The salient feature of the SBM is to overcome the fictitious boundary in the method of fundamental solution, while keeping all merits of the latter. The method offers several advantages over the classical domain or boundary discretization methods, among which are: meshless. The method requires neither domain nor boundary meshing but boundary-only discretization points; integration-free. The numerical integration of singular or nearly singular kernels could be otherwise troublesome, expensive, and complicated, as in the case, for example, the boundary element method; boundary-only discretization for homogeneous problems. The SBM shares all the advantages of the BEM over domain discretization methods such as the finite element or finite difference methods; to overcome the perplexing fictitious boundary in the method of fundamental solutions (see Figs. 1 and 2), thanks to the introduction of the concept of the origin intensity factor, which isolates the singularity of the fundamental solutions. The SBM provides a significant and promising alternative to popular boundary-type methods such as the BEM and MFS, in particular, for infinite domain, wave, thin-walled structures, and inverse problems. History of the singular boundary method The methodology of the SBM was firstly proposed by Chen and his collaborators in 2009. The basic idea is to introduce a
https://en.wikipedia.org/wiki/Taste%20receptor
A taste receptor or tastant is a type of cellular receptor which facilitates the sensation of taste. When food or other substances enter the mouth, molecules interact with saliva and are bound to taste receptors in the oral cavity and other locations. Molecules which give a sensation of taste are considered "sapid". Vertebrate taste receptors are divided into two families: Type 1, sweet, first characterized in 2001: – Type 2, bitter, first characterized in 2000: In humans there are 25 known different bitter receptors, in cats there are 12, in chickens there are three, and in mice there are 35 known different bitter receptors. Visual, olfactive, "sapictive" (the perception of tastes), trigeminal (hot, cool), mechanical, all contribute to the perception of taste. Of these, transient receptor potential cation channel subfamily V member 1 (TRPV1) vanilloid receptors are responsible for the perception of heat from some molecules such as capsaicin, and a CMR1 receptor is responsible for the perception of cold from molecules such as menthol, eucalyptol, and icilin. Tissue distribution The gustatory system consists of taste receptor cells in taste buds. Taste buds, in turn, are contained in structures called papillae. There are three types of papillae involved in taste: fungiform papillae, foliate papillae, and circumvallate papillae. (The fourth type - filiform papillae do not contain taste buds). Beyond the papillae, taste receptors are also in the palate and early parts of the digestive system like the larynx and upper esophagus. There are three cranial nerves that innervate the tongue; the vagus nerve, glossopharyngeal nerve, and the facial nerve. The glossopharyngeal nerve and the chorda tympani branch of the facial nerve innervate the TAS1R and TAS2R taste receptors. Next to the taste receptors in on the tongue, the gut epithelium is also equipped with a subtle chemosensory system that communicates the sensory information to several effector systems involved
https://en.wikipedia.org/wiki/Ad%20hominem
, short for argumentum ad hominem, is a term that refers to several types of arguments, most of which are fallacious. Typically this term refers to a rhetorical strategy where the speaker attacks the character, motive, or some other attribute of the person making an argument rather than attacking the substance of the argument itself. This avoids genuine debate by creating a diversion to some irrelevant but often highly charged issue. The most common form of this fallacy is "A makes a claim x, B asserts that A holds a property that is unwelcome, and hence B concludes that argument x is wrong". The valid types of ad hominem arguments are generally only encountered in specialized philosophical usage. These typically refer to the dialectical strategy of using the target's own beliefs and arguments against them, while not agreeing with the validity of those beliefs and arguments. Ad hominem arguments were first studied in ancient Greece; John Locke revived the examination of ad hominem arguments in the 17th century. Many contemporary politicians routinely use ad hominem attacks, which can be encapsulated to a derogatory nickname for a political opponent. History The various types of ad hominem arguments have been known in the West since at least the ancient Greeks. Aristotle, in his work Sophistical Refutations, detailed the fallaciousness of putting the questioner but not the argument under scrutiny. His description was somewhat different from the modern understanding, referring to a class of sophistry that applies an ambiguously worded question about people to a specific person. The proper refutation, he wrote, is not to debate the attributes of the person (solutio ad hominem) but to address the original ambiguity. Many examples of ancient non-fallacious ad hominem arguments are preserved in the works of the Pyrrhonist philosopher Sextus Empiricus. In these arguments, the concepts and assumptions of the opponents are used as part of a dialectical strategy against t
https://en.wikipedia.org/wiki/Deep-level%20transient%20spectroscopy
Deep-level transient spectroscopy (DLTS) is an experimental tool for studying electrically active defects (known as charge carrier traps) in semiconductors. DLTS establishes fundamental defect parameters and measures their concentration in the material. Some of the parameters are considered as defect "finger prints" used for their identifications and analysis. DLTS investigates defects present in a space charge (depletion) region of a simple electronic device. The most commonly used are Schottky diodes or p-n junctions. In the measurement process the steady-state diode reverse polarization voltage is disturbed by a voltage pulse. This voltage pulse reduces the electric field in the space charge region and allows free carriers from the semiconductor bulk to penetrate this region and recharge the defects causing their non-equilibrium charge state. After the pulse, when the voltage returns to its steady-state value, the defects start to emit trapped carriers due to the thermal emission process. The technique observes the device space charge region capacitance where the defect charge state recovery causes the capacitance transient. The voltage pulse followed by the defect charge state recovery are cycled allowing an application of different signal processing methods for defect recharging process analysis. The DLTS technique has a higher sensitivity than almost any other semiconductor diagnostic technique. For example, in silicon it can detect impurities and defects at a concentration of one part in 1012 of the material host atoms. This feature together with a technical simplicity of its design made it very popular in research labs and semiconductor material production factories. The DLTS technique was pioneered by David Vern Lang at Bell Laboratories in 1974. A US Patent was awarded to Lang in 1975. DLTS methods Conventional DLTS In conventional DLTS the capacitance transients are investigated by using a lock-in amplifier or double box-car averaging technique whe
https://en.wikipedia.org/wiki/Lewis%20Goldberg
Lewis R. Goldberg is an American personality psychologist and a professor emeritus at the University of Oregon. He is closely associated with the lexical hypothesis that any culturally important personality characteristic will be represented in the language of that culture. This hypothesis led to a five factor structure of personality trait adjectives (which he dubbed the Big 5). When applied to personality items this structure is also known as the five-factor model (FFM) of personality. He is the creator of the International Personality Item Pool(IPIP), a website that provides public-domain personality measures. Education Lew Goldberg was born in Chicago, Illinois, on January 28, 1932. His early education took place at the Bret Harte elementary school in Chicago, and the Highland Park High School in Highland Park, Illinois. In 1953 Goldberg received an A.B. in social relations from Harvard University. He earned a Ph.D. in psychology from the University of Michigan in 1958 where his Ph.D. advisor was E. Lowell Kelly; Kelly provided Goldberg with training in the methodology of quantitative personality assessment. Academic career As an advanced graduate student at Michigan, Goldberg met Warren T. Norman, a new assistant professor, who became a lifelong friend and collaborator on issues in personality structure and assessment. Their initial work together on the generality of the Big 5, and their subsequent work on the lexical hypothesis has had a major impact upon the development of a consensual model of personality. After receiving his doctorate, Goldberg became a visiting assistant professor at Stanford University. Since 1960 he has taught at the University of Oregon, where he is professor emeritus. He is a senior scientist at the Oregon Research Institute, where he has carried out research since 1961. From 1962 to 1966, Goldberg served as a field selection officer for the United States Peace Corps. In 1966 he became a Fulbright professor at the Univers
https://en.wikipedia.org/wiki/SailTimer
SailTimer is a technology for sailboat navigation, which calculates optimal tacking angles, distances and times. Seafarers have had a dilemma for as long as sailboats could use lift to sail upwind. Should you head off the wind more, to get more speed (but at a longer distance)? Or should you sail upwind more, to lessen the distance (but at slower speed)? Different sailboats have varying capabilities for sailing upwind because of the condition of their sails, the type of rigging, and the ability to track in the water based on their keel or centerboard design. That means there is not a single rule for everyone; each vessel has its own individual polar plot of boat speed for all of the possible wind angles. The mathematical calculations for determining the tacking angles and times with polar plots would not be practical to do mentally. Rationale Civilian access to GPS navigation became available circa 1989. Estimated time of arrival (ETA) is a standard GPS chartplotter parameter, based on the assumption that the route will be a straight line to the destination. However, sailboats typically tack back and forth, creating a longer distance than a straight line. SailTimer evaluates tacking distances and boat speeds on different points of sail to determine optimal tacking headings. During the original development of the SailTimer software in 2005–2006, the term TTD (tm) was coined for "Tacking Time to Destination". Safe Navigation Sailboats can be pushed downwind by the wind, or can use lift to move across or into the wind. However, for most sailboats the boat speed drops if sailing closer than 45 degrees to the wind. The highest speed for most sailboats is reaching with the wind from the side or the aft quarter (a beam reach or a broad reach). If the destination is upwind but the sailboat goes fastest heading away from this direction, this poses a significant problem: how to choose tack headings with the best tradeoff between maximimizing speed and minimizi
https://en.wikipedia.org/wiki/Ad%20Quadratum%3A%20The%20Practical%20Application%20of%20Geometry%20in%20Medieval%20Architecture
Ad Quadratum: The Practical Application of Geometry in Medieval Architecture is an edited volume on the mathematical design of medieval architecture. It was edited by Nancy Y. Wu, published in 2002 by Ashgate Publishing, and reprinted in 2016 by Routledge. Title The title, ad quadratum, refers to a phrase used by medieval architects to describe building designs based on the geometry of the square, including the use of ratios based on polygonal geometry such as the square root of two ratio between the sides and diagonal of the square. The phrase has also been used previously for other publications on the geometric study of medieval architecture, notably in a 1921 book by Frederik Macody Lund, Ad Quadratum: A Study of the Geometrical Bases of Classic and Medieval Religious Architecture. Topics After a preface by Wu and W. W. Clark, and an introduction by Eric Fernie, Ad Quadratum includes eleven chapters: "Geometry on a Carolingian wall", by Warren Sanderson, studies the triangular composition of paintings on a wall in Trier, Germany. "A proposal for constructing the plan and elevation of a Romanesque church using three measures", by Marie-Therèse Zenner, studies the using a "very complex and farraginous hypothetical geometrical layout". "Measure and proportion in Romanesque architecture", by James Addis, studies the same church as Zenner, instead finding a modular system of measurements based on the Roman foot. "A schematic plan for Norwich Cathedral", by Nigel Hiscock, proposes a complex design principle for Norwich Cathedral combining triangular, square, and pentagonal forms, criticized by Christian Freigang as "hardly plausible" for such a repetitive and consistently designed building. "The plan of Saint-Quentin: Pentagon and square in the genesis of high Gothic design", by Ellen M. Shortell, shows the appearance of pentagonal forms beginning around the 13th century, in the Basilica of Saint-Quentin. "The hand of the mind: The ground plan of Reims as a case stu
https://en.wikipedia.org/wiki/Spinodal%20decomposition
Spinodal decomposition is a mechanism by which a single thermodynamic phase spontaneously separates into two phases (without nucleation). Decomposition occurs when there is no thermodynamic barrier to phase separation. As a result, phase separation via decomposition does not require the nucleation events resulting from thermodynamic fluctuations, which normally trigger phase separation. Spinodal decomposition is observed when mixtures of metals or polymers separate into two co-existing phases, each rich in one species and poor in the other. When the two phases emerge in approximately equal proportion (each occupying about the same volume or area), characteristic intertwined structures are formed that gradually coarsen (see animation). The dynamics of spinodal decomposition is commonly modeled using the Cahn–Hilliard equation. Spinodal decomposition is fundamentally different from nucleation and growth. When there is a nucleation barrier to the formation of a second phase, time is taken by the system to overcome that barrier. As there is no barrier (by definition) to spinodal decomposition, some fluctuations (in the order parameter that characterizes the phase) start growing instantly. Furthermore, in spinodal decomposition, the two distinct phases start growing in any location uniformly throughout the volume, whereas a nucleated phase change begins at a discrete number of points. Spinodal decomposition occurs when a homogenous phase becomes thermodynamically unstable. An unstable phase lies at a maximum in free energy. In contrast, nucleation and growth occur when a homogenous phase becomes metastable. That is, another biphasic system becomes lower in free energy, but the homogenous phase remains at a local minimum in free energy, and so is resistant to small fluctuations. J. Willard Gibbs described two criteria for a metastable phase: that it must remain stable against a small change over a large area. History In the early 1940s, Bradley reported the observa
https://en.wikipedia.org/wiki/Tony%20DeBlase
Tony DeBlase (1942–2000), also known as Anthony DeBlase, was part of the BDSM and leather subcultures. He was the designer of the leather pride flag. Leather and BDSM activities In 1979, he founded DungeonMaster magazine (about sadomasochistic technique), which ran from 1979 until 1994. He also wrote for that magazine. In 1982, under the pen name Fledermaus, he published a collection of fictional sadomasochism stories, titled The Fledermaus Anthology. In 1986, the leather magazine Drummer was sold to DeBlase, who sold it in 1991 to Martijn Bakker, owner of RoB Amsterdam. On May 28, 1989, DeBlase first presented the leather pride flag (which he had designed), at International Mister Leather. Initial reaction to the flag was mixed. According to DeBlase's article A Leather Pride Flag, On September 18, 1990, Clive Platman (Mr. Australia Drummer) presented DeBlase with an Australian version of the flag, incorporating the southern cross, which is from the Australian national flag, with the original design of the leather pride flag. In 1991, the Leather Archives & Museum (the LA&M) was founded by DeBlase and Chuck Renslow, "as a community archives, library, and museum of Leather, kink, fetish, and BDSM history and culture." DeBlase served as Vice President of the Board of Directors there from 1992 until 2000. He also began a Leather History Timeline, which was set up in the LA&M on the south wall of the main exhibit gallery. His papers (among other things) are held in the LA&M. At International Mr. Leather 1999 DeBlase presented one of three original leather pride flags which he assembled as a prototype to the Leather Archives & Museum. Awards In 1987 DeBlase received NLA's Man of the Year award. In 1990 he received the Business Person of the Year Award as part of the Pantheon of Leather Awards. He also received the Pantheon's Lifetime Achievement Award in 1994, their Forebear Award in 1997, and their Community Choice (Man) Award in 2001. In 2010 he was inducted
https://en.wikipedia.org/wiki/GitHub
GitHub, Inc. () is a platform and cloud-based service for software development and version control using Git, allowing developers to store and manage their code. It provides the distributed version control of Git plus access control, bug tracking, software feature requests, task management, continuous integration, and wikis for every project. Headquartered in California, it has been a subsidiary of Microsoft since 2018. It is commonly used to host open source software development projects. , GitHub reported having over 100 million developers and more than 372 million repositories, including at least 28 million public repositories. It is the world's largest source code host . About Founding Development of the GitHub platform began on October 19, 2007. The site was launched in April 2008 by Tom Preston-Werner, Chris Wanstrath, P. J. Hyett and Scott Chacon after it had been available for a few months as a beta release. GitHub has an annual keynote called GitHub Universe. Organizational structure GitHub, Inc. was originally a flat organization with no middle managers; in other words, "everyone is a manager" (self-management). Employees could choose to work on projects that interested them (open allocation), but the chief executive set salaries. (i.e. Individual or groups of company executive leaders decides on project aims and development, including funding) In 2014, GitHub, Inc. added a layer of middle management in response to serious harassment allegations against its senior leadership. As a result of the scandal, Tom Preston-Werner resigned from his position as CEO. Finance GitHub was a bootstrapped start-up business, which in its first years provided enough revenue to be funded solely by its three founders and start taking on employees. In July 2012, four years after the company was founded, Andreessen Horowitz invested $100 million in venture capital. In July 2015 GitHub raised another $250 million of venture capital in a series B round. Investors were S
https://en.wikipedia.org/wiki/Thermal%20time%20scale
In astrophysics, the thermal time scale or Kelvin–Helmholtz time scale is the approximate time it takes for a star to radiate away its total kinetic energy content at its current luminosity rate. Along with the nuclear and free-fall (aka dynamical) time scales, it is used to estimate the length of time a particular star will remain in a certain phase of its life and its lifespan if hypothetical conditions are met. In reality, the lifespan of a star is greater than what is estimated by the thermal time scale because as one fuel becomes scarce, another will generally take its place – hydrogen burning gives way to helium burning, which is replaced by carbon burning. Stellar astrophysics The size of a star as well as its energy output generally determine a star's thermal lifetime because the measurement is independent of the type of fuel normally found at its center. Indeed, the thermal time scale assumes that there is no fuel at all inside the star and simply predicts the length of time it would take for the resulting change in outputted energy to reach the surface of the star and become visually apparent to an outside observer. where is the gravitational constant, is the mass of the star, is the radius of the star, and is the star's luminosity. As an example, the Sun's thermal time scale is approximately 15.7 million years.
https://en.wikipedia.org/wiki/Precursor%20%28physics%29
Precursors are characteristic wave patterns caused by dispersion of an impulse's frequency components as it propagates through a medium. Classically, precursors precede the main signal, although in certain situations they may also follow it. Precursor phenomena exist for all types of waves, as their appearance is only predicated on the prominence of dispersion effects in a given mode of wave propagation. This non-specificity has been confirmed by the observation of precursor patterns in different types of electromagnetic radiation (microwaves, visible light, and terahertz radiation) as well as in fluid surface waves and seismic waves. History Precursors were first theoretically predicted in 1914 by Arnold Sommerfeld for the case of electromagnetic radiation propagating through a neutral dielectric in a region of normal dispersion. Sommerfeld's work was expanded in the following years by Léon Brillouin, who applied the saddle point approximation to compute the integrals involved. However, it was not until 1969 that precursors were first experimentally confirmed for the case of microwaves propagating in a waveguide, and much of the experimental work observing precursors in other types of waves has only been done since the year 2000. This experimental lag is mainly due to the fact that in many situations, precursors have a much smaller amplitude than the signals that give rise to them (a baseline figure given by Brillouin is six orders of magnitude smaller). As a result, experimental confirmations could only be done after technology became available to detect precursors. Basic theory As a dispersive phenomenon, the amplitude at any distance and time of a precursor wave propagating in one dimension can be expressed by the Fourier integral where is the Fourier transform of the initial impulse and the complex exponential represents the individual component wavelets summed in the integral. To account for the effects of dispersion, the phase of the exponential must inc
https://en.wikipedia.org/wiki/Inter-working%20function
The inter-working function (IWF) is a method for interfacing a wireless telecommunication network with the public switched telephone network (PSTN). The IWF converts the data transmitted over the air interface into a format suitable for the PSTN. IWF contains both the hardware and software elements that provide the rate adaptation and protocol conversion between PSTN and the wireless network. Some systems require more IWF capability than others, depending on the network which is being connected. The IWF also incorporates a "modem bank", which may be used when, for example, the GSM data terminal equipment (DTE) exchanges data with a land DTE connected via analogue modem The IWF provides the function to enable the GSM system to interface with the various forms of public and private data networks currently available. The basic features of the IWF are: Data rate adaption Protocol conversion
https://en.wikipedia.org/wiki/UK%20Atomic%20Energy%20Authority%20Constabulary
The United Kingdom Atomic Energy Authority Constabulary was the armed security police force of the United Kingdom Atomic Energy Authority. The force existed for 50 years, operating from 1955, until 1 April 2005. On 1 April 2005, the Civil Nuclear Constabulary (CNC) was established in adherence to the Energy Act 2004, replacing the UKAEA Constabulary largely because a number of nuclear sites were poorly guarded, a force with more specialist attributes was needed to combat the possibility of terrorist threats and it was seen as an essential part of making the force independent of any one of the nuclear operators. The force was made up of 650 armed personnel (the majority of whom were from the UKAEA Constabulary), protecting sixteen atomic sites and protecting transportation of nuclear materials around the United Kingdom and abroad. At that time UKAEAC/CNC came under the Department for Trade and Industry. See also Civil Nuclear Constabulary
https://en.wikipedia.org/wiki/Spectroradiometer
A spectroradiometer is a light measurement tool that is able to measure both the wavelength and amplitude of the light emitted from a light source. Spectrometers discriminate the wavelength based on the position the light hits at the detector array allowing the full spectrum to be obtained with a single acquisition. Most spectrometers have a base measurement of counts which is the un-calibrated reading and is thus impacted by the sensitivity of the detector to each wavelength. By applying a calibration, the spectrometer is then able to provide measurements of spectral irradiance, spectral radiance and/or spectral flux. This data is also then used with built in or PC software and numerous algorithms to provide readings or Irradiance (W/cm2), Illuminance (lux or fc), Radiance (W/sr), Luminance (cd), Flux (Lumens or Watts), Chromaticity, Color Temperature, Peak and Dominant Wavelength. Some more complex spectrometer software packages also allow calculation of PAR μmol/m2/s, Metamerism, and candela calculations based on distance and include features like 2- and 20-degree observer, baseline overlay comparisons, transmission and reflectance. Spectrometers are available in numerous packages and sizes covering many wavelength ranges. The effective wavelength (spectral) range of a spectrometer is determined not only by the grating dispersion ability but also depends on the detectors' sensitivity range. Limited by the semiconductor's band gap the silicon-based detector responds to 200-1100 nm while the InGaAs based detector is sensitive to 900-1700 nm (or out to 2500 nm with cooling). Lab/Research spectrometers often cover a broad spectral range from UV to NIR and require a PC. There are also IR Spectrometers that require higher power to run a cooling system. Many Spectrometers can be optimized for a specific range i.e. UV, or VIS and combined with a second system to allow more precise measurements, better resolution, and eliminate some of the more common errors found in br
https://en.wikipedia.org/wiki/Predictive%20adaptive%20response
A predictive adaptive response (PAR) is a developmental trajectory taken by an organism during a period of developmental plasticity in response to perceived environmental cues. This PAR does not confer an immediate advantage to the developing organism; however, if the PAR correctly anticipates the postnatal environment it will be advantageous in later life, if the environment the organism is born into differs from that anticipated by the PAR it will result in a mismatch. PAR mechanisms were first recognized in research done on human fetuses that investigated whether poor nutrition results in the inevitable diagnosis of Type 2 diabetes in later life. PARs are thought to occur through epigenetic mechanisms that alter gene expression, such as DNA methylation and histone modification, and do not involve changes to the DNA sequence of the developing organism. Examples of PARs include greater helmet development in Daphnia cucullata in response to maternal exposure to predator pheromones, rats exposed to glucocorticoid during late gestation led to an intolerance to glucose as adults, and coat thickness determination in vole pups by the photoperiod length experienced by the mother. Two hypotheses to explain PAR are the "thrifty phenotype" hypothesis and the developmental plasticity hypothesis. The thrifty phenotype hypothesis The thrifty phenotype hypothesis is the idea that if an organism suffers from inadequate nutrition in fetal development it will subsequently be predisposed to certain genetic outcomes as an adult. A study done examining glucose tolerance of individuals born during a famine in the Netherlands in 1944-1945 favors the “thrifty phenotype” hypothesis. The results of the experiment showed that exposure to famine, particularly in late gestation, led to a decrease in the glucose tolerance of the adults. Other studies on humans have shown cardiovascular and diabetes mortality has been shown to correspond to the nutrition uptake of the parents and grandparents
https://en.wikipedia.org/wiki/Jean%20Leray
Jean Leray (; 7 November 1906 – 10 November 1998) was a French mathematician, who worked on both partial differential equations and algebraic topology. Life and career He was born in Chantenay-sur-Loire (today part of Nantes). He studied at École Normale Supérieure from 1926 to 1929. He received his Ph.D. in 1933. In 1934 Leray published an important paper that founded the study of weak solutions of the Navier–Stokes equations. In the same year, he and Juliusz Schauder discovered a topological invariant, now called the Leray–Schauder degree, which they applied to prove the existence of solutions for partial differential equations lacking uniqueness. From 1938 to 1939 he was professor at the University of Nancy. He did not join the Bourbaki group, although he was close with its founders. His main work in topology was carried out while he was in a prisoner of war camp in Edelbach, Austria from 1940 to 1945. He concealed his expertise on differential equations, fearing that its connections with applied mathematics could lead him to be asked to do war work. Leray's work of this period proved seminal to the development of spectral sequences and sheaves. These were subsequently developed by many others, each separately becoming an important tool in homological algebra. He returned to work on partial differential equations from about 1950. He was professor at the University of Paris from 1945 to 1947, and then at the Collège de France until 1978. He was awarded the Malaxa Prize (Romania, 1938), the Grand Prix in mathematical sciences (French Academy of Sciences, 1940), the Feltrinelli Prize (Accademia dei Lincei, 1971), the Wolf Prize in Mathematics (Israel, 1979), and the Lomonosov Gold Medal (Moscow, 1988). He was an elected to the American Academy of Arts and Sciences and the American Philosophical Society in 1959 and the United States National Academy of Sciences in 1965. See also Leray–Schauder theorem
https://en.wikipedia.org/wiki/Sudanese%20goat%20marriage%20incident
In 2006, a South Sudanese man named Charles Tombe was forced to "marry" a goat with which he was caught engaging in sexual activity, in the Hai Malakal suburb of Juba, at the time part of Sudan. The owner of the goat subdued the perpetrator and asked village elders to consider the matter. One elder noted that he and the other elders found the perpetrator, tied up by the owner, at the door of the goat shed. The goat's owner reported that, "They said I should not take him to the police, but rather let him pay a dowry for my goat because he used it as his wife." The perpetrator was thus ordered to "marry" the goat, pay the cost of the goat and pay a dowry of SD 15,000 (equating to US$50 in 2006, the GDP per capita was US$1,522 for 2008), with half of the dowry up front. The goat apparently acquired the name "Rose" during the elders' deliberations as part of a joke. On 3 May 2007, it was reported that the goat had died, having choked on a plastic bag. The goat was survived by a four-month-old male kid. In November 2013, the South Sudan Law Society called for a review of all South Sudan's laws to abolish bizarre or cruel practices under customary law, such as "a man being forced to marry a goat called "Rose" after deflowering her." Press attention The story, first published on 24 February 2006 on the BBC website, attracted massive attention and was republished on numerous newspapers, blogs and other websites. Even a year after publication, the story was consistently among the BBC's 10 most emailed articles, with many visitors to the BBC news site passing the tale on to friends. The story received over 100,000 hits on five successive days long after its original publication, and was read by millions of people. The BBC, astonished at this popularity, wondered if there was a campaign to keep the tale at the top of its rankings; however, an investigation by its senior software engineer, Gareth Owen, determined that the demand was genuine. The BBC honoured the goat with
https://en.wikipedia.org/wiki/Reflection%20lines
Engineers use reflection lines to judge a surface's quality. Reflection lines reveal surface flaws, particularly discontinuities in normals indicating that the surface is not . Reflection lines may be created and examined on physical surfaces or virtual surfaces with the help of computer graphics. For example, the shiny surface of an automobile body is illuminated with reflection lines by surrounding the car with parallel light sources. Virtually, a surface can be rendered with reflection lines by modulating the surfaces point-wise color according to a simple calculation involving the surface normal, viewing direction and a square wave environment map. Mathematical definition Consider a point on a surface with (normalized) normal . If an observer views this point from infinity at view direction then the reflected view direction is: (The vector is decomposed into its normal part and tangential part . Upon reflection, the tangential part is kept and the normal part is negated.) For reflection lines we consider the surface surrounded by parallel lines with direction , representing infinite, non-dispersive light sources. For each point on we determine which line is seen from direction . The position on each line is of no interest. Define the vector to be the reflection direction projected onto a plane that is orthogonal to : and similarly let be the viewing direction projected onto : Finally, define to be the direction lying in perpendicular to and : Using these vectors, the *reflection line function* is a scalar function mapping points on the surface to angles between and : where is the atan2 function producing a number in the range . ( and can be viewed as a local coordinate system in with -axis in direction and -axis in direction .) Finally, to render the reflection lines positive values are mapped to a light color and non-positive values to a dark color. Highlight lines Highlight lines are a view-independent alternative to ref
https://en.wikipedia.org/wiki/Transcription%20activator-like%20effector%20nuclease
Transcription activator-like effector nucleases (TALEN) are restriction enzymes that can be engineered to cut specific sequences of DNA. They are made by fusing a TAL effector DNA-binding domain to a DNA cleavage domain (a nuclease which cuts DNA strands). Transcription activator-like effectors (TALEs) can be engineered to bind to practically any desired DNA sequence, so when combined with a nuclease, DNA can be cut at specific locations. The restriction enzymes can be introduced into cells, for use in gene editing or for genome editing in situ, a technique known as genome editing with engineered nucleases. Alongside zinc finger nucleases and CRISPR/Cas9, TALEN is a prominent tool in the field of genome editing. TALE DNA-binding domain TAL effectors are proteins that are secreted by Xanthomonas bacteria via their type III secretion system when they infect plants. The DNA binding domain contains a repeated highly conserved 33–34 amino acid sequence with divergent 12th and 13th amino acids. These two positions, referred to as the Repeat Variable Diresidue (RVD), are highly variable and show a strong correlation with specific nucleotide recognition. This straightforward relationship between amino acid sequence and DNA recognition has allowed for the engineering of specific DNA-binding domains by selecting a combination of repeat segments containing the appropriate RVDs. Notably, slight changes in the RVD and the incorporation of "nonconventional" RVD sequences can improve targeting specificity. DNA cleavage domain The non-specific DNA cleavage domain from the end of the FokI endonuclease can be used to construct hybrid nucleases that are active in a yeast assay. These reagents are also active in plant cells and in animal cells. Initial TALEN studies used the wild-type FokI cleavage domain, but some subsequent TALEN studies also used FokI cleavage domain variants with mutations designed to improve cleavage specificity and cleavage activity. The FokI domain function
https://en.wikipedia.org/wiki/Lippmann%20plate
Gabriel Lippmann conceived a two-step method to record and reproduce colours, variously known as direct photochromes, interference photochromes, Lippmann photochromes, Photography in natural colours by direct exposure in the camera or the Lippmann process of colour photography. Lippmann won the Nobel Prize in Physics for this work in 1908. A Lippmann plate is a clear glass plate (having no anti-halation backing), coated with an almost transparent (very low silver halide content) emulsion of extremely fine grains, typically 0.01 to 0.04 micrometres in diameter. Consequently, Lippmann plates have an extremely high resolving power exceeding 400 lines/mm. Method In Lippmann's method, a glass plate is coated with an ultra fine grain light-sensitive film using the Albumen Process containing potassium bromide, then dried, sensitized in the silver bath, washed, irrigated with cyanine solution, and dried again. The back of the film is then brought into optical contact with a reflective surface. This is done by mounting the plate in a specialized holder with pure mercury behind the film. When it is exposed in the camera through the glass side of the plate, the light rays which strike the transparent light-sensitive film are reflected back on themselves and, by interference, create standing waves. The standing waves cause exposure of the emulsion in diffraction patterns. The developed and fixated diffraction patterns constitute a Bragg condition in which diffuse, white light is scattered in a specular fashion and undergoes constructive interference in accordance to Bragg's law. The result is an image having very similar colours as the original using a black and white photographic process. For this method Lippmann won the Nobel Prize in Physics in 1908. The colour image can only be viewed in the reflection of a diffuse light source from the plate, making the field of view limited, and it cannot be copied. The technique was very insensitive with the emulsions of the time and
https://en.wikipedia.org/wiki/Asus%20Vivo
The Vivo is a lineup of portable computers developed by Asus. It consists of: laptops (VivoBooks) All-in-Ones (Vivo AiO) desktops (VivoPC) Stick PCs (VivoStick) Mini PCs (VivoMini) smartwatches (VivoWatch) computer mouse (VivoMouse) tablets (VivoTab). VivoBook Some Asus VivoBook models are branded under different series depending on regions and/or time. For example, the VivoBook E12 E203 used to be marketed under the VivoBook E Series but has since been marketed without 'E12' and under the Asus Laptop series. VivoBook 4K The Asus VivoBook 4K uses a 15.6" 16:9 IPS 4K (3840 x 2160) display with a color gamut of 72% NTSC, 100% sRGB, and 74% Adobe RGB. The laptop supports up to Intel Core i7 processor, up to 12GB of RAM, up to a 2TB HDD and up to a Nvidia 940M video card. The I/O consists of a combo audio jack, a VGA port, 2x USB 3.0 port(s), 1x USB 2.0 port(s), a RJ45 LAN Jack and a HDMI. VivoBook E Series The Asus VivoBook E Series is the successor to the EeeBook and Eee PC lineup of computers. Some of the VivoBook E Series laptops are simply rebadged EeeBook laptops such as the E402 and E202. The VivoBook E Series consists of the E200 (E200HA), E201 (E201NA), E202 (E202SA), E12 E203 (E203NAH and E203NA), E402 (E402SA, E402NA, E402BA and E402BP), E403 (E403SA and E403NA) and E502 (E502NA). E200 Reception Windows Central rated the E200HA 4 out of 5 concluded by stating that the E200H has a great design, good touchpad, runs quiet and cool and has good speakers but also commented that it has a bad display and oddly sized keyboard. pcverge gave the E200HA a rating of 74% commenting that it is very inexpensive, light and well-made build with excellent battery life but could be improved with better viewing angles, a better Keyboard and a larger touchpad. VivoBook F Series The VivoBook F Series consists of the F200 (F200MA, F200CA and F200LA), F450 ( F450CA and F450CC) and F550 (F550LD). VivoBook Max The VivoBook Max Series consists of the X441 (X441SA, X441UV,
https://en.wikipedia.org/wiki/Close%20stool
A close stool was an early type of portable toilet, made in the shape of a cabinet or box at sitting height with an opening in the top. The external structure contained a pewter or earthenware chamberpot to receive the user's excrement and urine when they sat on it; this was normally covered (closed) by a folding lid. "Stool" has two relevant meanings: as a type of seat and as human feces. Close stools were used from the Middle Ages (the Oxford English Dictionary gives the first citation as 1410) until the introduction of the indoor flush toilet. Other names Records of the English court mention the "close stool" and detail its construction. As an example, the furniture maker and upholsterer William Green made a "close stool" in August 1537 for the Lady Mary. The stool was upholstered with crimson velvet and a silkwoman, Mistress Margery Vaughan, provided crimson silk fringes and ribbons for its decoration. Green made a leather carrying case for the stool. Lady Jane Grey ordered crimson velvet to cover two close stools in July 1553. In Scotland, equivalent close stools appear in inventories and were sometimes called "dry stools" or "stools of ease". James V of Scotland and his daughter Mary, Queen of Scots, both owned silk canopies which were suspended from the ceiling over the stool. The close stool was sometimes called a necessary stool or a night stool. The eighteenth-century euphemism was convenience; the term was further euphemised in the nineteenth century with the term night commode, which John Gloag suggested may have derived its significance from a "balance night stool" described in Thomas Sheraton's Cabinet Dictionary (London, 1803). Sheraton's design was "made to have the appearance of a small commode standing upon legs; when it is used the seat part presses down to a proper height by the hand, and afterwards it rises by means of lead weights, hung to the seat, by lines passing over pulleys at each end, all which are enclosed in a case." This appears to
https://en.wikipedia.org/wiki/VME%20eXtensions%20for%20Instrumentation
VME eXtensions for instrumentation bus (VXI bus) refers to standards for automated test based upon VMEbus. VXI defines additional bus lines for timing and triggering as well as mechanical requirements and standard protocols for configuration, message-based communication, multi-chassis extension, and other features. In 2004, the 2eVME extension was added to the VXI bus specification, giving it a maximum data rate of 160 MB/s. The basic building block of a VXI system is the mainframe or chassis. This contains up to 13 slots into which various modules (instruments) can be added. The mainframe also contains all the power supply requirements for the rack and the instruments it contains. Instruments in the form of VXI Modules then fit the slots in the rack. VXI bus modules are typically 6U in height (see Eurocard) and C-size (unlike VME bus modules which are more commonly B-size). It is therefore possible to configure a system to meet a particular requirement by selecting the required instruments. The basic architecture of the instrument system is described in US patent 4,707,834. This patent was freely licensed by Tektronix to the VXIbus Consortium. The VXIbus grew from the VME bus specification, it was established in 1987 by Hewlett Packard (now Keysight Technologies), Racal Instruments (now Astronics Test Systems), Colorado Data Systems, Wavetek and Tektronix. VXI is promoted by the VXIbus Consortium, whose sponsor members are currently (in alphabetical order) Astronics Test Systems (formerly Racal Instruments), Bustec, Keysight Technologies, National Instruments, Teradyne, and VTI Instruments (formerly known as VXI Technology) . ZTEC Instruments] is a participating Executive Member. VXI's core market is in Military and Avionics Automatic Test Systems. The VXIplug&play Alliance specified additional hardware and software interoperability standards, such as the Virtual Instrument Software Architecture (VISA), although the alliance was eventually merged with the I
https://en.wikipedia.org/wiki/Root%20effect
The Root effect is a physiological phenomenon that occurs in fish hemoglobin, named after its discoverer R. W. Root. It is the phenomenon where an increased proton or carbon dioxide concentration (lower pH) lowers hemoglobin's affinity and carrying capacity for oxygen. The Root effect is to be distinguished from the Bohr effect where only the affinity to oxygen is reduced. Hemoglobins showing the Root effect show a loss of cooperativity at low pH. This results in the Hb-O2 dissociation curve being shifted downward and not just to the right. At low pH, hemoglobins showing the Root effect don't become fully oxygenated even at oxygen tensions up to 20kPa. This effect allows hemoglobin in fish with swim bladders to unload oxygen into the swim bladder against a high oxygen gradient. The effect is also noted in the choroid rete, the network of blood vessels which carries oxygen to the retina. In the absence of the Root effect, retia will result in the diffusion of some oxygen directly from the arterial blood to the venous blood, making such systems less effective for the concentration of oxygen. It has also been hypothesized that the loss of affinity is used to provide more oxygen to red muscle during acidotic stress.
https://en.wikipedia.org/wiki/Half-Life%20%28video%20game%29
Half-Life is a 1998 first-person shooter (FPS) game developed by Valve Corporation and published by Sierra Studios for Microsoft Windows. It was Valve's debut product and the first game in the Half-Life series. The player assumes the role of Gordon Freeman, a scientist who must escape the Black Mesa Research Facility after it is invaded by aliens following a disastrous scientific experiment. The gameplay consists of combat, exploration, and puzzle-solving. Valve was disappointed with the lack of innovation in the FPS genre, and aimed to create an immersive world rather than a "shooting gallery." Unlike other games of the time, the player has almost uninterrupted control of the player character, and the story is told mostly in scripted sequences rather than cutscenes. Valve developed it using GoldSrc, a heavily modified version of the Quake engine, licensed from id Software. The science fiction novelist Marc Laidlaw was hired to craft the plot and assist with design. Half-Life received acclaim for its graphics, gameplay, and narrative, and won more than 50 PC "Game of the Year" awards. It is considered one of the most influential FPS games and one of the greatest video games ever made. By 2008, it had sold more than nine million copies. It was followed by the expansion packs Opposing Force (1999) and Blue Shift (2001), developed by Gearbox Software. It was ported to the PlayStation 2 in 2001, along with the multiplayer expansion Decay, and to OS X and Linux in 2013. Valve ported Half-Life to its Source engine as Half-Life: Source in 2004. In 2020, Crowbar Collective released an unofficial remake, Black Mesa. Half-Life inspired numerous fan-made mods, some of which became standalone games, such as Counter-Strike, Day of Defeat, and Sven Co-op. It was followed by Half-Life 2 (2004), Half-Life 2: Episode One (2006), Half-Life 2: Episode Two (2007), and Half-Life: Alyx (2020). Gameplay Half-Life is a first-person shooter (FPS) that requires the player to perform co
https://en.wikipedia.org/wiki/CARD9
Caspase recruitment domain-containing protein 9 is an adaptor protein of the CARD-CC protein family, which in humans is encoded by the CARD9 gene. It mediates signals from pattern recognition receptors to activate pro-inflammatory and anti-inflammatory cytokines, regulating inflammation. Homozygous mutations in CARD9 are associated with defective innate immunity against yeasts, like Candida and dermatophytes. Function CARD9 is a member of the CARD protein family, which is defined by the presence of a characteristic caspase-associated recruitment domain (CARD). This protein was identified by its selective association with the CARD domain of BCL10, a positive regulator and NF-κB activation. It is thought to function as a molecular scaffold for the assembly of a BCL10 signaling complex that activates NF-κB. Several alternatively spliced transcript variants have been observed, but their full-length nature is not clearly defined. Clinical significance In 2006, it became clear that Card9 plays important roles within the innate immune response against yeasts. Card9 mediates signals from so called pattern recognition receptors (Dectin-1) to downstream signalling pathways such as NF-κB and by this activates pro-inflammatory cytokines (TNF, IL-23, IL-6, IL-2) and an anti-inflammatory cytokine (IL-10) and subsequently an appropriate innate and adaptive immune response to clear an infection. An autosomal recessive form of susceptibility to chronic mucocutaneous candidiasis was found in 2009 to be associated with homozygous mutations in CARD9. Deep dermatophytosis and Card9 deficiency reported in an Iranian family led to its discovery in 17 people from Tunisian, Algerian, and Moroccan families with deep dermatophytosis. CARD9 mutations have been associated with inflammatory diseases such as ankylosing spondylitis and inflammatory bowel disease (Crohn's Disease and Ulcerative Colitis). A genetic variant, c.IVS11+1G>C was found to be protective against crohn's disease, ul
https://en.wikipedia.org/wiki/Union%20%28set%20theory%29
In set theory, the union (denoted by ∪) of a collection of sets is the set of all elements in the collection. It is one of the fundamental operations through which sets can be combined and related to each other. A refers to a union of zero () sets and it is by definition equal to the empty set. For explanation of the symbols used in this article, refer to the table of mathematical symbols. Union of two sets The union of two sets A and B is the set of elements which are in A, in B, or in both A and B. In set-builder notation, . For example, if A = {1, 3, 5, 7} and B = {1, 2, 4, 6, 7} then A ∪ B = {1, 2, 3, 4, 5, 6, 7}. A more elaborate example (involving two infinite sets) is: A = {x is an even integer larger than 1} B = {x is an odd integer larger than 1} As another example, the number 9 is not contained in the union of the set of prime numbers {2, 3, 5, 7, 11, ...} and the set of even numbers {2, 4, 6, 8, 10, ...}, because 9 is neither prime nor even. Sets cannot have duplicate elements, so the union of the sets {1, 2, 3} and {2, 3, 4} is {1, 2, 3, 4}. Multiple occurrences of identical elements have no effect on the cardinality of a set or its contents. Algebraic properties Binary union is an associative operation; that is, for any sets Thus, the parentheses may be omitted without ambiguity: either of the above can be written as Also, union is commutative, so the sets can be written in any order. The empty set is an identity element for the operation of union. That is, for any set Also, the union operation is idempotent: All these properties follow from analogous facts about logical disjunction. Intersection distributes over union and union distributes over intersection The power set of a set together with the operations given by union, intersection, and complementation, is a Boolean algebra. In this Boolean algebra, union can be expressed in terms of intersection and complementation by the formula where the superscript denotes the
https://en.wikipedia.org/wiki/Modal%20algebra
In algebra and logic, a modal algebra is a structure such that is a Boolean algebra, is a unary operation on A satisfying and for all x, y in A. Modal algebras provide models of propositional modal logics in the same way as Boolean algebras are models of classical logic. In particular, the variety of all modal algebras is the equivalent algebraic semantics of the modal logic K in the sense of abstract algebraic logic, and the lattice of its subvarieties is dually isomorphic to the lattice of normal modal logics. Stone's representation theorem can be generalized to the Jónsson–Tarski duality, which ensures that each modal algebra can be represented as the algebra of admissible sets in a modal general frame. A Magari algebra (or diagonalizable algebra) is a modal algebra satisfying . Magari algebras correspond to provability logic. See also Interior algebra Heyting algebra
https://en.wikipedia.org/wiki/Two-variable%20logic
In mathematical logic and computer science, two-variable logic is the fragment of first-order logic where formulae can be written using only two different variables. This fragment is usually studied without function symbols. Decidability Some important problems about two-variable logic, such as satisfiability and finite satisfiability, are decidable. This result generalizes results about the decidability of fragments of two-variable logic, such as certain description logics; however, some fragments of two-variable logic enjoy a much lower computational complexity for their satisfiability problems. By contrast, for the three-variable fragment of first-order logic without function symbols, satisfiability is undecidable. Counting quantifiers The two-variable fragment of first-order logic with no function symbols is known to be decidable even with the addition of counting quantifiers, and thus of uniqueness quantification. This is a more powerful result, as counting quantifiers for high numerical values are not expressible in that logic. Counting quantifiers actually improve the expressiveness of finite-variable logics as they allow to say that there is a node with neighbors, namely . Without counting quantifiers variables are needed for the same formula. Connection to the Weisfeiler-Leman algorithm There is a strong connection between two-variable logic and the Weisfeiler-Leman (or color refinement) algorithm. Given two graphs, then any two nodes have the same stable color in color refinement if and only if they have the same type, that is, they satisfy the same formulas in two-variable logic with counting.
https://en.wikipedia.org/wiki/Dinosaur%20tooth
Dinosaur teeth have been studied since 1822 when Mary Ann Mantell (1795-1869) and her husband Dr Gideon Algernon Mantell (1790-1852) discovered an Iguanodon tooth in Sussex in England. Unlike mammal teeth, individual dinosaur teeth are generally not considered by paleontologists to be diagnostic to the genus or species level for unknown taxa, due morphological convergence and variability between teeth. and many historically named tooth taxa like Paronychodon and Richardoestesia are today considered nomina dubia, and are used as form taxa to refer to isolated teeth from other localities displaced considerably in time and space from the type specimens. However, it is possible to refer isolated teeth to known taxa provided that the tooth morphology is known and the teeth originate from a similar time and place. Some of the most important anatomical information about dinosaur teeth is collected from polished, microscopically thin sections (histology), including the types of dental tissues present, tooth wear, tooth replacement patterns, how the teeth are attached, and the frequency of replacement. The actual material comprising a dinosaur tooth is not very different to teeth in modern animals. Most significant differences are in how the teeth fit together and continually regrew, with some examples shedding old teeth and others reabsorbing old teeth as they would grind down under chewing throughout a dinosaurs life. Background The use of histology in paleontology has traditionally been more focused on examining long bones such as the femur or the humerus. Previous work on long bone histology revealed differences in the growth patterns of polar dinosaurs, identified a case of dwarfism in Europasaurus, reconstructed the life history of Dysalotosaurus by examining multiple specimens of different ontogenetic stages, and suggested that Psittacosaurus underwent a postural change from a quadruped to biped as it matured. By contrast, dental histology has not been looked at
https://en.wikipedia.org/wiki/Flash%20mob%20computing
Flash mob computing or flash mob computer is a temporary ad hoc computer cluster running specific software to coordinate the individual computers into one single supercomputer. A flash mob computer is distinct from other types of computer clusters in that it is set up and broken down on the same day or during a similar brief amount of time and involves many independent owners of computers coming together at a central physical location to work on a specific problem and/or social event. Flash mob computer derives its name from the more general term flash mob which can mean any activity involving many people co-ordinated through virtual communities coming together for brief periods of time for a specific task or event. Flash mob computing is a more specific type of flash mob for the purpose of bringing people and their computers together to work on a single task or event. History The first flash mob computer was created on April 3, 2004 at the University of San Francisco using software written at USF called FlashMob (not to be confused with the more general term flash mob). The event, called FlashMob I, was a success. There was a call for computers on the computer news website Slashdot. An article in The New York Times "Hey, Gang, Let’s Make Our Own Supercomputer" brought a lot of attention to the effort. More than 700 computers were brought to the gym at the University of San Francisco, and were wired to a network donated by Foundry Networks. At FlashMob I the participants were able to run a benchmark on 256 of the computers, and achieved a peak rate of 180 Gflops (billions of calculations per second), though this computation stopped three quarters of the way due to a node failure. The best, complete run used 150 computers and resulted in 77 Gflops. FlashMob I was run off a bootable CD-ROM that ran a copy of Morphix Linux, which was only available for the x86 platform. Despite these efforts, the project was unable to achieve its original goal of running a clus
https://en.wikipedia.org/wiki/Entomophthoramycosis
Entomophthoramycosis (or Entomophthoromycosis) is a mycosis caused by Entomophthorales. Examples include basidiobolomycosis and conidiobolomycosis. Signs and symptoms Diagnosis A culture of the infected tissue of the individual suspected of having Entomophthoramycosis Treatment Treatment for phycomycosis is very difficult and includes surgery when possible. Postoperative recurrence is common. Antifungal drugs show only limited effect on the disease, but itraconazole and terbinafine hydrochloride are often used for two to three months following surgery. Humans with Basidiobolus infections have been treated with amphotericin B and potassium iodide. For pythiosis and lagenidiosis, a new drug targeting water moulds called caspofungin is available, but it is very expensive. Immunotherapy has been used successfully in humans and horses with pythiosis. Treatment for skin lesions is traditionally with potassium iodide, but itraconazole has also been used successfully.
https://en.wikipedia.org/wiki/CYP710%20family
Cytochrome P450, family 710, also known as CYP710, is a plant cytochrome P450 monooxygenase family, the proteins encoded by its family members are mainly sterol 22-desaturase, which was widely distributed in plants, and take participate in Phytosteroidogenesis. CYP710 family is considered to be the plant orthologous of fungi CYP61 family, which is lost in animal. The CYP61/CYP710 ancestor gene diverged from a gene duplication of ancestor CYP51 in early eukaryotes
https://en.wikipedia.org/wiki/Quantum%20%28video%20game%29
Quantum is a color vector arcade video game developed at General Computer Corporation for Atari, Inc. and released in December 1982. It was designed by Betty Ryan () who was the first female developer at GCC. The premise of the game is related loosely to quantum physics; the player directs a probe with a trackball to encircle atomic "particles" for points, without touching various other particles. Once the particles are surrounded by the probe's tail they are destroyed. Gameplay High score table To enter initials for a high score, the player uses the trackball to circle letters in the same fashion used during gameplay. If the player achieves the highest score on the table, the initials screen is preceded by another on which adept players can use the trackball to draw their initials. Legacy A screenshot of a clone called Tachyon was previewed in Atari 8-bit family magazine ANALOG Computing, but the game was never completed. A remake, Quantum Recharged, was released in August 2023 for the Nintendo Switch, PlayStation 4, PlayStation 5, Windows, Xbox One, and Xbox Series X/S. It is part of Atari SA's Recharged remake series. See also Disco No. 1 Libble Rabble
https://en.wikipedia.org/wiki/Mesoamerican%20Society%20for%20Ecological%20Economics
The Mesoamerican Society for Ecological Economics (SMEE) is a regional chapter of the International Society for Ecological Economics (ISEE). After its foundation in 2008 at Guatemala City, the organization has already celebrated its first International Conference in 2010 at Mexico City and will carry out the second International Conference, EcoEco Alternatives, between March 4 and 8 2014 at the main campus of the University of Costa Rica. This branch of the ISEE has a unique emphasis within ecological economics. Topics like social justice and the human value in environmental conservation prevail in this region. As a consequence of the strong influence from Joan Martinez Alier's "environmentalism of the poor or social environmentalism", major attention is given to ecological-distributive conflicts. Alier insists that in the South a struggle exists against these conflicts generated by economic growth, mainly by the North. These endeavors "attempt to preserve the access of the communities to natural resources and services." On top of the negative effects on the environment by economic distribution, the cultural influence is also widely debated. For instance, the anthropologist Arturo Escobar suggests that culturally-driven preferences are one of the main factors degrading the environment. For example, society naturally gives privilege to the capitalist model that distributes natural resources with the purposes of production and profit, instead of endorsing the agroforestal ecosystem model, which is less harmful to the environment. As part of this alternate perception in Mesoamerica, Ecological economics doesn't consider that the economic valuation of natural resources nor environmental norms are effective solutions to these social-environmental conflicts. On the other hand, an alternative based on community-based conservation and the management of sustainability is more advocated upon. By adding the latter cultural perspective, the three pillars of sustainable devel
https://en.wikipedia.org/wiki/Dancing%20plague%20of%201518
The dancing plague of 1518, or dance epidemic of 1518 (), was a case of dancing mania that occurred in Strasbourg, Alsace (modern-day France), in the Holy Roman Empire from July 1518 to September 1518. Somewhere between 50 and 400 people took to dancing for weeks. There are many theories behind the phenomenon, the most popular being stress-induced mass hysteria, suggested by John Waller. Other theories include ergot and religious explanations. There is controversy concerning the number of deaths. Events The outbreak began in July 1518 when a woman called Frau Troffea began to dance fervently in a street in Strasbourg. Troffea kept up the constant dancing for a week. Soon, three dozen others joined in. By August, the "dancing plague" had claimed 400 victims. Dancers were beginning to collapse. It is said some even died from a stroke or heart attack. No one knew what caused this reaction, which meant no one understood how to remedy it. By early September, the outbreak began to subside, when the dancers were sent to a mountain shrine to pray for absolution. Historical documents, including "physician notes, cathedral sermons, local and regional chronicles, and even notes issued by the Strasbourg city council" are clear that the victims danced; it is not known why. Historical sources agree that there was an outbreak of dancing after a single woman started dancing, and the dancing did not seem to die down. It lasted for such a long time that it even attracted the attention of the authorities; until the council gave up authority to the physicians, who prescribed the afflicted to "dance themselves free of it." There are claims that guild halls were refurbished to accommodate the dancing, as well as musicians and strong people to help keep those dealing with the dancing mania to stay upright. This backfired, and the council was forced to ban public dancing as people danced in fear it was a punishment from Saint Vitus; and to be "free of sin" many joined in on the dancing
https://en.wikipedia.org/wiki/Glycinamide
Glycinamide is a organic compound with the molecular formula H2NCH2C(O)NH2. It is the amide derivative of the amino acid glycine. It is a water-soluble, white solid. Amino acid amides, such as glycinamide are prepared by treating the amino acid ester with ammonia. It is a ligand for transition metals. The hydrochloride salt of glycinamide, glycinamide hydrochloride, is one of Good's buffers with a pH in the physiological range. Glycinamide hydrochloride has a pKa near the physiological pH (8.20 at 20°C), making it useful in cell culture work. Its ΔpKa/°C is -0.029 and it has a solubility in water at 0 °C of 6.4 M. Glycinamide is a reagent used in the synthesis of glycineamide ribonucleotide (an intermediate in de novo purine biosynthesis).
https://en.wikipedia.org/wiki/Hooley%27s%20delta%20function
In mathematics, Hooley's delta function (), also called Erdős--Hooley delta-function, defines the maximum number of divisors of in for all , where is the Euler's number. The first few terms of this sequence are . History The sequence was first introduced by Paul Erdős in 1974, then studied by Christopher Hooley in 1979. In 2023, Dimitris Koukoulopoulos and Terence Tao proved that the sum of the first terms, , for . In particular, the average order of to is for any . Later in 2023 Kevin Ford, Koukoulopoulos, and Tao proved the lower bound , where , fixed , and . Usage This function measures the tendency of divisors of a number to cluster. The growth of this sequence is limited by where is the number of divisors of . See also Divisor function Euler's number
https://en.wikipedia.org/wiki/Segre%27s%20theorem
In projective geometry, Segre's theorem, named after the Italian mathematician Beniamino Segre, is the statement: Any oval in a finite pappian projective plane of odd order is a nondegenerate projective conic section. This statement was assumed 1949 by the two Finnish mathematicians G. Järnefelt and P. Kustaanheimo and its proof was published in 1955 by B. Segre. A finite pappian projective plane can be imagined as the projective closure of the real plane (by a line at infinity), where the real numbers are replaced by a finite field . Odd order means that is odd. An oval is a curve similar to a circle (see definition below): any line meets it in at most 2 points and through any point of it there is exactly one tangent. The standard examples are the nondegenerate projective conic sections. In pappian projective planes of even order greater than four there are ovals which are not conics. In an infinite plane there exist ovals, which are not conics. In the real plane one just glues a half of a circle and a suitable ellipse smoothly. The proof of Segre's theorem, shown below, uses the 3-point version of Pascal's theorem and a property of a finite field of odd order, namely, that the product of all the nonzero elements equals -1. Definition of an oval In a projective plane a set of points is called oval, if: (1) Any line meets in at most two points. If the line is an exterior (or passing) line; in case a tangent line and if the line is a secant line. (2) For any point there exists exactly one tangent at , i.e., . For finite planes (i.e. the set of points is finite) we have a more convenient characterization: For a finite projective plane of order (i.e. any line contains points) a set of points is an oval if and only if and no three points are collinear (on a common line). Pascal's 3-point version Theorem Let be an oval in a pappian projective plane of characteristic . is a nondegenerate conic if and only if statement (P3) holds: (P3): Let
https://en.wikipedia.org/wiki/Terminal%20and%20nonterminal%20symbols
In formal languages, terminal and nonterminal symbols are the lexical elements used in specifying the production rules constituting a formal grammar. Terminal symbols are the elementary symbols of the language defined as part of a formal grammar. Nonterminal symbols (or syntactic variables) are replaced by groups of terminal symbols according to the production rules. The terminals and nonterminals of a particular grammar are in two completely separate sets. Terminal symbols Terminal symbols are symbols that may appear in the outputs of the production rules of a formal grammar and which cannot be changed using the rules of the grammar. Applying the rules recursively to a source string of symbols will usually terminate in a final output string consisting only of terminal symbols. Consider a grammar defined by two rules. In this grammar, the symbol Б is a terminal symbol and Ψ is both a non-terminal symbol and the start symbol. The production rules for creating strings are as follows: The symbol Ψ can become БΨ The symbol Ψ can become Б Here Б is a terminal symbol because no rule exists which would change it into something else. On the other hand, Ψ has two rules that can change it, thus it is nonterminal. A formal language defined or generated by a particular grammar is the set of strings that can be produced by the grammar and that consist only of terminal symbols. Diagram 1 illustrates a string that can be produced with this grammar. Nonterminal symbols Nonterminal symbols are those symbols that can be replaced. They may also be called simply syntactic variables. A formal grammar includes a start symbol, a designated member of the set of nonterminals from which all the strings in the language may be derived by successive applications of the production rules. In fact, the language defined by a grammar is precisely the set of terminal strings that can be so derived. Context-free grammars are those grammars in which the left-hand side of each production
https://en.wikipedia.org/wiki/Flag%20of%20Christmas%20Island
The flag of Christmas Island was unofficially adopted in 1986 after being chosen the winner in a competition for a flag for the territory. It was designed by Tony Couch of Sydney, Australia. The flag was made official in 2002 on Australia Day, when the administrator of the territory, Bill Taylor, presented the flag to the Christmas Island Shire. Design The flag of Christmas Island consists of a green and blue background, split from the top left corner to the bottom right. These colours are intended to represent the land and sea respectively. The Southern Cross constellation appears in the bottom left of the flag similar to the flag of Australia. In the top right, the golden bosun bird (Phaethon lepturus fulvus, one of six subspecies of the white-tailed tropicbird) appears. It is considered to be a symbol of the Island. The last motif appears in the centre of the flag on a golden disc is the map of the island in green. The disc itself was originally only included to offset the green colour of the map, but has become linked to the mining industry. History Creation In 1986, the Christmas Island Assembly announced a competition to design both a flag and a coat of arms for the territory. There was a prize fund of $100, and some 69 entries were submitted. The winning submission was created by Tony Couch, a resident of Sydney who had previously worked on Christmas Island. The new flag was announced on 14 April 1986, by the Christmas Island Assembly. Implementation The first attempt to make the flag official occurred in 1995 when the Minister of the Islands at the time took the view that implementation could take place on Australia Day 1996 via a formal announcement by the Administrator rather than an amendment to the Christmas Island Act 1958. Although this was agreed, the declaration never took place. Subsequently, Christmas Island official Gary Dunt revived the issue in 2001 and the flag was formally declared the official flag of Christmas Island on Australia Day 2
https://en.wikipedia.org/wiki/Distonic%20ion
Distonic ions are chemical species that contain two ionic charges on the same molecule, separated by two or more carbon or heteroatoms. A feature of distonic radical ions is that their charges and radical sites are in different locations (on separate atoms), unlike regular radicals where the formal charge and unpaired electron are in the same location. These molecular species are created by ionization of either zwitterions or diradicals; ultimately, a neutral molecule loses an electron. Through experimental research distonic radicals have been found to be extremely stable gas phase ions and can be separated into different classes depending on the inherent features of the charged portion of the ion. History In 1984 scientists Bouma, Radom and Yates originated the term through extensive experimental research but they were not the first to deal with distonic ions. Experiments date back to the 1970s with Gross and McLafferty who were the first to propose the idea of such a species. Ion structure Several efficient techniques are available to detect the presence of distonic ions; the most appropriate method will depend on the ion's internal energy and lifespan. Collisions between ions and uncharged molecules allow one to detect the location of the radical and charge site in order to confirm that the ion is not just a regular radical ion. When a molecule is ionized and can structurally be classified as a distonic ion, the molecule's kinetics and thermodynamic properties have been greatly altered. However, additional chemical properties are based on the reactions of the central excited ions. Mass spectrometry techniques are used to study their chemistry. Experimental data Distonic ions have been extensively examined due to their unique behavior and how commonly they can occur. It has been shown that in most cases distonic ions have a bonding arrangement corresponding to that of the original molecule before ionization occurred; but that distonic ions are less stable t
https://en.wikipedia.org/wiki/Displacement%E2%80%93length%20ratio
The displacement–length ratio (DLR or D/L ratio) is a calculation used to express how heavy a boat is relative to its waterline length. DLR was first published in It is calculated by dividing a boat's displacement in long tons (2,240 pounds) by the cube of one one-hundredth of the waterline length (in feet): DLR can be used to compare the relative mass of various boats no matter what their length. A DLR less than 200 is indicative of a racing boat, while a DLR greater than 300 or so is indicative of a heavy cruising boat. See also Sail Area-Displacement ratio
https://en.wikipedia.org/wiki/Trolox%20equivalent%20antioxidant%20capacity
The Trolox equivalent antioxidant capacity (TEAC) assay measures the antioxidant capacity of a given substance, as compared to the standard, Trolox. Most commonly, antioxidant capacity is measured using the ABTS Decolorization Assay. Other antioxidant capacity assays which use Trolox as a standard include the diphenylpicrylhydrazyl (DPPH), oxygen radical absorbance capacity (ORAC) and ferric reducing ability of plasma (FRAP) assays. The TEAC assay is often used to measure the antioxidant capacity of foods, beverages and nutritional supplements.
https://en.wikipedia.org/wiki/Symplectic%20integrator
In mathematics, a symplectic integrator (SI) is a numerical integration scheme for Hamiltonian systems. Symplectic integrators form the subclass of geometric integrators which, by definition, are canonical transformations. They are widely used in nonlinear dynamics, molecular dynamics, discrete element methods, accelerator physics, plasma physics, quantum physics, and celestial mechanics. Introduction Symplectic integrators are designed for the numerical solution of Hamilton's equations, which read where denotes the position coordinates, the momentum coordinates, and is the Hamiltonian. The set of position and momentum coordinates are called canonical coordinates. (See Hamiltonian mechanics for more background.) The time evolution of Hamilton's equations is a symplectomorphism, meaning that it conserves the symplectic 2-form . A numerical scheme is a symplectic integrator if it also conserves this 2-form. Symplectic integrators also might possess, as a conserved quantity, a Hamiltonian which is slightly perturbed from the original one (only true for a small class of simple cases). By virtue of these advantages, the SI scheme has been widely applied to the calculations of long-term evolution of chaotic Hamiltonian systems ranging from the Kepler problem to the classical and semi-classical simulations in molecular dynamics. Most of the usual numerical methods, like the primitive Euler scheme and the classical Runge–Kutta scheme, are not symplectic integrators. Methods for constructing symplectic algorithms Splitting methods for separable Hamiltonians A widely used class of symplectic integrators is formed by the splitting methods. Assume that the Hamiltonian is separable, meaning that it can be written in the form This happens frequently in Hamiltonian mechanics, with T being the kinetic energy and V the potential energy. For the notational simplicity, let us introduce the symbol to denote the canonical coordinates including both the position and
https://en.wikipedia.org/wiki/Universal%20gateway
A universal gateway is a device that transacts data between two or more data sources using communication protocols specific to each. Sometimes called a universal protocol gateway, this class of product is designed as a computer appliance, and is used to connect data from one automation system to another. Typical applications Typical applications include: M2M Communications – machine to machine communications between machines from different vendors, typically using different communication protocols. This is often a requirement to optimize the performance of a production line, by effectively communicating machine states upstream and downstream of a piece of equipment. Machine idle times can trigger lower power operation. Inventory Levels can be more effectively managed on a per station basis, by knowing the upstream and downstream demands. M2E Communications – machine to enterprise communications is typically managed through database interactions. In this case, EATM technology is typically leveraged for data interoperability. However, many enterprise systems have real-time data interfaces. When real-time interfaces are involved, a universal gateway, with its ability to support many protocols simultaneously becomes the best choice. In all cases, communications can fall over many different transports, RS-232, RS-485, Ethernet, etc. Universal Gateways have the ability to communicate between protocols and over different transports simultaneously. Design Hardware platform – Industrial Computer, Embedded Computer, Computer Appliance Communications software – Software (Drivers) to support one or more Industrial Protocols. Communications is typically polled or change based. Great care is typically taken to leverage communication protocols for the most efficient transactions of data (Optimized message sizes, communications speeds, and data update rates). Typical protocols; Rockwell Automation CIP, Ethernet/IP, Siemens Industrial Ethernet, Modbus TCP. There
https://en.wikipedia.org/wiki/Evolutionary%20origin%20of%20religion
The evolutionary origin of religion and religious behavior is a field of study related to evolutionary psychology, the origin of language and mythology, and cross-cultural comparison of the anthropology of religion. Some subjects of interest include Neolithic religion, evidence for spirituality or cultic behavior in the Upper Paleolithic, and similarities in great ape behavior. Nonhuman religious behavior Humanity's closest living relatives are common chimpanzees and bonobos. These primates share a common ancestor with humans who lived between six and eight million years ago. It is for this reason that chimpanzees and bonobos are viewed as the best available surrogate for this common ancestor. Barbara King argues that while non-human primates are not religious, they do exhibit some traits that would have been necessary for the evolution of religion. These traits include high intelligence, a capacity for symbolic communication, a sense of social norms, and realization of "self" continuity. There is inconclusive evidence that Homo neanderthalensis may have buried their dead, which would be evidence of mortuary ritual. The use of burial rituals is thought to be evidence of religious activity, but there is no other evidence that religion existed in human culture before humans reached behavioral modernity. Other lines of evidence have revealed that Homo neanderthalensis made cave art, which would require a high level of symbolic thinking paralleling religious thought. Elephants perform rituals for their dead. They demonstrate long periods of silence and mourning at the point of death; later, elephants return to grave sites and caress the remains. Some evidence suggests that many species grieve death and loss. Relevant prerequisites for human religion Increased brain size In this set of theories, the religious mind is one consequence of a brain that is large enough to formulate religious and philosophical ideas. During human evolution, the hominid brain tripled in si
https://en.wikipedia.org/wiki/H-vector
In algebraic combinatorics, the h-vector of a simplicial polytope is a fundamental invariant of the polytope which encodes the number of faces of different dimensions and allows one to express the Dehn–Sommerville equations in a particularly simple form. A characterization of the set of h-vectors of simplicial polytopes was conjectured by Peter McMullen and proved by Lou Billera and Carl W. Lee and Richard Stanley (g-theorem). The definition of h-vector applies to arbitrary abstract simplicial complexes. The g-conjecture stated that for simplicial spheres, all possible h-vectors occur already among the h-vectors of the boundaries of convex simplicial polytopes. It was proven in December 2018 by Karim Adiprasito. Stanley introduced a generalization of the h-vector, the toric h-vector, which is defined for an arbitrary ranked poset, and proved that for the class of Eulerian posets, the Dehn–Sommerville equations continue to hold. A different, more combinatorial, generalization of the h-vector that has been extensively studied is the flag h-vector of a ranked poset. For Eulerian posets, it can be more concisely expressed by means of a noncommutative polynomial in two variables called the cd-index. Definition Let Δ be an abstract simplicial complex of dimension d − 1 with fi i-dimensional faces and f−1 = 1. These numbers are arranged into the f-vector of Δ, An important special case occurs when Δ is the boundary of a d-dimensional convex polytope. For k = 0, 1, …, d, let The tuple is called the h-vector of Δ. In particular, , , and , where is the Euler characteristic of . The f-vector and the h-vector uniquely determine each other through the linear relation from which it follows that, for , In particular, . Let R = k[Δ] be the Stanley–Reisner ring of Δ. Then its Hilbert–Poincaré series can be expressed as This motivates the definition of the h-vector of a finitely generated positively graded algebra of Krull dimension d as the numerator of i
https://en.wikipedia.org/wiki/Hadamard%20regularization
In mathematics, Hadamard regularization (also called Hadamard finite part or Hadamard's partie finie) is a method of regularizing divergent integrals by dropping some divergent terms and keeping the finite part, introduced by . showed that this can be interpreted as taking the meromorphic continuation of a convergent integral. If the Cauchy principal value integral exists, then it may be differentiated with respect to to obtain the Hadamard finite part integral as follows: Note that the symbols and are used here to denote Cauchy principal value and Hadamard finite-part integrals respectively. The Hadamard finite part integral above (for ) may also be given by the following equivalent definitions: The definitions above may be derived by assuming that the function is differentiable infinitely many times at , that is, by assuming that can be represented by its Taylor series about . For details, see . (Note that the term in the second equivalent definition above is missing in but this is corrected in the errata sheet of the book.) Integral equations containing Hadamard finite part integrals (with unknown) are termed hypersingular integral equations. Hypersingular integral equations arise in the formulation of many problems in mechanics, such as in fracture analysis. Example Consider the divergent integral Its Cauchy principal value also diverges since To assign a finite value to this divergent integral, we may consider The inner Cauchy principal value is given by Therefore, Note that this value does not represent the area under the curve , which is clearly always positive.
https://en.wikipedia.org/wiki/Gaussian%20q-distribution
In mathematical physics and probability and statistics, the Gaussian q-distribution is a family of probability distributions that includes, as limiting cases, the uniform distribution and the normal (Gaussian) distribution. It was introduced by Diaz and Teruel. It is a q-analog of the Gaussian or normal distribution. The distribution is symmetric about zero and is bounded, except for the limiting case of the normal distribution. The limiting uniform distribution is on the range -1 to +1. Definition Let q be a real number in the interval [0, 1). The probability density function of the Gaussian q-distribution is given by where The q-analogue [t]q of the real number is given by The q-analogue of the exponential function is the q-exponential, E, which is given by where the q-analogue of the factorial is the q-factorial, [n]q!, which is in turn given by for an integer n > 2 and [1]q! = [0]q! = 1. The cumulative distribution function of the Gaussian q-distribution is given by where the integration symbol denotes the Jackson integral. The function Gq is given explicitly by where Moments The moments of the Gaussian q-distribution are given by where the symbol [2n − 1]!! is the q-analogue of the double factorial given by See also Q-Gaussian process
https://en.wikipedia.org/wiki/Strain%20%28mechanics%29
In mechanics, strain is defined as relative deformation, compared to a position configuration. Different equivalent choices may be made for the expression of a strain field depending on whether it is defined with respect to the initial or the final configuration of the body and on whether the metric tensor or its dual is considered. Strain has dimension of a length ratio, with SI base units of meter per meter (m/m). Hence strains are dimensionless and are usually expressed as a decimal fraction or a percentage. Parts-per notation is also used, e.g., parts per million or parts per billion (sometimes called "microstrains" and "nanostrains", respectively), corresponding to μm/m and nm/m. Strain can be formulated as the spatial derivative of displacement: where is the identity tensor. The displacement of a body may be expressed in the form , where is the reference position of material points of the body; displacement has units of length and does not distinguish between rigid body motions (translations and rotations) and deformations (changes in shape and size) of the body. The spatial derivative of a uniform translation is zero, thus strains measure how much a given displacement differs locally from a rigid-body motion. A strain is in general a tensor quantity. Physical insight into strains can be gained by observing that a given strain can be decomposed into normal and shear components. The amount of stretch or compression along material line elements or fibers is the normal strain, and the amount of distortion associated with the sliding of plane layers over each other is the shear strain, within a deforming body. This could be applied by elongation, shortening, or volume changes, or angular distortion. The state of strain at a material point of a continuum body is defined as the totality of all the changes in length of material lines or fibers, the normal strain, which pass through that point and also the totality of all the changes in the angle between pa
https://en.wikipedia.org/wiki/Rainbow%20flag
A rainbow flag is a multicolored flag consisting of the colors of the rainbow. The designs differ, but many of the colors are based on the seven spectral colors of the visible light spectrum. History In the 18th century, American Revolutionary War writer Thomas Paine proposed that a rainbow flag be used as a maritime flag to signify neutral ships in time of war. Contemporary international uses of a rainbow flag dates to the beginning of the 20th century. The International Co-operative Alliance adopted a rainbow flag in 1925. A similar flag (ca. 1920) is used in Andean indigenism in Peru and Bolivia to represent the legacy of the Inca Empire. Since 1961, the international peace flag, also known as the PACE flag, has been especially popular in Italy and to a lesser extent Europe and the rest of the world. Since 1978, when it was created to represent gay pride, the rainbow pride flag has evolved as a symbol of the LGBT movement (in 1978, the community as a whole was referred to and described as the "Gay" community; the term "LGBT" did not yet exist.) There are several independent rainbow flags in use today. Rainbow flags in various cultures and movements Reformation (1525) The reformer Thomas Müntzer (1489–1525) connected socially revolutionary claims with his religious preaching (Genesis 9:11-17, Isaiah 40:8, 1 Peter 1:25). He is often portrayed with a rainbow banner in his hand. The Thomas Müntzer statue in the German town of Stolberg also shows him holding a rainbow banner in his hand. In the German Peasants' War of the 16th century, Müntzer's rainbow banner together with the peasants' boot ("Bundschuh") was used as the sign of a new era, of hope and of social change. The choice of the rainbow in the form of a flag harks back to the rainbow as a symbol of biblical promise. According to the Bible, God used the rainbow as a sign to Noah that there would never again be a worldwide flood, also known as the Rainbow covenant. Armenian Republic proposed flag (1
https://en.wikipedia.org/wiki/Altruism
Altruism is the principle and practice of concern for the well-being and/or happiness of other humans or animals. While objects of altruistic concern vary, it is an important moral value in many cultures and religions. It may be considered a synonym of selflessness, the opposite of selfishness. The word altruism was popularized (and possibly coined) by the French philosopher Auguste Comte in French, as , for an antonym of egoism. He derived it from the Italian , which in turn was derived from Latin , meaning "other people" or "somebody else". Altruism, as observed in populations of organisms, is when an individual performs an action at a cost to themselves (in terms of e.g. pleasure and quality of life, time, probability of survival or reproduction) that benefits, directly or indirectly, another individual, without the expectation of reciprocity or compensation for that action. Altruism can be distinguished from feelings of loyalty or concern for the common good. The latter are predicated upon social relationships, whilst altruism does not consider relationships. Whether "true" altruism is possible in human psychology is a subject of debate. The theory of psychological egoism suggests that no act of sharing, helping, or sacrificing can be truly altruistic, as the actor may receive an intrinsic reward in the form of personal gratification. The validity of this argument depends on whether such intrinsic rewards qualify as "benefits". The term altruism may also refer to an ethical doctrine that claims that individuals are morally obliged to benefit others. Used in this sense, it is usually contrasted with egoism, which claims individuals are morally obligated to serve themselves first. Effective altruism is the use of evidence and reason to determine the most effective ways to benefit others. The notion of altruism The concept of altruism has a history in philosophical and ethical thought. The term was coined in the 19th century by the founding sociologist and ph
https://en.wikipedia.org/wiki/C-element
In digital computing, the Muller C-element (C-gate, hysteresis flip-flop, coincident flip-flop, or two-hand safety circuit) is a small binary logic circuit widely used in design of asynchronous circuits and systems. It outputs 0 when all inputs are 0, it outputs 1 when all inputs are 1, and it retains its output state otherwise. It was specified formally in 1955 by David E. Muller and first used in ILLIAC II computer. In terms of the theory of lattices, the C-element is a semimodular distributive circuit, whose operation in time is described by a Hasse diagram. The C-element is closely related to the rendezvous and join elements, where an input is not allowed to change twice in succession. In some cases, when relations between delays are known, the C-element can be realized as a sum-of-product (SOP) circuit. Earlier techniques for implementing the C-element include Schmitt trigger, Eccles-Jordan flip-flop and last moving point flip-flop. Truth table and delay assumptions For two input signals the C-element is defined by the equation , which corresponds to the following truth table: This table can be turned into a circuit using the Karnaugh map. However, the obtained implementation is naive, since nothing is said about delay assumptions. To understand under what conditions the obtained circuit is workable, it is necessary to do additional analysis, which reveals that delay1 is a propagation delay from node 1 via environment to node 3, delay2 is a propagation delay from node 1 via internal feedback to node 3, delay1 must be greater than delay2. Thus, the naive implementation is correct only for slow environment. The definition of C-element can be generalized for multiple-valued logic , or even for continuous signals: For example, the truth table for a balanced ternary C-element with two inputs is Implementations of the C-element Depending on the requirements to the switching speed and power consumption, the C-element can be realized as a coarse- or fine-grain
https://en.wikipedia.org/wiki/Normally%20flat%20ring
In algebraic geometry, a normally flat ring along a proper ideal I is a local ring A such that is flat over for each integer . The notion was introduced by Hironaka in his proof of the resolution of singularities as a refinement of equimultiplicity and was later generalized by Alexander Grothendieck and others.
https://en.wikipedia.org/wiki/Trichome
Trichomes (; ) are fine outgrowths or appendages on plants, algae, lichens, and certain protists. They are of diverse structure and function. Examples are hairs, glandular hairs, scales, and papillae. A covering of any kind of hair on a plant is an indumentum, and the surface bearing them is said to be pubescent. Algal trichomes Certain, usually filamentous, algae have the terminal cell produced into an elongate hair-like structure called a trichome. The same term is applied to such structures in some cyanobacteria, such as Spirulina and Oscillatoria. The trichomes of cyanobacteria may be unsheathed, as in Oscillatoria, or sheathed, as in Calothrix. These structures play an important role in preventing soil erosion, particularly in cold desert climates. The filamentous sheaths form a persistent sticky network that helps maintain soil structure. Plant trichomes Plant trichomes have many different features that vary between both species of plants and organs of an individual plant. These features affect the subcategories that trichomes are placed into. Some defining features include the following: Unicellular or multicellular Straight (upright with little to no branching), spiral (corkscrew-shaped) or hooked (curved apex) Presence of cytoplasm Glandular (secretory) vs. eglandular Tortuous, simple (unbranched and unicellular), peltate (scale-like), stellate (star-shaped) Adaxial vs. abaxial, referring to whether trichomes are present, respectively, on the upper surface (adaxial) or lower surface (abaxial) of a leaf or other lateral organ. In a model organism, Cistus salviifolius, there are more adaxial trichomes present on this plant because this surface suffers from more ultraviolet (UV), solar irradiance light stress than the abaxial surface. Trichomes can protect the plant from a large range of detriments, such as UV light, insects, transpiration, and freeze intolerance. Aerial surface hairs Trichomes on plants are epidermal outgrowths of various kinds
https://en.wikipedia.org/wiki/Memory%20bandwidth
Memory bandwidth is the rate at which data can be read from or stored into a semiconductor memory by a processor. Memory bandwidth is usually expressed in units of bytes/second, though this can vary for systems with natural data sizes that are not a multiple of the commonly used 8-bit bytes. Memory bandwidth that is advertised for a given memory or system is usually the maximum theoretical bandwidth. In practice the observed memory bandwidth will be less than (and is guaranteed not to exceed) the advertised bandwidth. A variety of computer benchmarks exist to measure sustained memory bandwidth using a variety of access patterns. These are intended to provide insight into the memory bandwidth that a system should sustain on various classes of real applications. Measurement conventions There are three different conventions for defining the quantity of data transferred in the numerator of "bytes/second": The bcopy convention: counts the amount of data copied from one location in memory to another location per unit time. For example, copying 1 million bytes from one location in memory to another location in memory in one second would be counted as 1 million bytes per second. The bcopy convention is self-consistent, but is not easily extended to cover cases with more complex access patterns, for example three reads and one write. The Stream convention: sums the amount of data that the application code explicitly reads plus the amount of data that the application code explicitly writes. Using the previous 1 million byte copy example, the STREAM bandwidth would be counted as 1 million bytes read plus 1 million bytes written in one second, for a total of 2 million bytes per second. The STREAM convention is most directly tied to the user code, but may not count all the data traffic that the hardware is actually required to perform. The hardware convention: counts the actual amount of data read or written by the hardware, whether the data motion was explicitly reques
https://en.wikipedia.org/wiki/Transovarial%20transmission
Transovarial or transovarian transmission (transmission from parent to offspring via the ovaries) occurs in certain arthropod vectors as they transmit pathogens from parent to offspring. This process, used by a wide variety of parasites, is also known as vertical transmission. For example, Rickettsia rickettsii, carried within ticks, is passed on from parent to offspring tick by transovarial transmission. This is in contrast to parasites such as Rickettsia prowazekii, which are not passed on by transovarian transmission due to killing the vector that carries it (in this case, the human louse). Other examples of parasites that use this mechanism of transmission include the aedes mosquito vector of the yellow fever virus and in phlebotomine sandflies that transmit pappataci fever. Richard Dawkins in "The Extended Phenotype" notes that "bacterial endosymbionts of insects which are transmitted transovarially" share an interest in the "success of their host's gametes.....as well as the survival of their host's body." In this case, "the interest of the host genes and parasite genes might not be quite identical, but they would..... be very much closer than the case of fluke and snail." where host and parasite have different means of propagation into the next generation, and therefore more divergent interests. See also Transstadial transmission
https://en.wikipedia.org/wiki/Serena%20Software
With operations in 11 countries, Serena Software Inc. is an American software company that provides IT management products to enterprises. Serena solutions offer a process orchestration approach and span the areas of development, DevOps and IT management. Serena is headquartered in Silicon Valley, California and has 400+ employees. The company has been a subsidiary of the British software and information company Micro Focus since 2016. History Serena Software acquired Merant Software in 2004 which provided Serena ownership rights to Polytron Version Control System (PVCS). In May 2004, Serena announced it would acquire TeamShare, producers of the defect management system TeamTrack. This later became Serena Business Mashups (SBM), which was shortly renamed Serena Business Manager. In November 2005, Serena Software announced that it had agreed to be acquired by Silver Lake Partners in a leveraged buyout transaction. The buyout was announced at a price of per share in cash, and the transaction occurred in 2006. Following the transaction, Silver Lake, a technology focused private equity firm acquired about 70% of the Serena stock. The Silver Lake Partners acquisition resulted in no immediate changes to the executive management team and daily operations continued much as they had before. In October 2006, Serena announced the acquisition of Pacific Edge Software to enter the emerging project portfolio management (PPM) market. Mariner, a leading provider of Project and Portfolio Management (PPM) solutions was determined to be a good match for Serena's existing portfolio of products. After Pacific Edge, Serena made a number of more minor acquisitions, purchasing prototyping technology from Apptero, requirements management technology from Integrated Chipware and asset discovery technology from Data Sciences Corporation. In September 2008 Serena announced the acquisition of Projity, an open source software as a service (SaaS)-based project management company. John Nu
https://en.wikipedia.org/wiki/SnapTag
SnapTag, invented by SpyderLynk, is a 2D mobile barcode alternative similar to a QR code, but that uses an icon or company logo and code ring rather than a square pattern of black dots. Similar to a QR code, SnapTags can be used to take consumers to a brand’s website, but can also facilitate mobile purchases, coupon downloads, free sample requests, video views, promotional entries, Facebook Likes, Pinterest Pins, Twitter Follows, Posts and Tweets. SnapTags offer back-end data mining capabilities. Use in mobile operating systems SnapTags can be used in Google's mobile Android operating system and iOS devices (iPhone/iPod/iPad) using The SnapTag Reader App or third party apps that have integrated the SnapTag Reader SDK. SnapTags can also be used by standard camera phones by taking a picture of the SnapTag and texting it to the designated short code or email address.
https://en.wikipedia.org/wiki/Ecotropica
Ecotropica is the peer-reviewed international scientific journal of the European Society for Tropical Ecology, GTOE. The journal first appeared in 1995 and received an impact factor in 2011. Ecotropica covers all aspects of tropical ecology, and normally appears twice annually. Papers reflect results of original research (major papers, short communications) or review important fields in tropical ecology. Current editor is Marco Tschapka at the University of Ulm in Germany. One of the former editors was Elisabeth Kalko. External links Ecotropica, an international journal of tropical ecology European Society for Tropical Ecology Ecology journals
https://en.wikipedia.org/wiki/Whole%20number%20rule
In chemistry, the whole number rule states that the masses of the isotopes are whole number multiples of the mass of the hydrogen atom. The rule is a modified version of Prout's hypothesis proposed in 1815, to the effect that atomic weights are multiples of the weight of the hydrogen atom. It is also known as the Aston whole number rule after Francis W. Aston who was awarded the Nobel Prize in Chemistry in 1922 "for his discovery, by means of his mass spectrograph, of isotopes, in a large number of non-radioactive elements, and for his enunciation of the whole-number rule." Law of definite proportions The law of definite proportions was formulated by Joseph Proust around 1800 and states that all samples of a chemical compound will have the same elemental composition by mass. The atomic theory of John Dalton expanded this concept and explained matter as consisting of discrete atoms with one kind of atom for each element combined in fixed proportions to form compounds. Prout's hypothesis In 1815, William Prout reported on his observation that the atomic weights of the elements were whole multiples of the atomic weight of hydrogen. He then hypothesized that the hydrogen atom was the fundamental object and that the other elements were a combination of different numbers of hydrogen atoms. Aston's discovery of isotopes In 1920, Francis W. Aston demonstrated through the use of a mass spectrometer that apparent deviations from Prout's hypothesis are predominantly due to the existence of isotopes. For example, Aston discovered that neon has two isotopes with masses very close to 20 and 22 as per the whole number rule, and proposed that the non-integer value 20.2 for the atomic weight of neon is due to the fact that natural neon is a mixture of about 90% neon-20 and 10% neon-22). A secondary cause of deviations is the binding energy or mass defect of the individual isotopes. Discovery of the neutron During the 1920s, it was thought that the atomic nucleus was made o
https://en.wikipedia.org/wiki/Numerical%20response
The numerical response in ecology is the change in predator density as a function of change in prey density. The term numerical response was coined by M. E. Solomon in 1949. It is associated with the functional response, which is the change in predator's rate of prey consumption with change in prey density. As Holling notes, total predation can be expressed as a combination of functional and numerical response. The numerical response has two mechanisms: the demographic response and the aggregational response. The numerical response is not necessarily proportional to the change in prey density, usually resulting in a time lag between prey and predator populations. For example, there is often a scarcity of predators when the prey population is increasing. Demographic response The demographic response consists of changes in the rates of predator reproduction or survival due to a changes in prey density. The increase in prey availability translates into higher energy intake and reduced energy output. This is different from an increase in energy intake due to increased foraging efficiency, which is considered a functional response. This concept can be articulated in the Lotka-Volterra Predator-Prey Model. a = conversion efficiency: the fraction of prey energy assimilated by the predator and turned into new predators P = predator density V = prey density m = predator mortality c = capture rate Demographic response consists of a change in dP/dt due to a change in V and/or m. For example, if V increases, then predator growth rate (dP/dt) will increase. Likewise if the energy intake increases (due to greater food availability) and a decrease in energy output (from foraging), then predator mortality (m) will decrease and predator growth rate (dP/dt) will increase. In contrast, the functional response consists of a change in conversion efficiency (a) or capture rate (c). The relationship between available energy and reproductive efforts can be explained with the life his
https://en.wikipedia.org/wiki/River%20delta
A river delta is a landform shaped like a triangle, created by the deposition of sediment that is carried by a river and enters slower-moving or stagnant water. This occurs when a river enters an ocean, sea, estuary, lake, reservoir, or (more rarely) another river that cannot carry away the supplied sediment. It is so named because its triangle shape resembles the Greek letter Delta. The size and shape of a delta are controlled by the balance between watershed processes that supply sediment, and receiving basin processes that redistribute, sequester, and export that sediment. The size, geometry, and location of the receiving basin also plays an important role in delta evolution. River deltas are important in human civilization, as they are major agricultural production centers and population centers. They can provide coastline defense and can impact drinking water supply. They are also ecologically important, with different species' assemblages depending on their landscape position. On geologic timescales, they are also important carbon sinks. Etymology A river delta is so named because the shape of the Nile Delta approximates the triangular uppercase Greek letter delta. The triangular shape of the Nile Delta was known to audiences of classical Athenian drama; the tragedy Prometheus Bound by Aeschylus refers to it as the "triangular Nilotic land", though not as a "delta". Herodotus's description of Egypt in his Histories mentions the Delta fourteen times, as "the Delta, as it is called by the Ionians", including describing the outflow of silt into the sea and the convexly curved seaward side of the triangle. Despite making comparisons to other river-systems' deltas, Herodotus did not describe them as "deltas". The Greek historian Polybius likened the land between the Rhône and Isère rivers to the Nile Delta, referring to both as islands, but did not apply the word delta. According to the Greek geographer Strabo, the Cynic philosopher Onesicritus of Astypalaea,
https://en.wikipedia.org/wiki/Zero%20of%20a%20function
In mathematics, a zero (also sometimes called a root) of a real-, complex-, or generally vector-valued function , is a member of the domain of such that vanishes at ; that is, the function attains the value of 0 at , or equivalently, is the solution to the equation . A "zero" of a function is thus an input value that produces an output of 0. A root of a polynomial is a zero of the corresponding polynomial function. The fundamental theorem of algebra shows that any non-zero polynomial has a number of roots at most equal to its degree, and that the number of roots and the degree are equal when one considers the complex roots (or more generally, the roots in an algebraically closed extension) counted with their multiplicities. For example, the polynomial of degree two, defined by has the two roots (or zeros) that are 2 and 3. If the function maps real numbers to real numbers, then its zeros are the -coordinates of the points where its graph meets the x-axis. An alternative name for such a point in this context is an -intercept. Solution of an equation Every equation in the unknown may be rewritten as by regrouping all the terms in the left-hand side. It follows that the solutions of such an equation are exactly the zeros of the function . In other words, a "zero of a function" is precisely a "solution of the equation obtained by equating the function to 0", and the study of zeros of functions is exactly the same as the study of solutions of equations. Polynomial roots Every real polynomial of odd degree has an odd number of real roots (counting multiplicities); likewise, a real polynomial of even degree must have an even number of real roots. Consequently, real odd polynomials must have at least one real root (because the smallest odd whole number is 1), whereas even polynomials may have none. This principle can be proven by reference to the intermediate value theorem: since polynomial functions are continuous, the function value must cross zero, in the
https://en.wikipedia.org/wiki/Ultra%20HD%20Blu-ray
Ultra HD Blu-ray (4K Ultra HD, UHD-BD, or 4K Blu-ray) is a digital optical disc data storage format that is an enhanced variant of Blu-ray. Ultra HD Blu-ray discs are incompatible with existing standard Blu-ray players. Ultra HD Blu-ray supports 4K UHD (3840 × 2160 pixel resolution) video at frame rates up to 60 progressive frames per second, encoded using High-Efficiency Video Coding. The discs support both high dynamic range by increasing the color depth to 10-bit per color and a greater color gamut than supported by conventional Blu-ray video by using the Rec. 2020 color space. Ultra HD Blu-Ray discs also support a 12-bit per color container via Dolby Vision. Dolby Vision content on 4K UHD Blu-Ray can also be mastered for 10,000 nits peak brightness, whereas standard HDR10 can only achieve a maximum of 4,000 nits of brightness. Moreover, Dolby Vision makes use of dynamic metadata, which adjusts the brightness and tone mapping per scene. In contrast, standard HDR10 only makes use of static metadata, which sets the same brightness and tone mapping for the entirety of the content. The format is supported on Microsoft's Xbox One X, One S, Series X, and Sony's PlayStation 5. Software made for the PlayStation 5 can use 100 GB UHD Blu-ray discs. The first Ultra HD Blu-ray Discs were officially released in the United States on February 14, 2016. To differentiate retail Ultra HD Blu-ray releases, the format usually uses a black opaque or slightly transparent keep case packaging format (as opposed to blue). The case size is the same as that of a normal Blu-ray disc. The Blu-ray Disc Association also completed the specification for an 8K Blu-ray format supporting 8K UHD (7680 × 4320 pixel resolution) videos for use in Japan. Specifications The specification for 4K Blu-ray allows for three disc capacities, each with its own data rate: 50 GB at 72 or 92 Mbit/s, and 66 GB and 100 GB at 92, 123, or 144 Mbit/s. On 66 GB and 100 GB discs, the pits and lands are not narrower
https://en.wikipedia.org/wiki/Einstein%E2%80%93Weyl%20geometry
An Einstein–Weyl geometry is a smooth conformal manifold, together with a compatible Weyl connection that satisfies an appropriate version of the Einstein vacuum equations, first considered by and named after Albert Einstein and Hermann Weyl. Specifically, if is a manifold with a conformal metric , then a Weyl connection is by definition a torsion-free affine connection such that where is a one-form. The curvature tensor is defined in the usual manner by and the Ricci curvature is The Ricci curvature for a Weyl connection may fail to be symmetric (its skew part is essentially the exterior derivative of .) An Einstein–Weyl geometry is then one for which the symmetric part of the Ricci curvature is a multiple of the metric, by an arbitrary smooth function: The global analysis of Einstein–Weyl geometries is generally more subtle than that of conformal geometry. For example, the Einstein cylinder is a global static conformal structure, but only one period of the cylinder (with the conformal structure of the de Sitter metric) is Einstein–Weyl. Citations
https://en.wikipedia.org/wiki/LydA%20holin%20family
The LydA Holin (LydA Holin) Family (TC# 1.E.4), named after the lydA gene which codes for its prototype member (LydA of E. coli; TC# 1.E.4.1.1), belongs to the Holin Superfamily III. Members of this family have 3 transmembrane segments (TMSs) and appear to possess between 90 and 120 amino acyl residues (aas). A representative list of proteins belonging to this family can be found in the Transporter Classification Database. See also Holin Lysin Transporter Classification Database
https://en.wikipedia.org/wiki/TPBG
Trophoblast glycoprotein, also known as TPBG, 5T4, Wnt-Activated Inhibitory Factor 1 or WAIF1, is a human protein encoded by a TPBG gene. TPBG is an antagonist of Wnt/β-catenin signalling pathway. Clinical significance 5T4 is an antigen expressed in a number of carcinomas. It is an N-glycosylated transmembrane 72 kDa glycoprotein containing eight leucine-rich repeats. 5T4 is often referred to as an oncofetal antigen due to its expression in foetal trophoblast (where it was first discovered) or trophoblast glycoprotein (TPBG). 5T4 is found in tumors including the colorectal, ovarian, and gastric. Its expression is used as a prognostic aid in these cases. It has very limited expression in normal tissue but is widespread in malignant tumours throughout their development. One study found that 5T4 was present in 85% of a cohort of 72 colorectal carcinomas and in 81% of a cohort of 27 gastric carcinomas. Its confined expression appears to give 5T4 the potential to be a target for T cells in cancer immunotherapy. There has been extensive research into its role in antibody-directed immunotherapy through the use of the high-affinity murine monoclonal antibody, mAb5T4, to deliver response modifiers (such as staphylococcus aureus superantigen) accurately to a tumor. 5T4 is also the target of the cancer vaccine TroVax which is in clinical trials for the treatment of a range of different solid tumour types. Interactions TPBG has been shown to interact with GIPC1.
https://en.wikipedia.org/wiki/Ramanujan%20theta%20function
In mathematics, particularly -analog theory, the Ramanujan theta function generalizes the form of the Jacobi theta functions, while capturing their general properties. In particular, the Jacobi triple product takes on a particularly elegant form when written in terms of the Ramanujan theta. The function is named after mathematician Srinivasa Ramanujan. Definition The Ramanujan theta function is defined as for . The Jacobi triple product identity then takes the form Here, the expression denotes the -Pochhammer symbol. Identities that follow from this include and and This last being the Euler function, which is closely related to the Dedekind eta function. The Jacobi theta function may be written in terms of the Ramanujan theta function as: Integral representations We have the following integral representation for the full two-parameter form of Ramanujan's theta function: The special cases of Ramanujan's theta functions given by and also have the following integral representations: This leads to several special case integrals for constants defined by these functions when (cf. theta function explicit values). In particular, we have that and that Application in string theory The Ramanujan theta function is used to determine the critical dimensions in Bosonic string theory, superstring theory and M-theory.
https://en.wikipedia.org/wiki/Evolution%3A%20The%20Origin%20of%20Species
Evolution: The Origin of Species is a card game created by Dmitriy Knorre and Sergey Machin in 2010. The game is inspired by the evolutionary biology. It was published by SIA Rightgames RBG. English, French and German game editions were published in 2011. Two or more players create their own animals, make them evolve and hunt in order to survive. In 2014, North Star Games published game Evolution. The original authors were part of the design crew. Rules Place definition in match The player with the largest number of victory points at the end of the game is the winner. The rankings of players in match are determined as follows: Preparation The deck is shuffled. Then each player gets 6 cards from the top of the deck to their hands. They roll dice to determine the first player. The game turn structure Each turn of the game consists of four phases: During each phase players act in order moving clockwise. The player who can't or doesn't want to act passes. Each phase ends when nobody can or want to act. Development phase This phase consists of several rounds. During the phase players can play their cards by putting them from their hands onto the table. Each player may play each card either as an animal or as a trait of an existing animal. If card is played as a trait, it is put underneath the corresponding animal. Some cards have two traits, but only one trait can be used - the chosen one. No animal can have two identical traits except "fat tissue" trait. Pairwise traits (i.e. "communication") are played only onto a pair of animals. Such cards are placed between the two cards onto which they are played. No duplet of animals can have two identical pairwise traits. However an animal can have several different pairwise traits with another animal. Food bank determination phase The amount of food available during this turn is determined at this time. Food bank estimated this way: Using two game sets or one game set with first expansion: Using two game set
https://en.wikipedia.org/wiki/Gamelan%20notation
Notation plays a relatively minor role in the oral traditions of Indonesian gamelan but, in Java and Bali, several systems of gamelan notation were devised beginning at the end of the 19th century, initially for archival purposes. Kepatihan Kepatihan is a type of cipher musical notation that was devised for the notation of the Indonesian gamelan. History The system was devised around 1900 at the Kepatihan (the Grand Vizier's compound) in Surakarta, and was based upon the Galin-Paris-Chevé system, imported in the nineteenth century by Christian missionaries to allow the notation of hymns. It superseded several other notation systems of Javanese origin devised around the same time. Notation The pitches of the seven-tone pélog tuning system are designated by the numbers 1, 2, 3, 4, 5, 6, and 7; while the five-tone slendro pitches are notated as 1, 2, 3, 5, and 6. The octaves are noted by dots above and below the numbers, as in Chinese jianpu, although of course the pitches do not correspond. A dot over a note indicates the octave above, and a dot below a note represents the octave below. Two dots over a note indicate a note two octaves higher than standard, and so on. Depending on the tuning of the individual gamelan, it is often possible to hear the pitches 1, 2, 3, 5, and 6 of slendro as an anhemitonic pentatonic scale, do-re-mi-sol-la. However, in the pélog system pitches are simply numbered from low to high 1–7 and there is no question of interpreting these sounds diatonically. As the pélog scale is essentially a five-note scale, the notes 4 and 7 function similarly to 'accidentals' in Western terms: a 4 may serve as a 'sharp' or raised 3 (common in patet lima or nem) or as a 'flat' or lowered 5 (usual in patet barang). Similarly 7 functions as a 'flat' 1 in patet lima or nem; 1 in patet barang may function as a 'sharp' 7, but is more often interpreted as a temporary change of pathet. By default, kepatihan notes are assumed all to have the same duration. D
https://en.wikipedia.org/wiki/Alfred%20Bourne%20%28cricketer%29
Alfred Bourne (16 April 1848 – 17 July 1931) was an English cricketer. He played four first-class matches for Cambridge University Cricket Club in 1870. See also List of Cambridge University Cricket Club players
https://en.wikipedia.org/wiki/Polyominoid
In geometry, a polyominoid (or minoid for short) is a set of equal squares in 3D space, joined edge to edge at 90- or 180-degree angles. The polyominoids include the polyominoes, which are just the planar polyominoids. The surface of a cube is an example of a hexominoid, or 6-cell polyominoid, and many other polycubes have polyominoids as their boundaries. Polyominoids appear to have been first proposed by Richard A. Epstein. Classification 90-degree connections are called hard; 180-degree connections are called soft. This is because, in manufacturing a model of the polyominoid, a hard connection would be easier to realize than a soft one. Polyominoids may be classified as hard if every junction includes a 90° connection, soft if every connection is 180°, and mixed otherwise, except in the unique case of the monominoid, which has no connections of either kind. The set of soft polyominoids is equal to the set of polyominoes. As with other polyforms, two polyominoids that are mirror images may be distinguished. One-sided polyominoids distinguish mirror images; free polyominoids do not. Enumeration The table below enumerates free and one-sided polyominoids of up to 6 cells. Generalization to higher dimensions In general one can define an n,k-polyominoid as a polyform made by joining k-dimensional hypercubes at 90° or 180° angles in n-dimensional space, where 1≤k≤n. Polysticks are 2,1-polyominoids. Polyominoes are 2,2-polyominoids. The polyforms described above are 3,2-polyominoids. Polycubes are 3,3-polyominoids.
https://en.wikipedia.org/wiki/Commodore%20LCD
The Commodore LCD (sometimes known in short as the CLCD) is an LCD-equipped laptop made by Commodore International. It was presented at the January 1985 Consumer Electronics Show, but never released. The CLCD was not directly compatible with other Commodore home computers, but its built-in Commodore BASIC 3.6 interpreter could run programs written in the Commodore 128's BASIC 7.0, as long as these programs did not include system-specific POKE commands. Like the Commodore 264 and Radio Shack TRS-80 Model 100 series computers, the CLCD had several built-in ROM-based office application programs. The CLCD featured a 1 MHz Rockwell 65C102 CPU (a CMOS 6502 variant) and 32 KB of RAM (expandable to 64 KB internally). The BASIC interpreter and application programs were built into 96 KB of ROM.
https://en.wikipedia.org/wiki/General%20Federation%20of%20Agriculture%2C%20Food%2C%20Tobacco%20and%20Related%20Services%20Workers
The General Federation of Agriculture, Food, Tobacco and Related Services Workers (, FGTA) is a trade union representing workers in a range of related industries in France. The union was founded in 1975, when the National Federation of Agricultural Workers merged with the Federation of Workers in Food and Retail. Like its predecessors, the union affiliated to Workers' Force. From formation until 1997, it was led by Jean-Marie Goube. In 1988, the Federation of Tobacco and Matches merged in. By 1995, the union claimed 14,000 members.
https://en.wikipedia.org/wiki/Britains%20Deetail
Britains' Deetail toy soldiers were a popular product in the 1970s and 1980s. Manufactured in England by W. Britain, the 1/32 (54mm) scale plastic figures were finished with hand painted details and came with sturdy Zamak metal bases. In the early 1990s production moved to China before eventually being phased out. Development In 1971, Britains began phasing out its plastic Herald Miniatures produced in Hong Kong since 1966, with Herald eventually ceasing production in 1976. New Deetail figures were produced moulded in PVC plastic rather than polythene and using plug-in type arms, which were glued to bodies resulting in poses previously unavailable. Figures were moulded with a tee shaped "footlug" on the feet of each figure that allowed secure attachment to sturdy metal bases. These rectangular metal bases ensured figures stood better than rival manufacturers products and paid homage to Britains hollowcast metal figures as well as being thought by consumers to be of "better value" due to their heavier weight. The first Deetail figures produces were based upon Second World War (WW2) American and German infantry. Sculptor Rod Cameron rented uniforms from Berman's and Nathan's theatrical costumes with Cameron giving model Les Harden his air rifle to pose with. In 1978 Britains developed the Super Deetail range using an overmoulding process whereby different coloured plastics came together in one figure. The initial release were modern British paratroopers with red berets. Range Figures were generally produced in six different standing poses and represented various historical periods from medieval to the modern era including: Knights and Turks Waterloo American Wild West - Cowboys and Indians, 7th Cavalry, etc. American Civil War African Desert Military - WW2 British, American, German; Modern; Guards, etc. Space Mounted figures were also released including a horse on metal base with rider glued in place. Riders and horses were presented in various different
https://en.wikipedia.org/wiki/12-Hydroxyeicosatetraenoic%20acid
12-Hydroxyeicosatetraenoic acid (12-HETE) is a derivative of the 20 carbon polyunsaturated fatty acid, arachidonic acid, containing a hydroxyl residue at carbon 12 and a 5Z,8Z,10E,14Z Cis–trans isomerism configuration (Z=cis, E=trans) in its four double bonds. It was first found as a product of arachidonic acid metabolism made by human and bovine platelets through their 12S-lipoxygenase (i.e. ALOX12) enzyme(s). However, the term 12-HETE is ambiguous in that it has been used to indicate not only the initially detected "S" stereoisomer, 12S-hydroxy-5Z,8Z,10E,14Z-eicosatetraenoic acid (12(S)-HETE or 12S-HETE), made by platelets, but also the later detected "R" stereoisomer, 12(R)-hydroxy-5Z,8Z,10E,14Z-eicosatetraenoic acid (also termed 12(R)-HETE or 12R-HETE) made by other tissues through their 12R-lipoxygenase enzyme, ALOX12B. The two isomers, either directly or after being further metabolized, have been suggested to be involved in a variety of human physiological and pathological reactions. Unlike hormones which are secreted by cells, travel in the circulation to alter the behavior of distant cells, and thereby act as Endocrine signalling agents, these arachidonic acid metabolites act locally as Autocrine signalling and/or Paracrine signaling agents to regulate the behavior of their cells of origin or of nearby cells, respectively. In these roles, they may amplify or dampen, expand or contract cellular and tissue responses to disturbances. Production In humans, Arachidonate 12-lipoxygenase (12-LO, 12-LOX, ALO12, or platelet type 12-lipoxygenase) is encoded by the ALOX12 gene and expressed primarily in platelets and skin. ALOX12 metabolizes arachidonic acid almost exclusively to 12(S)-hydroperoxy-5Z,8Z,10E,14Z-eicosatetraenoic acid (12(S)-HpETE or 12S-HpETE). Arachidonate 15-lipoxygenase-1 (15-LO-1, 15-LOX-1, ALOX15), which is expressed in far more tissues that ALOX12, metabolizes arachidonic acid primarily to 15(S)-HpETE along with other metabolites of the 15
https://en.wikipedia.org/wiki/Quantum%20Experiments%20at%20Space%20Scale
Quantum Experiments at Space Scale (QUESS; ), is a Chinese research project in the field of quantum physics. Tiangong-2 is China's second Space Laboratory module which was launched on 15 September 2016. Tiangong-2 carries a total of 14 mission and experiment packages, including Space-Earth quantum key distribution () and laser communications experiment to facilitate space-to-ground quantum communication. A satellite, nicknamed Micius or Mozi () after the ancient Chinese philosopher, is operated by the Chinese Academy of Sciences, as well as ground stations in China. The University of Vienna and the Austrian Academy of Sciences are running the satellite's European receiving stations. QUESS is a proof-of-concept mission designed to facilitate quantum optics experiments over long distances to allow the development of quantum encryption and quantum teleportation technology. Quantum encryption uses the principle of entanglement to facilitate communication that can absolutely detect whether a third party has intercepted a message in transit thus denying undetected decryption. By producing pairs of entangled photons, QUESS will allow ground stations separated by many thousands of kilometres to establish secure quantum channels. QUESS itself has limited communication capabilities: it needs line-of-sight, and can only operate when not in sunlight. Further Micius satellites were planned, including a global network by 2030. The mission cost was around US$100 million in total. Mission The initial experiment demonstrated quantum key distribution (QKD) between Xinjiang Astronomical Observatory near Ürümqi and Xinglong Observatory near Beijing – a great-circle distance of approximately . In addition, QUESS tested Bell's inequality at a distance of – further than any experiment to date – and teleported a photon state between Shiquanhe Observatory in Ali, Tibet Autonomous Region, and the satellite. This requires very accurate orbital maneuvering and satellite tracking so th
https://en.wikipedia.org/wiki/Nuclear%20Instruments%20and%20Methods%20in%20Physics%20Research
Nuclear Instruments and Methods in Physics Research (Nucl. Instrum. Methods Phys. Res.) is a peer-reviewed scientific journal published by Elsevier. It was established in 1957 as Nuclear Instruments. It focuses on detectors descriptions and data analysis methods. History Nuclear Instruments (1957–1958) Nuclear Instruments and Methods (1959–1981) Nuclear Instruments and Methods in Physics Research (1981–present) Nuclear Instruments and Methods in Physics Research Section A: Accelerators, Spectrometers, Detectors and Associated Equipment (1984–present) Nuclear Instruments and Methods in Physics Research Section B: Beam Interactions with Materials and Atoms (1984–present) External links Nuclear Instruments Nuclear Instruments and Methods Nuclear Instruments and Methods in Physics Research Nuclear Instruments and Methods in Physics Research Section A: Accelerators, Spectrometers, Detectors and Associated Equipment Nuclear Instruments and Methods in Physics Research Section B: Beam Interactions with Materials and Atoms Elsevier academic journals Academic journals established in 1957 English-language journals Nuclear physics journals Research methods journals
https://en.wikipedia.org/wiki/Knowledge%20integration
Knowledge integration is the process of synthesizing multiple knowledge models (or representations) into a common model (representation). Compared to information integration, which involves merging information having different schemas and representation models, knowledge integration focuses more on synthesizing the understanding of a given subject from different perspectives. For example, multiple interpretations are possible of a set of student grades, typically each from a certain perspective. An overall, integrated view and understanding of this information can be achieved if these interpretations can be put under a common model, say, a student performance index. The Web-based Inquiry Science Environment (WISE), from the University of California at Berkeley has been developed along the lines of knowledge integration theory. Knowledge integration has also been studied as the process of incorporating new information into a body of existing knowledge with an interdisciplinary approach. This process involves determining how the new information and the existing knowledge interact, how existing knowledge should be modified to accommodate the new information, and how the new information should be modified in light of the existing knowledge. A learning agent that actively investigates the consequences of new information can detect and exploit a variety of learning opportunities; e.g., to resolve knowledge conflicts and to fill knowledge gaps. By exploiting these learning opportunities the learning agent is able to learn beyond the explicit content of the new information. The machine learning program KI, developed by Murray and Porter at the University of Texas at Austin, was created to study the use of automated and semi-automated knowledge integration to assist knowledge engineers constructing a large knowledge base. A possible technique which can be used is semantic matching. More recently, a technique useful to minimize the effort in mapping validation and vi
https://en.wikipedia.org/wiki/N%286%29-Carboxymethyllysine
N(6)-Carboxymethyllysine (CML), also known as Nε-(carboxymethyl)lysine, is an advanced glycation endproduct (AGE). CML has been the most used marker for AGEs in food analysis. Recently, it has been demonstrated that gut microbiota mediates an aging-associated decline in gut barrier function, allowing AGEs to leak into the bloodstream from the gut and impairing microglial function in the brain. It is suggested that the amount of CML in human blood samples may correlated with age. A humanized monoclonal antibody which binds to N6 – carboxymethyl lysine shows considerable promise as a possible therapeutic agent for treating pancreatic cancer.
https://en.wikipedia.org/wiki/Volanesorsen
Volanesorsen, sold under the brand name Waylivra, is a triglyceride-reducing drug. It is a second-generation 2'-O-methoxyethyl (2'-MOE) chimeric antisense therapeutic oligonucleotide (ASO) that targets the messenger RNA for apolipoprotein C3 (apo-CIII). The most common side effects include reduced platelet levels and reactions at the site of the injection such as pain, swelling, itching, or bruising. Volanesorsen, is an 'antisense oligonucleotide,' a very short piece of synthetic RNA (a type of genetic material). It has been designed to block the production of a protein that slows down the breakdown of fats called apolipoprotein C-III. By blocking the production of this protein, the medicine reduces the level of triglycerides in the blood and, as a result, fat accumulation in the body, which is expected to reduce the risk of pancreatitis. Medical uses Familial chylomicronaemia syndrome (FCS) (also known as type I hyperlipoproteinaemia) is an inherited disease where people have abnormally high levels of some types of fat called triglycerides in their blood. The excess fat accumulates in organs such as the spleen and liver, which become abnormally enlarged. Fat accumulation can also cause repeated bouts of pancreatitis (inflammation of the pancreas) and xanthomas (formation of yellow fatty deposits just under the skin, generally around joints) Volanesorsen is indicated as an adjunct to diet in adults with genetically confirmed familial chylomicronemia syndrome (FCS) and at high risk for pancreatitis, in whom response to diet and triglyceride lowering therapy has been inadequate. History It is in Phase III clinical trials for the treatment of hypertriglycidemia, familial chylomicronemia syndrome and familial partial lipodystrophy. The drug was discovered and developed by Ionis Pharmaceuticals. Volanesorsen was designated an orphan drug by the European Medicines Agency (EMA) in February 2014, for phosphorothioate oligonucleotide targeted to apolipoprotein C-III
https://en.wikipedia.org/wiki/Weil%E2%80%93Petersson%20metric
In mathematics, the Weil–Petersson metric is a Kähler metric on the Teichmüller space Tg,n of genus g Riemann surfaces with n marked points. It was introduced by using the Petersson inner product on forms on a Riemann surface (introduced by Hans Petersson). Definition If a point of Teichmüller space is represented by a Riemann surface R, then the cotangent space at that point can be identified with the space of quadratic differentials at R. Since the Riemann surface has a natural hyperbolic metric, at least if it has negative Euler characteristic, one can define a Hermitian inner product on the space of quadratic differentials by integrating over the Riemann surface. This induces a Hermitian inner product on the tangent space to each point of Teichmüller space, and hence a Riemannian metric. Properties stated, and proved, that the Weil–Petersson metric is a Kähler metric. proved that it has negative holomorphic sectional, scalar, and Ricci curvatures. The Weil–Petersson metric is usually not complete. Generalizations The Weil–Petersson metric can be defined in a similar way for some moduli spaces of higher-dimensional varieties.
https://en.wikipedia.org/wiki/Group%20Domain%20of%20Interpretation
Group Domain of Interpretation or GDOI is a cryptographic protocol for group key management. The GDOI protocol is specified in an IETF Standard, RFC 6407, and is based on Internet Security Association and Key Management Protocol (ISAKMP), RFC 2408, and Internet Key Exchange version 1 (IKE). Whereas IKE is run between two peers to establish a "pair-wise security association", GDOI protocol is run between a group member and a "group controller/key server" (controller) and establishes a security association among two or more group members. Functional Overview GDOI "interprets" IKE or ISAKMP for the group security domain in addition to pair-wise security associations. GDOI uses an IKE v1 Phase 1 security association for authenticating a GDOI member to a GDOI controller. The IKE/GDOI Phase 1 cryptographic protocol exchange protects a new type of Phase 2 exchange in which the member requests ("pulls") group state from the controller. The "group key" is the most important state in a GDOI member. The group key encrypts keys that decrypt application data. Thus, the group key is also called a "key encrypting key" in GDOI. A group's key-encrypting key is used for the "Rekey Security Association". Once the "Rekey-SA" is established, the GDOI controller can send ("push") unsolicited updates to the group security association to members over multicast, broadcast or unicast channels. This is why GDOI is called a "multicast key management system" as it uses and supports multicast messaging for very large groups. These multicast messages are unsolicited messages and are therefore called "push" messages, which are unsolicited messages sent from the controller to the members; explicit requests from a member to a controller are called "pull" messages in GDOI. Thus GDOI group key updates are pushed and can reach any number of group members with a single efficient transmission from the controller. GDOI group key updates also serve to remove members from groups. RFC 2627 de
https://en.wikipedia.org/wiki/Orchidology
Orchidology is the scientific study of orchids. It is an organismal-level branch of botany. See also List of orchidologists
https://en.wikipedia.org/wiki/California%20Botanical%20Society
The California Botanical Society was founded by Willis Linn Jepson in 1913, since when it has advanced the knowledge of botanical sciences in the Western United States Services The society services are: the journal Madroño, published since 1916; annual banquets in various California locations along with educational lectures; research support on green plants of Baja California, (enabled by the Annetta Carter Memorial Fund); graduate student support (together with the annual banquet); and community discussions with professional botanists. Journal Madroño is the quarterly peer-reviewed scientific journal of the Society. It was established in 1916 and focuses on botany in the western part of North America. Articles are published in English or Spanish. The current editor is Matt Ritter (California Polytechnic State University). See also University and Jepson Herbaria Cneoridium dumosum (Nuttall) Hooker F. Collected March 26, 1960, at an Elevation of about 1450 Meters on Cerro Quemazón, 15 Miles South of Bahía de Los Angeles, Baja California, México, Apparently for a Southeastward Range Extension of Some 140 Miles
https://en.wikipedia.org/wiki/Joanne%20M.%20Holden
Joanne Merson Holden (September 5, 1946 – October 2, 2014) was an American nutritionist known for her food composition research. She served as head of the USDA Agricultural Research Service's Nutrient Data Laboratory and at the time of her death was ranked among the top 1% most cited researchers in agricultural science. Education Joanne Holden received her Bachelor of Science degree in Foods and Nutrition from the University of Delaware in Newark, Delaware before going on to earn her Master of Science degree from the University of Maryland in 1974. Career & contributions Over the course of her career, Holden published over 100 papers on food composition data production, collection, management, and documentation and nutrition analysis. Holden and her colleagues from the Agricultural Research Service's Nutrient Data Laboratory published 31 papers in the JFCA (Journal of Food Composition and Analysis). Holden was listed as an ISI Highly Cited researcher in 2014, meaning that she ranked among the top 1% most cited researchers in the field of agricultural science. Her work contributed to the USDA National Nutrient Database for Standard Reference (also known as SR24), which is the standard source for food composition data in the U.S. for public and private sector uses. Among other projects, this work included significant updates to the nutrient information for fresh beef and pork products in 2012, as changes and improvements in animal husbandry practices and industry procedures over time affect nutrient content. In addition, Holden furthered global food science education by serving as co-director (1994 to at least 2007) of the International Postgraduate Course for the Production, Management, and Use of Food Composition Data, a project of the UN Food and Agriculture Organization's International Network of Food Data Systems (INFOODS).
https://en.wikipedia.org/wiki/Fairness%20measure
Fairness measures or metrics are used in network engineering to determine whether users or applications are receiving a fair share of system resources. There are several mathematical and conceptual definitions of fairness. Transmission Control Protocol fairness Congestion control mechanisms for new network transmission protocols or peer-to-peer applications must interact well with Transmission Control Protocol (TCP). TCP fairness requires that a new protocol receive a no larger share of the network than a comparable TCP flow. This is important as TCP is the dominant transport protocol on the Internet, and if new protocols acquire unfair capacity they tend to cause problems such as congestion collapse. This was the case with the first versions of RealMedia's streaming protocol: it was based on UDP and was widely blocked at organizational firewalls until a TCP-based version was developed. TCP throughput unfairness over WiFi is a critical problem and needs further investigations. Jain's fairness index Raj Jain's equation, rates the fairness of a set of values where there are users, is the throughput for the th connection, and is the sample coefficient of variation . The result ranges from (worst case) to 1 (best case), and it is maximum when all users receive the same allocation. This index is when users equally share the resource, and the other users receive zero allocation. This metric identifies underutilized channels and is not unduly sensitive to atypical network flow patterns. To achieve a given fairness level , one approximate method is to let , where and A is an arbitrary factor, typically used for normalization. This gives an allocation with a fairness close to F, and the allocation can then be refined to get even closer. Note this also allows for a prioritization of allocation, as the s will be sorted. An exact method is to let , where solves . A simple way to calculate is to use Newton's Method on , which converges consistently and fairly q
https://en.wikipedia.org/wiki/Virbhadra%E2%80%93Ellis%20lens%20equation
The Virbhadra-Ellis lens equation in astronomy and mathematics relates to the angular positions of an unlensed source , the image , the Einstein bending angle of light , and the angular diameter lens-source and observer-source distances. . This lens equation is useful for studying gravitational lensing in a strong gravitational field.
https://en.wikipedia.org/wiki/Adverse%20food%20reaction
An adverse food reaction is an adverse response by the body to food or a specific type of food. The most common adverse reaction is a food allergy, which is an adverse immune response to either a specific type or a range of food proteins. However, other adverse responses to food are not allergies. These reactions include responses to food such as food intolerance, pharmacological reactions, and toxin-mediated reactions, as well as physical responses, such as choking.
https://en.wikipedia.org/wiki/IBM%20Basic%20Programming%20Support
IBM Basic Programming Support/360 (BPS), originally called Special Support, was a set of standalone programs for System/360 mainframes with a minimum of 8 KiB of memory. BPS was developed by IBM's General Products Division in Endicott, New York. The package included "assemblers, IOCS, compilers, sorts, and utilities but no governing control program." BPS components were introduced in a series of product announcements between 1964 and 1965. BPS came in two versions — a strictly punched card system and a magnetic tape based system which, contrary to the stated goals, kept a small supervisor permanently resident. Programming languages available were IBM Basic Assembly Language, IBM RPG, and FORTRAN IV (subset). Tape FORTRAN required 16 KiB of memory. There were also two versions of the BPS assembler, with the tape version having enhanced capabilities. BPS also had a "disk" counterpart called BOS/360. It also required 8 KiB of memory and supported disks such as the IBM 2311. The group responsible for BPS/BOS went on to develop DOS/360 and TOS/360 as a supposed "interim" solution when it became evident that OS/360 would be too large to run on 16 KiB systems. BPS and BOS could be used to run standalone applications on a minimal System/360. One application was the System/360 Work Station for remote job entry to a larger system. See also Punched card input/output
https://en.wikipedia.org/wiki/Salicylaldehyde
Salicylic aldehyde (2-hydroxybenzaldehyde) is the organic compound with the formula (C7 H6 O2) C6H4CHO-2-OH. Along with 3-hydroxybenzaldehyde and 4-hydroxybenzaldehyde, it is one of the three isomers of hydroxybenzaldehyde. This colorless oily liquid has a bitter almond odor at higher concentration. Salicylaldehyde is a key precursor to a variety of chelating agents, some of which are commercially important. Production Salicylaldehyde is prepared from phenol and chloroform by heating with sodium hydroxide or potassium hydroxide in a Reimer–Tiemann reaction: Alternatively, it is produced by condensation of phenol or its derivatives with formaldehyde to give hydroxybenzyl alcohol, which is oxidized to the aldehyde. Salicylaldehydes in general may be prepared by other ortho-selective formylation reactions from the corresponding phenol, for instance by the Duff reaction, Reimer–Tiemann reaction, or by treatment with paraformaldehyde in the presence of magnesium chloride and a base. Natural occurrences Salicylaldehyde was identified as a characteristic aroma component of buckwheat. It is also one of the components of castoreum, the exudate from the castor sacs of the mature North American beaver (Castor canadensis) and the European beaver (Castor fiber), used in perfumery. Furthermore, salicylaldehyde occurs in the larval defensive secretions of several leaf beetle species that belong the subtribe Chrysomelina. An example for a leaf beetle species that produces salicylaldehyde is the red poplar leaf beetle Chrysomela populi. Reactions and applications Salicylaldehyde is used to make the following: Oxidation with hydrogen peroxide gives catechol (1,2-dihydroxybenzene) (Dakin reaction). Etherification with chloroacetic acid followed by cyclisation gives the heterocycle benzofuran (coumarone). The first step in this reaction to the substituted benzofuran is called the Rap–Stoermer condensation after E. Rap (1895) and R. Stoermer (1900). Salicylaldehyde is co
https://en.wikipedia.org/wiki/American%20Entomological%20Society
The American Entomological Society was founded on March 1, 1859. It is the oldest continuously operating entomology society in the Western Hemisphere, and one of the oldest scientific societies in the United States. It is headquartered in Philadelphia, Pennsylvania. The society publishes Entomological News, Transactions of the American Entomological Society, and Memoirs of the American Entomological Society. It is not affiliated in any way with the similarly named Entomological Society of America. See also List of entomology journals
https://en.wikipedia.org/wiki/ParaSail%20%28programming%20language%29
Parallel Specification and Implementation Language (ParaSail) is an object-oriented parallel programming language. Its design and ongoing implementation is described in a blog and on its official website. ParaSail uses a pointer-free programming model, where objects can grow and shrink, and value semantics are used for assignment. It has no global garbage collected heap. Instead, region-based memory management is used throughout. Types can be recursive, so long as the recursive components are declared optional. There are no global variables, no parameter aliasing, and all subexpressions of an expression can be evaluated in parallel. Assertions, preconditions, postconditions, class invariants, etc., are part of the standard syntax, using a Hoare-like notation. Any possible race conditions are detected at compile time. Initial design of ParaSail began in September 2009, by S. Tucker Taft. Both an interpreter using the ParaSail virtual machine, and an LLVM-based ParaSail compiler are available. Work stealing is used for scheduling ParaSail's light-weight threads. The latest version can be downloaded from the ParaSail website. Description The syntax of ParaSail is similar to Modula, but with a class-and-interface-based object-oriented programming model more similar to Java or C#. More recently, the parallel constructs of ParaSail have been adapted to other syntaxes, to produce Java-like, Python-like, and Ada-like parallel languages, dubbed, respectively, Javallel, Parython, and Sparkel (named after the Ada subset SPARK on which it is based). Compilers and interpreters for these languages are included with the ParaSail implementation. Examples The following is a Hello world program in ParaSail: func Hello_World(var IO) is IO.Println("Hello, World"); end func Hello_World; The following is an interface to a basic map module: interface BMap<Key_Type is Ordered<>; Element_Type is Assignable<>> is op "[]"() -> BMap; // Create an empty map func In
https://en.wikipedia.org/wiki/National%20Information%20Governance%20Board%20for%20Health%20and%20Social%20Care
The National Information Governance Board for Health and Social Care (NIGB) advised the United Kingdom government on information governance between 2008 and 2013. History The board was established under section 157 of the Health and Social Care Act 2008, with effect from October 2008, with a range of advisory functions relating to information governance. From January 2009, the NIGB also gained functions under section 251 of the NHS Act 2006 which had previously been held by the Patient Information Advisory Group (PIAG) until its abolition. These functions were to advise the Secretary of State for Health on the use of powers to set aside the common law duty of confidentiality in England, where identifiable patient information is needed and where consent is not practicable. From 1 April 2013, the NIGB's functions for monitoring and improving information governance practice transferred to the Care Quality Commission, which established a National Information Governance Committee to oversee this work. Functions relating to section 251 of the NHS Act 2006 (access to people's personal and confidential information for research purposes) were transferred to the Health Research Authority's Confidentiality Advisory Group. Terms of reference The key functions of the NIGB (excerpted from the legislation) were: to monitor the practice followed by relevant bodies in relation to the processing of relevant information; to keep the Secretary of State for Health, and such bodies as the Secretary of State for Health may designate by direction, informed about the practice being followed by relevant bodies in relation to the processing of relevant information; to publish guidance on the practice to be followed in relation to the processing of relevant information; to advise the Secretary of State for Health on particular matters relating to the processing of relevant information by any person; and to advise persons who process relevant information on such matters relating to t
https://en.wikipedia.org/wiki/Skin%20electrode%20gel
Skin electrode gels are used extensively to reduce electical resistance when using skin surface electrodes, both for medical instrumentation such as EEG, EMG and EKG, and for electrical stimulation treatments such as TENS, electrotherapy, electroconvulsive therapy and defibrillation. A typical electrode gel consists of a mixture of water, a thickener such as polyvinyl alcohol or carboxymethyl cellulose, and an electrolyte to increase conductivity. Some electrode gels have been shown to pose a potential fire risk when high-energy pulses are used, as for example when defibrillators are used. Skin electrode gels are also used by users of erotic electrostimulation to avoid skin burning.