source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/BKL%20singularity
A Belinski–Khalatnikov–Lifshitz (BKL) singularity is a model of the dynamic evolution of the universe near the initial gravitational singularity, described by an anisotropic, chaotic solution of the Einstein field equation of gravitation. According to this model, the universe is chaotically oscillating around a gravitational singularity in which time and space become equal to zero or, equivalently, the spacetime curvature becomes infinitely big. This singularity is physically real in the sense that it is a necessary property of the solution, and will appear also in the exact solution of those equations. The singularity is not artificially created by the assumptions and simplifications made by the other special solutions such as the Friedmann–Lemaître–Robertson–Walker, quasi-isotropic, and Kasner solutions. The model is named after its authors Vladimir Belinski, Isaak Khalatnikov, and Evgeny Lifshitz, then working at the Landau Institute for Theoretical Physics. The picture developed by BKL has several important elements. These are: Near the singularity the evolution of the geometry at different spatial points decouples so that the solutions of the partial differential equations can be approximated by solutions of ordinary differential equations with respect to time for appropriately defined spatial scale factors. This is called the BKL conjecture. For most types of matter the effect of the matter fields on the dynamics of the geometry becomes negligible near the singularity. Or, in the words of John Wheeler, "matter doesn't matter" near a singularity. The original BKL work posed a negligible effect for all matter but later they theorized that "stiff matter" (equation of state p = ε) equivalent to a massless scalar field can have a modifying effect on the dynamics near the singularity. The ordinary differential equations describing the asymptotics come from a class of spatially homogeneous solutions which constitute the Mixmaster dynamics: a complicated oscillat
https://en.wikipedia.org/wiki/Black%20hat%20%28computer%20security%29
A black hat (black hat hacker or blackhat) is a computer hacker who violates laws or typical ethical standards for nefarious purposes, such as cybercrime, cyberwarfare or malice. The term originates from 1950s westerns, when bad guys typically wore black hats and good guys white hats. A black hat is contrasted with a white hat. Additionally, there is a third category which is called a grey hat, a person who hacks with good intentions but sometimes without permission. Description Criminals who intentionally enter computer networks with malicious intent are known as "black hat hackers". They may distribute malware that steals data (particularly login credentials), financial information, or personal information (such as passwords or credit card numbers). This information is often then sold on the dark web. Additionally, malware can also be used to hold computers hostage or destroy files. Some hackers modify or destroy data in addition to stealing it. Even though hacking has become an important tool for governments to gather intelligence, black hats still tend to work alone or with organized crime groups to make easy money. Black hat hackers can be novices or experienced criminals. They are usually knowledgeable about infiltrating computer networks and getting around security protocols. Furthermore, they also create malware, a form of software that lets them access computer networks, monitor their victims' online activities, or lock their victims' devices. Black hat hackers can be involved in cyber espionage or protest, in addition to pursuing personal or financial gain. In fact, for some hackers, cybercrime may be an addictive experience. History The term "black hat" comes from the color scheme used in westerns from the 1950s, in which the bad guys wore black hats, and the good guys wore white or other light colors. One example of black hat activity was the WannaCry ransomware in May 2017. Around 400,000 computers in 150 countries were infected by it in the fir
https://en.wikipedia.org/wiki/Plasma%20protein%20binding
Plasma protein binding refers to the degree to which medications attach to blood proteins within the blood plasma. A drug's efficacy may be affected by the degree to which it binds. The less bound a drug is, the more efficiently it can traverse or diffuse through cell membranes. Common blood proteins that drugs bind to are human serum albumin, lipoprotein, glycoprotein, and α, β‚ and γ globulins. Binding (drug distribution) A drug in blood exists in two forms: bound and unbound. Depending on a specific drug's affinity for plasma proteins, a proportion of the drug may become bound to the proteins, with the remainder being unbound. If the protein binding is reversible, then a chemical equilibrium will exist between the bound and unbound states, such that: Protein + drug ⇌ Protein-drug complex Notably, it is the unbound fraction which exhibits pharmacologic effects. It is also the fraction that may be metabolized and/or excreted. For example, the "fraction bound" of the anticoagulant warfarin is 97%. This means that out of the amount of warfarin in the blood, 97% is bound to plasma proteins. The remaining 3% (the fraction unbound) is the fraction that is actually active and may be excreted. Protein binding can influence the drug's biological half-life. The bound portion may act as a reservoir or depot from which the drug is slowly released as the unbound form. Since the unbound form is being metabolized and/or excreted from the body, the bound fraction will be released in order to maintain equilibrium. Since albumin is alkalotic, acidic and neutral drugs will primarily bind to albumin. If albumin becomes saturated, then these drugs will bind to lipoprotein. Basic drugs will bind to the acidic alpha-1 acid glycoprotein. This is significant because various medical conditions may affect the levels of albumin, alpha-1 acid glycoprotein, and lipoproteins. Impact of the altered protein binding Only the unbound fraction of the drug undergoes metabolism in the liver a
https://en.wikipedia.org/wiki/Ultrafiltration%20%28kidney%29
In renal physiology, ultrafiltration occurs at the barrier between the blood and the filtrate in the glomerular capsule (Bowman's capsule) in the kidneys. As in nonbiological examples of ultrafiltration, pressure (in this case blood pressure) and concentration gradients lead to a separation through a semipermeable membrane (provided by the podocytes). The Bowman's capsule contains a dense capillary network called the glomerulus. Blood flows into these capillaries through the afferent arterioles and leaves through the efferent arterioles. The high hydrostatic pressure forces small molecules in the tubular fluid such as water, glucose, amino acids, sodium chloride and urea through the filter, from the blood in the glomerular capsule across the basement membrane of the Bowman's capsule and into the renal tubules. This process is called ultrafiltration; the resulting fluid, virtually free of large proteins and blood cells, is referred to as glomerular filtrate, or ultrafiltrate. Further modification of ultrafiltrate, by reabsorption and secretion, transforms it into urine. Glomerular pressure is about 75 millimeters of mercury (10 kPa). It is opposed by osmotic pressure (30 mmHg, 4.0 kPa) and hydrostatic pressure (20 mmHg, 2.7 kPa) of solutes present in capsular space. This difference in pressure is called effective pressure (25 mmHg, 3.3 kPa). In hemodialysis centers, ultrafiltration takes place in a hemofilter on the hemodialysis machines, when the blood pressure is greater than the dialysate pressure (difference = transmembrane pressure (TMP)). This removes fluid from the blood while keeping its blood cells intact. Selectivity The structures of the layers of the glomerulus determine their permeability-selectivity (permselectivity). For instance, small ions such as sodium and potassium pass freely, while larger plasma proteins, such as hemoglobin tetramers, haptoglobin bound hemoglobin and albumin have practically no permeability at all. Also, negatively charged
https://en.wikipedia.org/wiki/Weyl%20algebra
In abstract algebra, the Weyl algebra is the ring of differential operators with polynomial coefficients (in one variable), namely expressions of the form More precisely, let F be the underlying field, and let F[X] be the ring of polynomials in one variable, X, with coefficients in F. Then each fi lies in F[X]. ∂X is the derivative with respect to X. The algebra is generated by X and ∂X. The Weyl algebra is an example of a simple ring that is not a matrix ring over a division ring. It is also a noncommutative example of a domain, and an example of an Ore extension. The Weyl algebra is isomorphic to the quotient of the free algebra on two generators, X and Y, by the ideal generated by the element The Weyl algebra is the first in an infinite family of algebras, also known as Weyl algebras. The n-th Weyl algebra, An, is the ring of differential operators with polynomial coefficients in n variables. It is generated by Xi and ∂Xi, . Weyl algebras are named after Hermann Weyl, who introduced them to study the Heisenberg uncertainty principle in quantum mechanics. It is a quotient of the universal enveloping algebra of the Heisenberg algebra, the Lie algebra of the Heisenberg group, by setting the central element of the Heisenberg algebra (namely [X,Y]) equal to the unit of the universal enveloping algebra (called 1 above). The Weyl algebra is also referred to as the symplectic Clifford algebra. Weyl algebras represent the same structure for symplectic bilinear forms that Clifford algebras represent for non-degenerate symmetric bilinear forms. Generators and relations One may give an abstract construction of the algebras An in terms of generators and relations. Start with an abstract vector space V (of dimension 2n) equipped with a symplectic form ω. Define the Weyl algebra W(V) to be where T(V) is the tensor algebra on V, and the notation means "the ideal generated by". In other words, W(V) is the algebra generated by V subject only to the relation . Th
https://en.wikipedia.org/wiki/FreeS/WAN
FreeS/WAN, for Free Secure Wide-Area Networking, was a free software project, which implemented a reference version of the IPsec network security layer for Linux. The project goal of ubiquitous opportunistic encryption of Internet traffic was not realized, although it did contribute to general Internet encryption. The project was founded by John Gilmore, and administered for most of its duration by Hugh Daniel. John Ioannidis and Angelos Keromytis started the codebase while outside the United States of America prior to autumn 1997. Technical lead for the project was Henry Spencer, and later Michael Richardson. The IKE keying daemon (pluto) was maintained by D. Hugh Redelmeier while the IPsec kernel module (KLIPS) was maintained by Richard Guy Briggs. Sandy Harris was the main documentation person for most of the project, later Claudia Schmeing. The final FreeS/WAN version 2.06 was released on 22 April 2004. The earlier version 2.04 was forked to form two projects, Openswan and strongSwan. Openswan has since (2012) been forked to Libreswan. External links Project website Documentation Free security software History of software IPsec
https://en.wikipedia.org/wiki/Cubic%20IDE
Cubic IDE is a modular development environment (IDE) for AmigaOS (versions 3.5 and 3.9 only) and MorphOS. Its central editor is GoldED 8, which supports file type centric configuration. The specific features for developers include syntax highlighting for several programming languages (e.g. Hollywood), folding, a symbol browser, a project explorer, an installation assistant (to create installations), support for creating Rexx macros and documentation, makefile generation, dialogs to set compiler options, automatic completion of OS symbols and clickable compiler output (jump to error). Compiler integration is available for popular C/C++ compilers for the supported platforms: GCC, vbcc, SAS/C and StormC3. Several free compilers for AmigaOS3, PowerUP, WarpOS and MorphOS are included and integrated into the development environment. External links Amiga development software Integrated development environments MorphOS software
https://en.wikipedia.org/wiki/Graduated%20electronic%20decelerator
The graduated electronic decelerator (GED) is a device that delivers powerful electric shocks to the skin, described by the United Nations as torture, created by Matthew Israel for use on people at the Judge Rotenberg Center as part of the institution's behavior modification program. The school has since been condemned for torture by the United Nations special rapporteur on torture for its use of the GED and other inhumane punishments. In 2020, the device was banned in the United States by the Food and Drug Administration, however the ban was overturned in federal court a year later. In response, congress amended the Food, Drug, and Cosmetic Act to expand FDA's authority to ban such devices. Matthew Israel created the GED to replace the older punishments of spankings, pinches, and muscle squeezes, but continued to use restraints, sensory deprivation, and the withholding of food. These older punishments were often used in combination with the GED: For example, a student could be restrained to a board and then given several GED shocks in succession. While the school advertises its behavior modification program as safe, effective, and backed by science, these claims are disputed by independent experts. History The GED was created by Matthew Israel, the founder of Judge Rotenberg Center. Before it made use of electric shocks, the school used pinches, spankings, muscle squeezes, and a wide variety of other methods of aversive intervention including punitive restraints, sensory deprivation, and the withholding of food. Matthew Israel said that the school moved to electric shocks because “A lot of injuries were occurring” and also because it is more consistent. After the school began to use electric shocks as punishment, it phased out pinches, spankings, and muscle squeezes, but retained most other aversion interventions which were used alongside, and sometimes at the same time as, the electric shocks. For example, it was a common punishment to apply multiple GED shocks
https://en.wikipedia.org/wiki/T.50%20%28standard%29
ITU-T recommendation T.50 specifies the International Reference Alphabet (IRA), formerly International Alphabet No. 5 (IA5), a character encoding. ASCII is the U.S. variant of that character set. The original version from November 1988 corresponds to ISO 646. The current version is from September 1992. History At the beginning was the International Telegraph Alphabet No. 2 (ITA2), a five bits code. IA5 is an improvement based on seven bits bytes. Recommendation V.3 IA5 (1968) Initial version, superseded Recommendation V.3 IA5 (1972) Superseded Recommendation V.3 IA5 (1976-10) Superseded Recommendation V.3 IA5 (1980-11) Superseded Recommendation T.50 IA5 (1984-10) Superseded Recommendation T.50 IA5 (1988-11-25) Superseded Recommendation T.50 IRA (1992-09-18) In force Use This standard is referenced by other standards such as RFC 3966. It is also used by some analog modems such as Cisco ones. This standard is referenced by other standards such as RFC 3939 - Calling Line Identification for Voice Mail Messages. Character set The following table shows the IA5 character set. Each character is shown with the hex code of its Unicode equivalent. Standardisation Identical standard: ISO/IEC 646:1991 (Twinned) See also ITU T.51
https://en.wikipedia.org/wiki/Existentially%20closed%20model
In model theory, a branch of mathematical logic, the notion of an existentially closed model (or existentially complete model) of a theory generalizes the notions of algebraically closed fields (for the theory of fields), real closed fields (for the theory of ordered fields), existentially closed groups (for the theory of groups), and dense linear orders without endpoints (for the theory of linear orders). Definition A substructure M of a structure N is said to be existentially closed in (or existentially complete in) if for every quantifier-free formula φ(x1,…,xn,y1,…,yn) and all elements b1,…,bn of M such that φ(x1,…,xn,b1,…,bn) is realized in N, then φ(x1,…,xn,b1,…,bn) is also realized in M. In other words: If there is a tuple a1,…,an in N such that φ(a1,…,an,b1,…,bn) holds in N, then such a tuple also exists in M. This notion is often denoted . A model M of a theory T is called existentially closed in T if it is existentially closed in every superstructure N that is itself a model of T. More generally, a structure M is called existentially closed in a class K of structures (in which it is contained as a member) if M is existentially closed in every superstructure N that is itself a member of K. The existential closure in K of a member M of K, when it exists, is, up to isomorphism, the least existentially closed superstructure of M. More precisely, it is any extensionally closed superstructure M∗ of M such that for every existentially closed superstructure N of M, M∗ is isomorphic to a substructure of N via an isomorphism that is the identity on M. Examples Let σ = (+,×,0,1) be the signature of fields, i.e. + and × are binary function symbols and 0 and 1 are constant symbols. Let K be the class of structures of signature σ that are fields. If A is a subfield of B, then A is existentially closed in B if and only if every system of polynomials over A that has a solution in B also has a solution in A. It follows that the existentially closed members of K are
https://en.wikipedia.org/wiki/Wi-Fi%20Protected%20Setup
Wi-Fi Protected Setup (WPS) originally, Wi-Fi Simple Config, is a network security standard to create a secure wireless home network. Created by Cisco and introduced in 2006, the point of the protocol is to allow home users who know little of wireless security and may be intimidated by the available security options to set up Wi-Fi Protected Access, as well as making it easy to add new devices to an existing network without entering long passphrases. It is used by devices made by HP, Brother and Canon for their printers. WPS is a wireless method that is used to connect certain Wi-Fi devices such printers and security cameras to the Wi-Fi network without using any password. In addition, there is another way to connect called WPS Pin that is used by some devices to connect to the wireless network. Wi-Fi Protected Setup allows the owner of Wi-Fi privileges to block other users from using their household Wi-Fi. The owner can also allow people to use Wi-Fi. This can be changed by pressing the WPS button on the home router. A major security flaw was revealed in December 2011 that affects wireless routers with the WPS PIN feature, which most recent models have enabled by default. The flaw allows a remote attacker to recover the WPS PIN in a few hours with a brute-force attack and, with the WPS PIN, the network's WPA/WPA2 pre-shared key (PSK). Users have been urged to turn off the WPS PIN feature, although this may not be possible on some router models. Modes The standard emphasizes usability and security, and allows four modes in a home network for adding a new device to the network: PIN method In which a PIN has to be read from either a sticker or display on the new wireless device. This PIN must then be entered at the "representant" of the network, usually the network's access point. Alternately, a PIN provided by the access point may be entered into the new device. This method is the mandatory baseline mode and everything must support it. The Wi-Fi Direct specific
https://en.wikipedia.org/wiki/Universal%20polar%20stereographic%20coordinate%20system
The universal polar stereographic (UPS) coordinate system is used in conjunction with the universal transverse Mercator (UTM) coordinate system to locate positions on the surface of the Earth. Like the UTM coordinate system, the UPS coordinate system uses a metric-based cartesian grid laid out on a conformally projected surface. UPS covers the Earth's polar regions, specifically the areas north of 84°N and south of 80°S, which are not covered by the UTM grids, plus an additional 30 minutes of latitude extending into UTM grid to provide some overlap between the two systems. In the polar regions, directions can become complicated, with all geographic north–south lines converging at the poles. The difference between UPS grid north and true north can therefore be anything up to 180°—in some places, grid north is true south, and vice versa. UPS grid north is arbitrarily defined as being along the prime meridian in the Antarctic and the 180th meridian in the Arctic; thus, east and west on the grids when moving directly away from the pole are along the 90°E and 90°W meridians respectively. Projection system As the name indicates, the UPS system uses a stereographic projection. Specifically, the projection used in the system is a secant version based on an elliptical model of the earth. The scale factor at each pole is adjusted to 0.994 so that the latitude of true scale is 81.11451786859362545° (about 81° 06' 52.3") North and South. The scale factor inside the regions at latitudes higher than this parallel is too small, whereas the regions at latitudes below this line have scale factors that are too large, reaching 1.0016 at 80° latitude. The scale factor at the origin (the poles) is adjusted to minimize the overall distortion of scale within the mapped region. As with the Mercator projection, the region near the tangent (or secant) point on a Stereographic map remains very close to true scale for an angular distance of a few degrees. In the ellipsoidal model, a stereog
https://en.wikipedia.org/wiki/Guided%20local%20search
Guided local search is a metaheuristic search method. A meta-heuristic method is a method that sits on top of a local search algorithm to change its behavior. Guided local search builds up penalties during a search. It uses penalties to help local search algorithms escape from local minima and plateaus. When the given local search algorithm settles in a local optimum, GLS modifies the objective function using a specific scheme (explained below). Then the local search will operate using an augmented objective function, which is designed to bring the search out of the local optimum. The key is in the way that the objective function is modified. Overview Solution features To apply GLS, solution features must be defined for the given problem. Solution features are defined to distinguish between solutions with different characteristics, so that regions of similarity around local optima can be recognized and avoided. The choice of solution features depends on the type of problem, and also to a certain extent on the local search algorithm. For each feature a cost function is defined. Each feature is also associated with a penalty (initially set to 0) to record the number of occurrences of the feature in local minima. The features and costs often come directly from the objective function. For example, in the traveling salesman problem, “whether the tour goes directly from city X to city Y” can be defined to be a feature. The distance between X and Y can be defined to be the cost. In the SAT and weighted MAX-SAT problems, the features can be “whether clause C satisfied by the current assignments”. At the implementation level, we define for each feature an Indicator Function indicating whether the feature is present in the current solution or not. is 1 when solution exhibits property , 0 otherwise. Selective penalty modifications GLS computes the utility of penalising each feature. When the local search algorithm returns a local minimum x, GLS penalizes all tho
https://en.wikipedia.org/wiki/Encyclopedic%20Dictionary%20of%20Mathematics
The Encyclopedic Dictionary of Mathematics is a translation of the Japanese . The editor of the first and second editions was Shokichi Iyanaga; the editor of the third edition was Kiyosi Itô; the fourth edition was edited by the Mathematical Society of Japan. Editions ; paperback version of the 1987 edition
https://en.wikipedia.org/wiki/Binary%20quadratic%20form
In mathematics, a binary quadratic form is a quadratic homogeneous polynomial in two variables where a, b, c are the coefficients. When the coefficients can be arbitrary complex numbers, most results are not specific to the case of two variables, so they are described in quadratic form. A quadratic form with integer coefficients is called an integral binary quadratic form, often abbreviated to binary quadratic form. This article is entirely devoted to integral binary quadratic forms. This choice is motivated by their status as the driving force behind the development of algebraic number theory. Since the late nineteenth century, binary quadratic forms have given up their preeminence in algebraic number theory to quadratic and more general number fields, but advances specific to binary quadratic forms still occur on occasion. Pierre Fermat stated that if p is an odd prime then the equation has a solution iff , and he made similar statement about the equations , , and . and so on are quadratic forms, and the theory of quadratic forms gives a unified way of looking at and proving these theorems. Another instance of quadratic forms is Pell's equation . Binary quadratic forms are closely related to ideals in quadratic fields, this allows the class number of a quadratic field to be calculated by counting the number of reduced binary quadratic forms of a given discriminant. The classical theta function of 2 variables is , if is a positive definite quadratic form then is a theta function. Equivalence Two forms f and g are called equivalent if there exist integers such that the following conditions hold: For example, with and , , , and , we find that f is equivalent to , which simplifies to . The above equivalence conditions define an equivalence relation on the set of integral quadratic forms. It follows that the quadratic forms are partitioned into equivalence classes, called classes of quadratic forms. A class invariant can mean either a funct
https://en.wikipedia.org/wiki/Ordinal%20data
Ordinal data is a categorical, statistical data type where the variables have natural, ordered categories and the distances between the categories are not known. These data exist on an ordinal scale, one of four levels of measurement described by S. S. Stevens in 1946. The ordinal scale is distinguished from the nominal scale by having a ranking. It also differs from the interval scale and ratio scale by not having category widths that represent equal increments of the underlying attribute. Examples of ordinal data A well-known example of ordinal data is the Likert scale. An example of a Likert scale is: Examples of ordinal data are often found in questionnaires: for example, the survey question "Is your general health poor, reasonable, good, or excellent?" may have those answers coded respectively as 1, 2, 3, and 4. Sometimes data on an interval scale or ratio scale are grouped onto an ordinal scale: for example, individuals whose income is known might be grouped into the income categories $0–$19,999, $20,000–$39,999, $40,000–$59,999, ..., which then might be coded as 1, 2, 3, 4, .... Other examples of ordinal data include socioeconomic status, military ranks, and letter grades for coursework. Ways to analyse ordinal data Ordinal data analysis requires a different set of analyses than other qualitative variables. These methods incorporate the natural ordering of the variables in order to avoid loss of power. Computing the mean of a sample of ordinal data is discouraged; other measures of central tendency, including the median or mode, are generally more appropriate. General Stevens (1946) argued that, because the assumption of equal distance between categories does not hold for ordinal data, the use of means and standard deviations for description of ordinal distributions and of inferential statistics based on means and standard deviations was not appropriate. Instead, positional measures like the median and percentiles, in addition to descriptive statistics
https://en.wikipedia.org/wiki/Renal%E2%80%93hepatic%E2%80%93pancreatic%20dysplasia
Renal–hepatic–pancreatic dysplasia is an autosomal recessive congenital disorder characterized by pancreatic fibrosis, renal dysplasia and hepatic dysgenesis. It is usually fatal soon after birth.An association with NPHP3 has been described. It was characterized in 1959.
https://en.wikipedia.org/wiki/Kyubey
is the main antagonist of the 2011 anime series Puella Magi Madoka Magica and its 2013 sequel The Rebellion Story. They are a "messenger of magic" that can grant any wish to a certain girl, on the condition that she become a magical girl and fight against witches. It is later revealed that their true identity is Incubator. Kyubey is considered the main villain of the Puella Magi Madoka Magica series, as they tricked Madoka Kaname and the other protagonists in order to turn them into magical girls and, later on, into witches. Kyubey's true goal is countering entropy and staving off the heat death of the universe. Creation and design Kyubey was created and designed by writer Gen Urobuchi. As one of the primary villains in the series, producer Atsuhiro Iwakami stated that "the mash-up of cuteness and darkness is the central theme to Puella Magi Madoka Magica, and Kyubey is an epitome of that theme." A central goal in Urobuchi's writing was to highlight the moral and ethical dissonance between Kyubey and the young girls, which was done through actions in the series such as Kyubey eating its own corpse in order to recycle energy. Urobuchi compared Kyubey to monsters occurring in the works of horror fiction author H. P. Lovecraft, commenting of the character: "He is not evil, it is his lack of feelings that make him scary." Appearances In Madoka Magica Kyubey is sexless (though it apparently identifies as a male) extraterrestrial cat-like being posing as a familiar who can grant any wish to a certain girl, on the condition that she becomes a magical girl and fight against witches. When the chosen girl makes a contract with him, he extracts her soul and places it inside a soul gem, reconstructing their body into a shell that is more resilient in order to fight witches. He constantly tries to get Madoka Kaname to make a contract with him, as she allegedly possesses great magical potential within her that would allow her to become the most powerful magical girl. He can o
https://en.wikipedia.org/wiki/2003%20United%20States%20smallpox%20vaccination%20campaign
The 2003 United States smallpox vaccination campaign was a vaccination program announced by the White House on 13 December 2002 as preparedness for bioterrorism using smallpox virus. The campaign aimed to provide the smallpox vaccine to those who would respond to an attack, establishing Smallpox Response Teams and using DryVax (containing the NYCBOH strain) to mandatorily vaccinate half a million American military personnel, followed by half a million health care worker volunteers by January 2004. The first vaccine was administered to then-President George W. Bush. The campaign ended early in June 2003, with only 38,257 civilian health care workers vaccinated, after several hospitals refused to participate due to the risk of the live virus infecting vulnerable patients and skepticism about the risks of an attack, and after over 50 heart complications were reported by the CDC. That August, the US Institute of Medicine (IOM) criticized the programme for its costs and not considering other bioterrorism control measures such as surveillance. The adverse cardiac events, including two deaths, were however unlikely to have been caused by the vaccine. A 2005 IOM report noted that some of the problems of the campaign stemmed from administration officials overruling scientific advice on the numbers who should be vaccinated and a lack of communication by the CDC of the public health need, though it found that the campaign had increased general preparedness for sudden occurrences of infectious diseases like that year's monkeypox outbreak and the 2002–2004 SARS outbreak.
https://en.wikipedia.org/wiki/Begonia%20%C3%97%20tuberhybrida
Tuberous begonias (Begonia × tuberhybrida Voss, also known as the Tuberhybrida Group or the Tuberosa Group) are a group of Begonia cultivars, sometimes regarded as some of the most spectacular of the genus. One of the first hybrids produced was B. x sedenii in 1870, a cross between B. boliviensis, collected by botanist Richard Pearce and a species from the Andes. Another species from Peru, B. davisii (named after Walter Davis), was also used in early breeding. Nomenclature The University of Vermont recognises the following 13 groups: (S) Single – large single flowers, four usually flat tepals (flower part undistinguishable as sepal or petal) (Fr) Frilled, Crispa – large single flowers, tepal margins frilled or ruffled (Cr) Cristata, Crested – large single flowers, frilled or tufted center of tepals (N) Narcissiflora, Daffodil-flowered – large more or less double flowers, central tepals form "trumpet" (C) Camellia, Camelliflora – large double flowers resembling camellias, unruffled, solid colors (RC) Ruffled Camellia – camellia flowers ruffled on edges (R) Rosebud, Rosiflora – large double flowers with rose bud-like center (Car) Carnation, Fimbriata Plena – large double carnation-like flowers, tepals fringed on margins (P) Picotee – large usually double flowers like camellias, tepals with different color on margin blending with other color (M) Marginata – like Picotee only distinct non-blending line of color on margins (Mar) Marmorata, Marbled – like Camellia but rose-colored, blotched or spotted with white (HB) Hanging Basket, Pendula – stems trailing or pendant, large to small flowers single or double (Mul) Multiflora – low, bushy, compact plants with many small single or double flowers Other systems confusingly use binomial nomenclature to refer to flower types, by the terms Begonia grandiflora, Begonia multiflora and Begonia pendula. The last two correspond to groups 13 and 12, respectively. The first type, grandiflora (large flowered), is then
https://en.wikipedia.org/wiki/Peanut%20butter
Peanut butter is a food paste or spread made from ground, dry-roasted peanuts. It commonly contains additional ingredients that modify the taste or texture, such as salt, sweeteners, or emulsifiers. Consumed in many countries, it is the most commonly used of the nut butters, a group that also includes cashew butter and almond butter (though peanuts are not nuts, peanut butter is culinarily considered a nut butter). Peanut butter is a nutrient-rich food containing high levels of protein, several vitamins, and dietary minerals. It is typically served as a spread on bread, toast, or crackers, and used to make sandwiches (notably the peanut butter and jelly sandwich). It is also used in a number of breakfast dishes and desserts, such as granola, smoothies, crepes, cookies, brownies, or croissants. History The earliest references to peanut butter can be traced to Aztec and Inca civilizations, who ground roasted peanuts into a paste. However, several people can be credited with the invention of modern peanut butter and the processes involved in making it. The US National Peanut Board credits three modern inventors with the earliest patents related to the production of modern peanut butter. Marcellus Gilmore Edson of Montreal, Quebec, Canada, obtained the first patent for a method of producing peanut butter from roasted peanuts using heated surfaces in 1884. Edson's cooled product had "a consistency like that of butter, lard, or ointment" according to his patent application which described a process of milling roasted peanuts until the peanuts reached "a fluid or semi-fluid state". He mixed sugar into the paste to harden its consistency. A businessman from St. Louis named George Bayle produced and sold peanut butter in the form of a snack food in 1894. By 1917, American consumers used peanut products during periods of meat rationing, with government promotions of "meatless Mondays" when peanut butter was a favored choice. John Harvey Kellogg, known for his line of pr
https://en.wikipedia.org/wiki/Comparison%20of%20Q%26A%20sites
The following is a list of websites that follow a question-and-answer format. The list contains only websites for which an article exists, dedicated either wholly or at least partly to the websites. For the humor "Q&A site" format first popularized by Forum 2000 and The Conversatron, see Q&A comedy website. See also Comparison of civic technology platforms Comparison of Internet forum software Educational Technology List of Internet forums Knowledge market Q&A software – includes a comparison of self-hostable Q&A software
https://en.wikipedia.org/wiki/International%20Psychopharmacology%20Algorithm%20Project
The International Psychopharmacology Algorithm Project (IPAP) is a non-profit corporation whose purpose is to "enable, enhance, and propagate" use of algorithms for the treatment of some Axis I psychiatric disorders. Kenneth O Jobson founded the Project. The Dean Foundation provides funding. IPAP has organized and supported several international conferences on psychopharmacology algorithms. It has also supported the creation of several algorithms based on expert opinion. It is now in the process of creating "evidence-based algorithms," that is algorithms created by experts and annotated with the evidence that leads to these algorithms. A schizophrenia algorithm has been created and one on Post Traumatic Stress Disorder (PTSD) was released in July 2005. A general anxiety disorder (GAD) algorithm was released in 2006. Periodic updates of the algorithms are released as the basis of evidence changes. In addition, the algorithms are being translated into various non-English languages (Chinese, Japanese, Spanish, and Thai) as the availability of translators permits.
https://en.wikipedia.org/wiki/Paul%20Seymour%20%28mathematician%29
Paul D. Seymour (born 26 July 1950) is a British mathematician known for his work in discrete mathematics, especially graph theory. He (with others) was responsible for important progress on regular matroids and totally unimodular matrices, the four colour theorem, linkless embeddings, graph minors and structure, the perfect graph conjecture, the Hadwiger conjecture, claw-free graphs, χ-boundedness, and the Erdős–Hajnal conjecture. Many of his recent papers are available from his website. Seymour is currently the Albert Baldwin Dod Professor of Mathematics at Princeton University. He won a Sloan Fellowship in 1983, and the Ostrowski Prize in 2003; and (sometimes with others) won the Fulkerson Prize in 1979, 1994, 2006 and 2009, and the Pólya Prize in 1983 and 2004. He received an honorary doctorate from the University of Waterloo in 2008, one from the Technical University of Denmark in 2013, and one from the École normale supérieure de Lyon in 2022. He was an invited speaker in the 1986 International Congress of Mathematicians and a plenary speaker in the 1994 International Congress of Mathematicians. He became a Fellow of the Royal Society in 2022. Early life Seymour was born in Plymouth, Devon, England. He was a day student at Plymouth College, and then studied at Exeter College, Oxford, gaining a BA degree in 1971, and D.Phil in 1975. Career From 1974 to 1976 he was a college research fellow at University College of Swansea, and then returned to Oxford for 1976–1980 as a Junior Research Fellow at Merton College, Oxford, with the year 1978–79 at University of Waterloo. He became an associate and then a full professor at Ohio State University, Columbus, Ohio, between 1980 and 1983, where he began research with Neil Robertson, a fruitful collaboration that continued for many years. From 1983 until 1996, he was at Bellcore (Bell Communications Research), Morristown, New Jersey (now Telcordia Technologies). He was also an adjunct professor at Rutgers University f
https://en.wikipedia.org/wiki/Hardware%20code%20page
In computing, a hardware code page (HWCP) refers to a code page supported natively by a hardware device such as a display adapter or printer. The glyphs to present the characters are stored in the alphanumeric character generator's resident read-only memory (like ROM or flash) and are thus not user-changeable. They are available for use by the system without having to load any font definitions into the device first. Startup messages issued by a PC's System BIOS or displayed by an operating system before initializing its own code page switching logic and font management and before switching to graphics mode are displayed in a computer's default hardware code page. Code page assignments In North American IBM-compatible PCs, the hardware code page of the display adapter is typically code page 437. However, various portable machines as well as (Eastern) European, Arabic, Middle Eastern and Asian PCs used a number of other code pages as their hardware code page, including code page 100 ("Hebrew"), 151 ("Nafitha Arabic"), 667 ("Mazovia"), 737 ("Greek"), 850 ("Multilingual"), encodings like "Roman-8", "Kamenický", "KOI-8", "MIK", and others. Most display adapters support a single 8-bit hardware code page only. The bitmaps were often stored in an EPROM in a DIP socket. At most, the hardware code page to be activated was user-selectable via jumpers, configuration EEPROMs or CMOS setup. However, some of the display adapters designed for Eastern European, Arabic and Hebrew PCs supported multiple software-switchable hardware code pages, also named font pages, selectable via I/O ports or additional BIOS functions. In contrast to this, printers frequently support several user-switchable character sets, often including various variants of the 7-bit ISO/IEC 646 character sets such as code page 367 ("ISO/IEC 646-US / ASCII"), sometimes also a couple of 8-bit code pages like code page 437, 850, 851, 852, 853, 855, 857, 860, 861, 863, 865, and 866. Printers for the Eastern European
https://en.wikipedia.org/wiki/Free%20Software%20Foundation%20of%20India
The Free Software Foundation of India (FSFI) is the Indian sister organisation to the US-based Free Software Foundation. It was founded in Thiruvananthapuram (Trivandrum) (the capital of Kerala) in 2001 as a non-profit Company. The FSFI advocates to promote the use and development of free software in India. This includes educating people about free software, including how it can help the economy of a developing country like India. FSF India regards non-free software as not a solution, but a problem to be solved. Free software is sometimes locally called swatantra software in India. In 2003, after meeting with FSF founder Richard Stallman, the President of India Dr. Abdul Kalam urged Indian computer scientists and professionals to use free and open-source software in research and development. Logo The left side of the FSF India logo resembles the famed Ashoka Chakra, that also appears on the Indian national flag, while its right half is a depiction of a compact disc. The combination of both symbols is intended to draw attention to the similarity between political freedom and free software, the latter of which the FSF promotes. The Ashoka Chakra, in addition to being a long-standing Buddhist symbol depicting the wheel of time, also resembles Charkha of India's Freedom Struggle and thus in the current context, may symbolize the Indian independence movement which resulted in India's freedom from British occupation. See also Free Software Foundation Free Software Foundation Europe Free Software Movement of India Free Software Movement
https://en.wikipedia.org/wiki/Web%20hosting%20service
A web hosting service is a type of Internet hosting service that hosts websites for clients, i.e. it offers the facilities required for them to create and maintain a site and makes it accessible on the World Wide Web. Companies providing web hosting services are sometimes called web hosts. Typically, web hosting requires the following: one or more servers to act as the host(s) for the sites; servers may be physical or virtual colocation for the server(s), providing physical space, electricity, and Internet connectivity; Domain Name System configuration to define name(s) for the sites and point them to the hosting server(s); a web server running on the host; for each site hosted on the server: space on the server(s) to hold the files making up the site site-specific configuration often, a database; software and credentials allowing the client to access these, enabling them to create, configure, and modify the site; email connectivity allowing the host and site to send email to the client. History Until 1991, the Internet was restricted to use only "... for research and education in the sciences and engineering..." and was used for email, telnet, FTP and USENET traffic—but only a tiny number of web pages. The World Wide Web protocols had only just been written and not until the end of 1993 would there be a graphical web browser for Mac or Windows computers. Even after there was some opening up of Internet access, the situation was confused until 1995. To host a website on the internet, an individual or company would need their own computer or server. As not all companies had the budget or expertise to do this, web hosting services began to offer to host users' websites on their own servers, without the client needing to own the necessary infrastructure required to operate the website. The owners of the websites, also called webmasters, would be able to create a website that would be hosted on the web hosting service's server and published to the web
https://en.wikipedia.org/wiki/Abhyankar%27s%20inequality
Abhyankar's inequality is an inequality involving extensions of valued fields in algebra, introduced by . Abhyankar's inequality states that for an extension K/k of valued fields, the transcendence degree of K/k is at least the transcendence degree of the residue field extension plus the rank of the quotient of the valuation groups; here the rank of an abelian group is defined as .
https://en.wikipedia.org/wiki/Point%20accepted%20mutation
A point accepted mutation — also known as a PAM — is the replacement of a single amino acid in the primary structure of a protein with another single amino acid, which is accepted by the processes of natural selection. This definition does not include all point mutations in the DNA of an organism. In particular, silent mutations are not point accepted mutations, nor are mutations that are lethal or that are rejected by natural selection in other ways. A PAM matrix is a matrix where each column and row represents one of the twenty standard amino acids. In bioinformatics, PAM matrices are sometimes used as substitution matrices to score sequence alignments for proteins. Each entry in a PAM matrix indicates the likelihood of the amino acid of that row being replaced with the amino acid of that column through a series of one or more point accepted mutations during a specified evolutionary interval, rather than these two amino acids being aligned due to chance. Different PAM matrices correspond to different lengths of time in the evolution of the protein sequence. Biological background The genetic instructions of every replicating cell in a living organism are contained within its DNA. Throughout the cell's lifetime, this information is transcribed and replicated by cellular mechanisms to produce proteins or to provide instructions for daughter cells during cell division, and the possibility exists that the DNA may be altered during these processes. This is known as a mutation. At the molecular level, there are regulatory systems that correct most — but not all — of these changes to the DNA before it is replicated. One of the possible mutations that occurs is the replacement of a single nucleotide, known as a point mutation. If a point mutation occurs within an expressed region of a gene, an exon, then this will change the codon specifying a particular amino acid in the protein produced by that gene. Despite the redundancy in the genetic code, there is a possibility t
https://en.wikipedia.org/wiki/Merkle%20tree
In cryptography and computer science, a hash tree or Merkle tree is a tree in which every "leaf" (node) is labelled with the cryptographic hash of a data block, and every node that is not a leaf (called a branch, inner node, or inode) is labelled with the cryptographic hash of the labels of its child nodes. A hash tree allows efficient and secure verification of the contents of a large data structure. A hash tree is a generalization of a hash list and a hash chain. Demonstrating that a leaf node is a part of a given binary hash tree requires computing a number of hashes proportional to the logarithm of the number of leaf nodes in the tree. Conversely, in a hash list, the number is proportional to the number of leaf nodes itself. A Merkle tree is therefore an efficient example of a cryptographic commitment scheme, in which the root of the tree is seen as a commitment and leaf nodes may be revealed and proven to be part of the original commitment. The concept of a hash tree is named after Ralph Merkle, who patented it in 1979. Uses Hash trees can be used to verify any kind of data stored, handled and transferred in and between computers. They can help ensure that data blocks received from other peers in a peer-to-peer network are received undamaged and unaltered, and even to check that the other peers do not lie and send fake blocks. Hash trees are used in hash-based cryptography. Hash trees are also used in the InterPlanetary File System (IPFS), Btrfs and ZFS file systems (to counter data degradation); Dat protocol; Apache Wave protocol; Git and Mercurial distributed revision control systems; the Tahoe-LAFS backup system; Zeronet; the Bitcoin and Ethereum peer-to-peer networks; the Certificate Transparency framework; the Nix package manager and descendants like GNU Guix; and a number of NoSQL systems such as Apache Cassandra, Riak, and Dynamo. Suggestions have been made to use hash trees in trusted computing systems. The initial Bitcoin implementation of Merkle
https://en.wikipedia.org/wiki/AmigaOS%204%20version%20history
A new version of AmigaOS was released on December 24, 2006 after five years of development by Hyperion Entertainment (Belgium) under license from Amiga, Inc. for AmigaOne registered users. During the five years of development, users of AmigaOne machines could download from Hyperion repository Pre-Release Versions of AmigaOS 4.0 as long as these were made available. As witnessed by many users into Amiga discussion forum sites, these versions were stable and reliable, despite the fact that they are technically labeled as "pre-releases". The last stable version of AmigaOS 4.0 for AmigaOne computers is the "July 2007 Update", released for download 18 July 2007 to the registered users of AmigaOne machines. AmigaOS 4 Classic was released commercially for older Amiga computers with CyberstormPPC and BlizzardPPC accelerator cards in November 2007. It had previously been available only to developers and beta-testers. 4.0 versions The new version is PowerPC-native, finally abandoning the Motorola 68k processor. AmigaOS 4.0 will run on some PowerPC hardware, which currently only includes A1200, A3000 and A4000 with PowerPC accelerator boards and AmigaOne motherboards. Amiga, Inc.'s distribution policies for AmigaOS 4.0 and any later versions require that for third-party hardware the OS must be bundled with it, with the sole exception of Amigas with Phase 5 PowerPC accelerator boards, for which the OS will be sold separately. AmigaOS 4.0 Final introduced a new memory system based on the slab allocator. Features, among others: Fully skinnable GUI Virtualized memory Integrated viewer for PDF and other document formats Support for PowerPC (native) and 68k (interpreted/JIT) applications New drivers for various hardware New memory allocation system Support for file sizes larger than 2 GB Integrated Picasso 96 2D Graphics API Integrated Warp3D 3D Graphics API Developer pre-releases AmigaOS 4.0 (Developer Pre-release) The developer pre-release version was a snapshot
https://en.wikipedia.org/wiki/Hutchinson%27s%20triad
Hutchinson's triad is named after Sir Jonathan Hutchinson (1828–1913). It is a common pattern of presentation for late congenital syphilis, and consists of three phenomena: interstitial keratitis, malformed teeth (Hutchinson incisors and mulberry molars), and eighth nerve deafness. There may also be a deformity on the nose known as saddle nose deformity.
https://en.wikipedia.org/wiki/Fujitsu%20VP2000
The VP2000 was the second series of vector supercomputers from Fujitsu. Announced in December 1988, they replaced Fujitsu's earlier FACOM VP Model E Series. The VP2000 was succeeded in 1995 by the VPP300, a massively parallel supercomputer with up to 256 vector processors. The VP2000 was similar in many ways to their earlier designs, and in turn to the Cray-1, using a register-based vector processor for performance. For additional performance the vector units supported a special multiply-and-add instruction that could retire two results per clock cycle. This instruction "chain" is particularly common in many supercomputer applications. Another difference is that the main scalar units of the processor ran at half the speed of the vector unit. According to Amdahl's Law computers tend to run at the speed of their slowest unit, and in this case unless the program spent most of its time in the vector units, the slower scalar performance would make it 1/2 the performance of a Cray-1 at the same speed. The reason for this seemingly odd "feature" is unclear. One of the major complaints about the earlier VP series was their limited memory bandwidth—while the machines themselves had excellent performance in the processors, they were often starved for data. For the VP2000 series this was addressed by adding a second load/store unit to the scalar units, doubling memory bandwidth. Several versions of the machines were sold at different price points. The low-end VP2100 ran at an 8 ns cycle time and delivered only 0.5 GFLOPS (about 4-8 times the performance of a Cray), while the VP2200 and VP2400 decreased the cycle time to 4 ns and delivered between 1.25 and 2.5 GFLOPS peak. The high-end VP2600 ran at 3.2 ns and delivered 5 GFLOPS. All of the models came in the /10 versions with a single scalar processor, or the /20 with a second, while the 2200 and 2400 also came in a /40 configuration with four. Due to the additional load/store units, adding additional scalar units improved
https://en.wikipedia.org/wiki/Primary%20growth
Primary growth in plants is growth that takes place from the tips of roots or shoots. It leads to lengthening of roots and stems and sets the stage for organ formation. It is distinguished from secondary growth that leads to widening. Plant growth takes place in well defined plant locations. Specifically, the cell division and differentiation needed for growth occurs in specialized structures called meristems. These consist of undifferentiated cells (meristematic cells) capable of cell division. Cells in the meristem can develop into all the other tissues and organs that occur in plants. These cells continue to divide until they differentiate and then lose the ability to divide. Thus, the meristems produce all the cells used for plant growth and function. At the tip of each stem and root, an apical meristem adds cells to their length, resulting in the elongation of both. Examples of primary growth are the rapid lengthening growth of seedlings after they emerge from the soil and the penetration of roots deep into the soil. Furthermore, all plant organs arise ultimately from cell divisions in the apical meristems, followed by cell expansion and differentiation. In contrast, a growth process that involves thickening of stems takes place within lateral meristems that are located throughout the length of the stems. The lateral meristems of larger plants also extend into the roots. This thickening is secondary growth and is needed to give mechanical support and stability to the plant. The functions of a plant's growing tips – its apical (or primary) meristems – include: lengthening through cell division and elongation; organising the development of leaves along the stem; creating platforms for the eventual development of branches along the stem; laying the groundwork for organ formation by providing a stock of undifferentiated or incompletely differentiated cells that later develop into fully differentiated cells, thereby ultimately allowing the "spatial deployment
https://en.wikipedia.org/wiki/Fat%20comma
The fat comma (also termed hash rocket in Ruby and a fat arrow in JavaScript) is a syntactic construction that appears in a position in a function call (or definition) where a comma would usually appear. The original usage refers to the ")letters:(" construction in ALGOL 60. Newer usage refers to the "=>" operator present in some programming languages. It is primarily associated with PHP, Ruby and Perl programming languages, which use it to declare hashes. Using a fat comma to bind key-value pairs in a hash, instead of using a comma, is considered an example of good idiomatic Perl. In CoffeeScript and TypeScript, the fat comma is used to declare a function that is bound to this. # a typical, idiomatic use of the fat comma in Perl my %hash = ( first_name => "Larry", last_name => "Wall", ); Subtleties ALGOL 60 The ALGOL "fat comma" is semantically identical to the comma. In particular, whether letter strings are used, and what their contents are, need not match between the definition of a function and its uses. The following are equivalent: S(s-5, T, P) S(s-5) t: (T) p: (P) S(s-5) Temperature: (T) Pressure: (P) Perl The "fat comma" forces the word to its left to be interpreted as a string. Thus, where this would produce a run-time error under strict (barewords are not allowed): %bad_example = ( bad_bareword, "not so cool" ); the following use of the fat comma would be legal and idiomatic: %good_example = ( converted_to_string => "very monkish" ); This is because the token converted_to_string would be converted to the string literal "converted_to_string" which is a legal argument in a hash key assignment. The result is easier-to-read code, with a stronger emphasis on the name-value pairing of associative arrays. PHP In PHP, the fat comma is termed a double arrow, and is used to specify key/value relationships when declaring an array. Unlike in Perl, the double arrow does not treat what comes before it as a bare word, but rather evaluates it. Hence,
https://en.wikipedia.org/wiki/Gauss%E2%80%93Legendre%20quadrature
In numerical analysis, Gauss–Legendre quadrature is a form of Gaussian quadrature for approximating the definite integral of a function. For integrating over the interval , the rule takes the form: where n is the number of sample points used, wi are quadrature weights, and xi are the roots of the nth Legendre polynomial. This choice of quadrature weights wi and quadrature nodes xi is the unique choice that allows the quadrature rule to integrate degree polynomials exactly. Many algorithms have been developed for computing Gauss–Legendre quadrature rules. The Golub–Welsch algorithm presented in 1969 reduces the computation of the nodes and weights to an eigenvalue problem which is solved by the QR algorithm. This algorithm was popular, but significantly more efficient algorithms exist. Algorithms based on the Newton–Raphson method are able to compute quadrature rules for significantly larger problem sizes. In 2014, Ignace Bogaert presented explicit asymptotic formulas for the Gauss–Legendre quadrature weights and nodes, which are accurate to within double-precision machine epsilon for any choice of n ≥ 21. This allows for computation of nodes and weights for values of n exceeding one billion. History Carl Friedrich Gauss was the first to derive the Gauss–Legendre quadrature rule, doing so by a calculation with continued fractions in 1814. He calculated the nodes and weights to 16 digits up to order n=7 by hand. Carl Gustav Jacob Jacobi discovered the connection between the quadrature rule and the orthogonal family of Legendre polynomials. As there is no closed-form formula for the quadrature weights and nodes, for many decades people were only able to hand-compute them for small n, and computing the quadrature had to be done by referencing tables containing the weight and node values. By 1942 these values were only known for up to n=16. Definition For integrating f over with Gauss–Legendre quadrature, the associated orthogonal polynomials are Legendre p
https://en.wikipedia.org/wiki/Bromochlorofluoromethane
Bromochlorofluoromethane or fluorochlorobromomethane, is a chemical compound and trihalomethane derivative with the chemical formula CHBrClF. As one of the simplest possible stable chiral compounds, it is useful for fundamental research into this area of chemistry. However, its relative instability to hydrolysis, and lack of suitable functional groups, made separation of the enantiomers of bromochlorofluoromethane especially challenging, and this was not accomplished until almost a century after it was first synthesised, in March 2005, though it has now been done by a variety of methods. More recent research using bromochlorofluoromethane has focused on its potential use for experimental measurement of parity violation, a major unsolved problem in quantum physics. See also Bromochlorodifluoromethane, used in fire extinguishers Bromochlorofluoroiodomethane, a theoretical derivative with iodine replacing the hydrogen
https://en.wikipedia.org/wiki/Bittering%20agent
A bittering agent is a flavoring agent added to a food or beverage to impart a bitter taste, possibly in addition to other effects. While many substances are bitter to a greater or lesser degree, a few substances are used specifically for their bitterness, especially to balance other flavors, such as sweetness. Notable beverage examples include caffeine, found naturally in tea and coffee and added to many soft drinks, hops in beer, and quinine in tonic water. Food examples include bitter melon, which may be mixed into a stir fry or soup for its bitter flavor. Potent bittering agents may also be added to dangerous products as aversive agents to make them foul tasting, so as to prevent accidental poisoning. Examples including anti-freeze, household cleaning products and pesticides such as slug pellets. In general dangerous products with bright colours, which may be appealing to children, often contain agents such as denatonium. However, the efficacy of using bittering agents for this purpose is not conclusive. Beer Prior to the introduction of hops, many other bitter herbs and flowers were used as bittering agents in beer, in a mixture called gruit, which could include dandelion, burdock root, marigold, horehound (the German name for horehound means "mountain hops"), ground ivy, and heather. Also bog myrtle. More recently, some Chinese and Okinawan beer uses bitter melon as a bittering agent. Other substances Various other substances are used, including: Aloin Gesho, used in Tej, Ethiopian honey wine Other uses Other prominent uses of bittering agents include: Bitters – used as digestifs or flavorings Dandelion and burdock – traditional British soft drink
https://en.wikipedia.org/wiki/Autoimmune%20oophoritis
Autoimmune oophoritis is a rare autoimmune disease where the body's own immune system attacks the ovaries. This causes the ovaries to have inflammation, atrophy, and fibrosis. Such changes in the ovaries can cause them to not function properly. This disease is caused by primary ovarian insufficiency (POI), where reproduction and hormonal function of the ovaries stops before the age of 40. Signs and symptoms Autoimmune oophoritis can present with a wide variety of symptoms. It begins with the main symptom, which is amenorrhea, where there is an irregular or no menstrual period at all. Other symptoms are related to ovarian cysts, and more common ones are also listed below. A variation of symptoms can occur together, however, that depends on the person and the severity of the disease. Primary amenorrhea This amenorrhea is more specific to menstruation never occurring before. Secondary amenorrhea This type of amenorrhea is where menstruation occurred once puberty began but then suddenly stopped later on. Infertility Cramping Bloating Nausea Vomiting Sex hormone deficiency Hot flushes Lower abdominal pain Fever Malaise Vaginal discharge Symptoms associated with other coexisting autoimmune conditions Causes The underlying cause of autoimmune oophoritis is yet unknown. However, it is known that it can co-exit with lupus, pernicious anemia, myasthenia gravis, or other autoimmune conditions. Autoimmune oophoritis can also be associated with autoimmune endocrinopathy syndrome type I and type II. These conditions can coexists because they are all autoimmune diseases, this is called polyautoimmunity. There are some environmental causes for autoimmune diseases that have been discovered in different studies. Those consist of exposures to chemicals in the environment, such as vinyl chloride, metals, mycotoxins, and organic compounds. Genetics As Autoimmune oophoritis often occurs in the setting of autoimmune polyendocrine syndromes (APS), mutations in AIRE or HLA
https://en.wikipedia.org/wiki/PlanetPhysics
PlanetPhysics was a virtual community with several Internet sites supported by a non-profit organization registered in the USA in an open science, open data, peer-to-peer review mode that aimed to help make physics, and related mathematics, knowledge much more accessible, as well as to further develop physical, logical, computational and mathematical physics concepts. PlanetPhysics was also a free, collaborative, online physics, mathematical physics, computational physics and physical mathematics project, including original articles, lectures, books and encyclopedia entries. The emphasis was on openness, pedagogy, real-time content, rigour, interlinked content, and also based on a virtual community, or virtual group, of about 600 people with various physics, mathematical physics, physical mathematics, logic (such as quantum logic, relational logic and many-valued logics), axiomatics and mathematics interests. Content The main PlanetPhysics.org focus was on both original research and encyclopedic entries; with over 3,400 physics and mathematics concepts edited in LaTeX and rendered in HTML, the PlanetPhysics Encyclopedia is at present the largest Physics encyclopedia written in LaTeX, containing both introductory as well as advanced level presentations. Moreover, its sections on papers and expositions, are second only to CERN and arXiv physics preprint archives. In addition, the PlanetPhysics.org new websites also included extensive graphics illustrations of physics experiments, and also forum discussions. The emphasis was on modern physics contents, including advanced physics and mathematical physics concepts as well. The project hosted data containing physics, applied physics, engineering and mathematics books, lectures, preprints and research-level papers. A system for both private and semi-private messaging among users was also in place. , the Physics and Mathematical Physics projects hosted over 2,000 entries, containing more than 30,000 concepts in boo
https://en.wikipedia.org/wiki/Julius%20Petersen
Julius Peter Christian Petersen (16 June 1839, Sorø, West Zealand – 5 August 1910, Copenhagen) was a Danish mathematician. His contributions to the field of mathematics led to the birth of graph theory. Biography Petersen's interests in mathematics were manifold, including: geometry, complex analysis, number theory, mathematical physics, mathematical economics, cryptography and graph theory. His famous paper Die Theorie der regulären graphs was a fundamental contribution to modern graph theory as we know it today. In 1898, he presented a counterexample to Tait's claimed theorem about 1-factorability of 3-regular graphs, which is nowadays known as the "Petersen graph". In cryptography and mathematical economics he made contributions which today are seen as pioneering. He published a systematic treatment of geometrical constructions (with straightedge and compass) in 1880. A French translation was reprinted in 1990. A special issue of Discrete Mathematics has been dedicated to the 150th birthday of Petersen. Petersen, as he claimed, had a very independent way of thinking. In order to preserve this independence he made a habit to read as little as possible of other people's mathematics, pushing it to extremes. The consequences for his lack of knowledge of the literature of the time were severe. He spent a significant part of his time rediscovering already known results, in other cases already existing results had to be removed from a submitted paper and in other more serious cases a paper did not get published at all. He started from very modest beginnings, and by hard work, some luck and some good connections, moved steadily upward to a station of considerable importance. In 1891 his work received royal recognition through the award of the Order of the Dannebrog. Among mathematicians he enjoyed an international reputation. At his death –which was front-page news in Copenhagen– the socialist newspaper Social-Demokraten correctly sensed the popular appeal of his st
https://en.wikipedia.org/wiki/The%20California%20Wine%20Club
The California Wine Club is a wine club co-founded in 1990 by Bruce and Pam Boring in Ventura, California and is considered one of the largest mail-order wine businesses in the United States. The club offers hand-selected wines on a monthly basis from boutique and small, family-owned wineries in California, Oregon and Washington State, and other countries such as France, Italy, and Austria. Cork recycling program The California Wine Club partnered with ReCork, a natural wine cork recycling program, in July 2011 to conduct a nationwide cork drive. Uncorked, The California Wine Club newsletter Shipments from The California Wine Club include their newsletter, Uncorked. Uncorked is a full magazine featuring glossy pictures and articles. It also includes tasting notes for featured wines, interviews, recipes, and more. See also Wine clubs Wine tasting Aging of wine Aroma of wine
https://en.wikipedia.org/wiki/Libarchive
libarchive is a free and open-source library for reading and writing various archive and compression formats. It is written in C and works on most Unix-like systems and Windows. History libarchive's development was started in 2003 as part of the FreeBSD project. During the early years it was led by the FreeBSD project, but later it became an independent project. It was first released with FreeBSD 5.3 in November 2004. libarchive libarchive automatically detects and reads archive formats. If the archive is compressed, libarchive also detects and handles compression formats before evaluating the archive. libarchive is designed to minimize the copying of data internally for very optimal performance. Supported archive formats: 7z - read and write ar - read and write cab - read only cpio - read and write ISO9660- read and write lha & lzh - read only pax - read and write rar - read only shar - write only tar - read and write warc (ISO 28500:2009) - read and write xar - read and write zip - read and write Utilities libarchive provides command-line utilities called bsdtar and bsdcpio. These are complete re-implementation based on libarchive. These are the default system tar and cpio on FreeBSD, NetBSD, macOS and Windows. There is also bsdcat, designed to decompress a file to the standard output like zcat. Users libarchive was originally developed for FreeBSD, but is also used in NetBSD and macOS as part of those operating systems. bsdtar has been included in Windows since Windows 10 April 2018 Update. In May 2023, Microsoft announced Windows 11 will natively support additional archive formats such as 7z and RAR via libarchive.
https://en.wikipedia.org/wiki/Tank%20blanketing
Tank blanketing, also called gas sealing or tank padding, is the process of applying a gas to the empty space in a storage container. The term storage container here refers to any container that is used to store products, regardless of its size. Though tank blanketing is used for a variety of reasons, it typically involves using a buffer gas to protect products inside the storage container. A few of the benefits of blanketing include a longer life of the product in the container, reduced hazards, and longer equipment life cycles. Methods In 1970, Appalachian Controls Environmental (ACE) was the world’s first company to introduce a tank blanketing valve. There are now many ready-made systems available for purchase from a variety of process equipment companies. It is also possible to piece together your own system using a variety of different equipment. Regardless of which method is used, the basic requirements are the same. There must be a way of allowing the blanketing gas into the system, and a way to vent the gas should the pressure get too high. Since ACE introduced its valve many companies have engineered their own versions. Though many of the products available vary in features and applicability, the fundamental design is the same. When the pressure inside the container drops below a set point, a valve opens and allows the blanketing gas to enter. Once the pressure reaches the set point, the valve closes. As a safety feature, many systems include a pressure vent that opens when the pressure inside exceeds a maximum pressure set point. This helps to prevent the container from rupturing due to high pressure. Since most blanketing gas sources will provide gas at a much higher than desired pressure, a blanketing system will also use a pressure reducing valve to decrease the inlet pressure to the tank. Although it varies from application to application, blanketing systems usually operate at a slightly higher than atmospheric pressure (a few inches of water colu
https://en.wikipedia.org/wiki/Gelfand%E2%80%93Kirillov%20dimension
In algebra, the Gelfand–Kirillov dimension (or GK dimension) of a right module M over a k-algebra A is: where the supremum is taken over all finite-dimensional subspaces and . An algebra is said to have polynomial growth if its Gelfand–Kirillov dimension is finite. Basic facts The Gelfand–Kirillov dimension of a finitely generated commutative algebra A over a field is the Krull dimension of A (or equivalently the transcendence degree of the field of fractions of A over the base field.) In particular, the GK dimension of the polynomial ring Is n. (Warfield) For any real number r ≥ 2, there exists a finitely generated algebra whose GK dimension is r. In the theory of D-Modules Given a right module M over the Weyl algebra , the Gelfand–Kirillov dimension of M over the Weyl algebra coincides with the dimension of M, which is by definition the degree of the Hilbert polynomial of M. This enables to prove additivity in short exact sequences for the Gelfand–Kirillov dimension and finally to prove Bernstein's inequality, which states that the dimension of M must be at least n. This leads to the definition of holonomic D-modules as those with the minimal dimension n, and these modules play a great role in the geometric Langlands program.
https://en.wikipedia.org/wiki/Joe%20Harris%20%28mathematician%29
Joseph Daniel Harris (born August 17, 1951) is a mathematician at Harvard University working in the field of algebraic geometry. After earning an AB from Harvard College, where he took Math 55, he continued at Harvard to study for a PhD under Phillip Griffiths. Work During the 1980s, he was on the faculty of Brown University, moving to Harvard around 1988. He served as chair of the department at Harvard from 2002 to 2005. His work is characterized by its classical geometric flavor: he has claimed that nothing he thinks about could not have been imagined by the Italian geometers of the late 19th and early 20th centuries, and that if he has had greater success than them, it is because he has access to better tools. Harris is well known for several of his books on algebraic geometry, notable for their informal presentations: Principles of Algebraic Geometry , with Phillip Griffiths Geometry of Algebraic Curves, Vol. 1 , with Enrico Arbarello, Maurizio Cornalba, and Phillip Griffiths , with William Fulton , with David Eisenbud Moduli of Curves , with Ian Morrison. Fat Chance: Probability from 0 to 1, with Benedict Gross and Emily Riehl, 2019 As of 2018, Harris has supervised 50 PhD students, including Brendan Hassett, James McKernan, Rahul Pandharipande, Zvezdelina Stankova, and Ravi Vakil.
https://en.wikipedia.org/wiki/Von%20Neumann%20cardinal%20assignment
The von Neumann cardinal assignment is a cardinal assignment that uses ordinal numbers. For a well-orderable set U, we define its cardinal number to be the smallest ordinal number equinumerous to U, using the von Neumann definition of an ordinal number. More precisely: where ON is the class of ordinals. This ordinal is also called the initial ordinal of the cardinal. That such an ordinal exists and is unique is guaranteed by the fact that U is well-orderable and that the class of ordinals is well-ordered, using the axiom of replacement. With the full axiom of choice, every set is well-orderable, so every set has a cardinal; we order the cardinals using the inherited ordering from the ordinal numbers. This is readily found to coincide with the ordering via ≤c. This is a well-ordering of cardinal numbers. Initial ordinal of a cardinal Each ordinal has an associated cardinal, its cardinality, obtained by simply forgetting the order. Any well-ordered set having that ordinal as its order type has the same cardinality. The smallest ordinal having a given cardinal as its cardinality is called the initial ordinal of that cardinal. Every finite ordinal (natural number) is initial, but most infinite ordinals are not initial. The axiom of choice is equivalent to the statement that every set can be well-ordered, i.e. that every cardinal has an initial ordinal. In this case, it is traditional to identify the cardinal number with its initial ordinal, and we say that the initial ordinal is a cardinal. The -th infinite initial ordinal is written . Its cardinality is written (the -th aleph number). For example, the cardinality of is , which is also the cardinality of , , and (all are countable ordinals). So we identify with , except that the notation is used for writing cardinals, and for writing ordinals. This is important because arithmetic on cardinals is different from arithmetic on ordinals, for example  =  whereas  > . Also, is the smallest uncountable ordi
https://en.wikipedia.org/wiki/Fusobacteriaceae
The Fusobacteriaceae are a family of the bacterial order Fusobacteriales. See also List of bacteria genera List of bacterial orders
https://en.wikipedia.org/wiki/Buchholz%20hydra
In mathematics, especially mathematical logic, graph theory and number theory, the Buchholz hydra game is a type of hydra game, which is a single-player game based on the idea of chopping pieces off of a mathematical tree. The hydra game can be used to generate a rapidly growing function, , which eventually dominates all recursive functions that are provably total in "", and the termination of all hydra games is not provably total in . Rules The game is played on a hydra, a finite, rooted connected tree , with the following properties: The root of has a special label, usually denoted . Any other node of has a label . All nodes directly above the root of have a label . If the player decides to remove the top node of , the hydra will then choose an arbitrary , where is a current turn number, and then transform itself into a new hydra as follows. Let represent the parent of , and let represent the part of the hydra which remains after has been removed. The definition of depends on the label of : If the label of is 0 and is the root of , then = . If the label of is 0 but is not the root of , copies of and all its children are made, and edges between them and 's parent are added. This new tree is . If the label of is for some , then the first node below is labelled as . is then the subtree obtained by starting with and replacing the label of with and with 0. is then obtained by taking and replacing with . In this case, the value of does not matter. If the label of is , is obtained by replacing the label of with . If is the rightmost head of , is written. A series of moves is called a strategy, and a strategy is called a winning strategy if, after a finite amount of moves, the hydra reduces to its root. It has been proven that this always terminates, even though the hydra can get taller by massive amounts. Hydra theorem Buchholz's paper in 1987 showed that the canonical correspondence between a hydra and an infinitary well
https://en.wikipedia.org/wiki/Personalized%20genomics
Personalized genomics is the human genetics-derived study of analyzing and interpreting individualized genetic information by genome sequencing to identify genetic variations compared to the library of known sequences. International genetics communities have spared no effort from the past and have gradually cooperated to prosecute research projects to determine DNA sequences of the human genome using DNA sequencing techniques. The methods that are the most commonly used are whole exome sequencing and whole genome sequencing. Both approaches are used to identify genetic variations. Genome sequencing became more cost-effective over time, and made it applicable in the medical field, allowing scientists to understand which genes are attributed to specific diseases. Personalized medicine is an emerging practice in medicine that develops patient-specific treatments based on an individual's genetic profile. The treatment enables patients to experience maximized therapeutic effectiveness and minimized adverse effects. Personalized medicine has been widely accepted, and future-oriented changes in policy and infrastructure are implemented throughout the world to readily adopt into other fields. History Efforts to explore genes and heredity have been prolonged for over 100 years. From Gregor Mendel’s studies of inheritance, many researchers have dedicated themselves to scientific development via new discoveries and inventions, such as the DNA double helix discovered by Rosalind Franklin and the sanger sequencing invented by Frederick Sanger. Recognition of chromosomal disorders like trisomy in 1959-1960 has led scientists to realize that genes are related to the phenotypic distinctive diseases. It motivated scientists all over to carry out several projects together to understand the human genome. Human Genome Project Human Genome Project (HGP) is a research project conducted by universities and research centers throughout six countries with the primary goal of determining
https://en.wikipedia.org/wiki/Tammes%20problem
In geometry, the Tammes problem is a problem in packing a given number of points on the surface of a sphere such that the minimum distance between points is maximized. It is named after the Dutch botanist Pieter Merkus Lambertus Tammes (the nephew of pioneering botanist Jantina Tammes) who posed the problem in his 1930 doctoral dissertation on the distribution of pores on pollen grains. Mathematicians independent of Tammes began studying circle packing on the sphere in the early 1940s; it was not until twenty years later that the problem became associated with his name. It can be viewed as a particular special case of the generalized Thomson problem of minimizing the total Coulomb force of electrons in a spherical arrangement. Thus far, solutions have been proven only for small numbers of circles: 3 through 14, and 24. There are conjectured solutions for many other cases, including those in higher dimensions. See also Spherical code Kissing number problem Cylinder sphere packings
https://en.wikipedia.org/wiki/PAX6
Paired box protein Pax-6, also known as aniridia type II protein (AN2) or oculorhombin, is a protein that in humans is encoded by the PAX6 gene. Function PAX6 is a member of the Pax gene family which is responsible for carrying the genetic information that will encode the Pax-6 protein. It acts as a "master control" gene for the development of eyes and other sensory organs, certain neural and epidermal tissues as well as other homologous structures, usually derived from ectodermal tissues. However, it has been recognized that a suite of genes is necessary for eye development, and therefore the term of "master control" gene may be inaccurate. Pax-6 is expressed as a transcription factor when neural ectoderm receives a combination of weak Sonic hedgehog (SHH) and strong TGF-Beta signaling gradients. Expression is first seen in the forebrain, hindbrain, head ectoderm and spinal cord followed by later expression in midbrain. This transcription factor is most noted for its use in the interspecifically induced expression of ectopic eyes and is of medical importance because heterozygous mutants produce a wide spectrum of ocular defects such as aniridia in humans. Pax6 serves as a regulator in the coordination and pattern formation required for differentiation and proliferation to successfully take place, ensuring that the processes of neurogenesis and oculogenesis are carried out successfully. As a transcription factor, Pax6 acts at the molecular level in the signaling and formation of the central nervous system. The characteristic paired DNA binding domain of Pax6 utilizes two DNA-binding domains, the paired domain (PD), and the paired-type homeodomain (HD). These domains function separately via utilization by Pax6 to carry out molecular signaling that regulates specific functions of Pax6. An example of this lies in HD's regulatory involvement in the formation of the lens and retina throughout oculogenesis contrasted by the molecular mechanisms of control exhibited
https://en.wikipedia.org/wiki/Upper%20and%20lower%20bounds
In mathematics, particularly in order theory, an upper bound or majorant of a subset of some preordered set is an element of that is greater than or equal to every element of . Dually, a lower bound or minorant of is defined to be an element of that is less than or equal to every element of . A set with an upper (respectively, lower) bound is said to be bounded from above or majorized (respectively bounded from below or minorized) by that bound. The terms bounded above (bounded below) are also used in the mathematical literature for sets that have upper (respectively lower) bounds. Examples For example, is a lower bound for the set (as a subset of the integers or of the real numbers, etc.), and so is . On the other hand, is not a lower bound for since it is not smaller than every element in . The set has as both an upper bound and a lower bound; all other numbers are either an upper bound or a lower bound for that . Every subset of the natural numbers has a lower bound since the natural numbers have a least element (0 or 1, depending on convention). An infinite subset of the natural numbers cannot be bounded from above. An infinite subset of the integers may be bounded from below or bounded from above, but not both. An infinite subset of the rational numbers may or may not be bounded from below, and may or may not be bounded from above. Every finite subset of a non-empty totally ordered set has both upper and lower bounds. Bounds of functions The definitions can be generalized to functions and even to sets of functions. Given a function with domain and a preordered set as codomain, an element of is an upper bound of if for each in . The upper bound is called sharp if equality holds for at least one value of . It indicates that the constraint is optimal, and thus cannot be further reduced without invalidating the inequality. Similarly, a function defined on domain and having the same codomain is an upper bound of , if for each in
https://en.wikipedia.org/wiki/Mir-657%20microRNA%20precursor%20family
In molecular biology mir-657 microRNA is a short RNA molecule. MicroRNAs function to regulate the expression levels of other genes by several mechanisms. See also MicroRNA
https://en.wikipedia.org/wiki/Benbros
Benbros (London) Ltd was a British toy company in existence from the late 1940s to 1965. The company was founded by two brothers, Jack and Nathan Benenson, the name, a contraction of "Benenson Brothers", was adopted in 1951 and the company was based at 145, Gosport Road, Walthamstow. Benbros produced: Die cast metal model cars and farm machinery under the names of "Qualitoys" and "T.V. Series" that came packed in a box that looked like a television set that were later retitled "Mighty Midgets". "Zebra Toys" that came in black and white striped boxes that were better quality models. Hollow cast metal Robin Hood, cowboy and Indian figures and toy soldiers. Timpo Toys sold several of their moulds to Benbros, which can lead to collectors to having difficulty identifying which company made some figures.
https://en.wikipedia.org/wiki/World%20glyph%20set
The world glyph sets are character repertoires comprising a subset of Unicode characters. Their purpose is to provide an implementation guideline for producers of fonts for the representation of natural languages. Unlike Windows Glyph List 4 (WGL) it is specified by font foundries and not by operating system manufacturers. It is, however, very similar in glyph coverage to WGL4, but neither contains all the characters of the other. Digital fonts for the European and American market have traditionally often been sold in a standard ("Std") package for western languages and additional separate files to cover central ("CE"), eastern ("Baltic") and southern ("Turk") European languages in the Roman script and sometimes also packages to support for the Greek (monotonic) and Cyrillic scripts. With the advent of the OpenType format, which supports Unicode, all characters could be included in a single font file. Some font foundries continue to sell packages with differing glyph coverage where the basic one hardly covers more than Windows-1252. They introduced certain variants like Professional ("Pro") which supports all major languages written with Latin letters, Commercial ("Com") for international communication in office use, such as those covering Linotype Extended European Characterset (LEEC) or, adding Greek and Cyrillic, WGL or world glyph sets. The set is used in several font families by Linotype and Monotype such as Neue Frutiger W1G. Character table See also Adobe Glyph List DIN 91379 Unicode subset for Europe
https://en.wikipedia.org/wiki/Variable%20speed%20of%20light
A variable speed of light (VSL) is a feature of a family of hypotheses stating that the speed of light may in some way not be constant, for example, that it varies in space or time, or depending on frequency. Accepted classical theories of physics, and in particular general relativity, predict a constant speed of light in any local frame of reference and in some situations these predict apparent variations of the speed of light depending on frame of reference, but this article does not refer to this as a variable speed of light. Various alternative theories of gravitation and cosmology, many of them non-mainstream, incorporate variations in the local speed of light. Attempts to incorporate a variable speed of light into physics were made by Robert Dicke in 1957, and by several researchers starting from the late 1980s. VSL should not be confused with faster than light theories, its dependence on a medium's refractive index or its measurement in a remote observer's frame of reference in a gravitational potential. In this context, the "speed of light" refers to the limiting speed c of the theory rather than to the velocity of propagation of photons. Historical proposals Background Einstein's equivalence principle, on which general relativity is founded, requires that in any local, freely falling reference frame, the speed of light is always the same. This leaves open the possibility, however, that an inertial observer inferring the apparent speed of light in a distant region might calculate a different value. Spatial variation of the speed of light in a gravitational potential as measured against a distant observer's time reference is implicitly present in general relativity. The apparent speed of light will change in a gravity field and, in particular, go to zero at an event horizon as viewed by a distant observer. In deriving the gravitational redshift due to a spherically-symmetric massive body, a radial speed of light dr/dt can be defined in Schwarzschild coord
https://en.wikipedia.org/wiki/Sorption
Sorption is a physical and chemical process by which one substance becomes attached to another. Specific cases of sorption are treated in the following articles: Absorption "the incorporation of a substance in one state into another of a different state" (e.g., liquids being absorbed by a solid or gases being absorbed by a liquid); Adsorption The physical adherence or bonding of ions and molecules onto the surface of another phase (e.g., reagents adsorbed to a solid catalyst surface); Ion exchange An exchange of ions between two electrolytes or between an electrolyte solution and a complex. The reverse of sorption is desorption. Sorption rate The adsorption and absorption rate of a diluted solute in gas or liquid solution to a surface or interface can be calculated using Fick's laws of diffusion. See also Sorption isotherm
https://en.wikipedia.org/wiki/Fisher%E2%80%93Tippett%E2%80%93Gnedenko%20theorem
In statistics, the Fisher–Tippett–Gnedenko theorem (also the Fisher–Tippett theorem or the extreme value theorem) is a general result in extreme value theory regarding asymptotic distribution of extreme order statistics. The maximum of a sample of iid random variables after proper renormalization can only converge in distribution to one of 3 possible distributions: the Gumbel distribution, the Fréchet distribution, or the Weibull distribution. Credit for the extreme value theorem and its convergence details are given to Fréchet (1927), Fisher and Tippett (1928), Mises (1936) and Gnedenko (1943). The role of the extremal types theorem for maxima is similar to that of central limit theorem for averages, except that the central limit theorem applies to the average of a sample from any distribution with finite variance, while the Fisher–Tippet–Gnedenko theorem only states that if the distribution of a normalized maximum converges, then the limit has to be one of a particular class of distributions. It does not state that the distribution of the normalized maximum does converge. Statement Let be a sequence of independent and identically-distributed random variables with cumulative distribution function . Suppose that there exist two sequences of real numbers and such that the following limits converge to a non-degenerate distribution function: , or equivalently: . In such circumstances, the limit distribution belongs to either the Gumbel, the Fréchet or the Weibull family. In other words, if the limit above converges, then up to a linear change of coordinates will assume the form: or else for some parameter This is the cumulative distribution function of the generalized extreme value distribution (GEV) with extreme value index . The GEV distribution groups the Gumbel, Fréchet and Weibull distributions into a single one. Note that the second formula (the Gumbel distribution) is the limit of the first as goes to zero. Conditions of convergence The Fisher–
https://en.wikipedia.org/wiki/Eyferth%20study
The Eyferth study, conducted by German psychologist Klaus Eyferth, examined the IQs of white and racially-mixed children raised by single mothers in post-World War II West Germany. The mothers of the children studied were white German women, while their fathers were white and black members of the US occupation forces. In contrast to results obtained in many American studies, the average IQs of the children studied were roughly similar across racial groups, making the study an oft-cited piece of evidence in the debate about race and intelligence. Eyferth's study was first published under the title Eine Untersuchung der Neger-Mischlingskinder in Westdeutschland in the journal Vita Humana in 1959. Eyferth described further results in the paper Leistungen verschiedener Gruppen von Besatzungskindern in Hamburg-Wechsler Intelligenztest für Kinder (HAWIK), which was published in the journal Archiv für die gesamte Psychologie in 1961. Study design The children studied had been raised by their unmarried German mothers. Most of the fathers, white or black, had been members of the US occupation forces stationed in Germany. At the time of the study, the children were aged between 5 and 13 (mean age: 10). The mothers of the children were approximately matched for socio-economic status; they were mostly of low socioeconomic status. There were about 98 mixed race (black-white), and about 83 white children in the sample. The total sample consisted of about 5 percent of the German children known to have been fathered by black soldiers between 1945 and 1953, in addition to a matched sample of 83 German children whose fathers were white soldiers. Of the fathers of the mixed-race children, about 20 percent were French Africans and the remaining approximately 80 percent were African Americans. For assessing IQ, a German version of the WISC intelligence test (Hamburg Wechsler Intelligence Scale for Children, HAWIK) was used. Results The white children studied averaged an IQ of 97
https://en.wikipedia.org/wiki/Logic%20learning%20machine
Logic learning machine (LLM) is a machine learning method based on the generation of intelligible rules. LLM is an efficient implementation of the Switching Neural Network (SNN) paradigm, developed by Marco Muselli, Senior Researcher at the Italian National Research Council CNR-IEIIT in Genoa. LLM has been employed in many different sectors, including the field of medicine (orthopedic patient classification, DNA micro-array analysis and Clinical Decision Support Systems ), financial services and supply chain management. History The Switching Neural Network approach was developed in the 1990s to overcome the drawbacks of the most commonly used machine learning methods. In particular, black box methods, such as multilayer perceptron and support vector machine, had good accuracy but could not provide deep insight into the studied phenomenon. On the other hand, decision trees were able to describe the phenomenon but often lacked accuracy. Switching Neural Networks made use of Boolean algebra to build sets of intelligible rules able to obtain very good performance. In 2014, an efficient version of Switching Neural Network was developed and implemented in the Rulex suite with the name Logic Learning Machine. Also, an LLM version devoted to regression problems was developed. General Like other machine learning methods, LLM uses data to build a model able to perform a good forecast about future behaviors. LLM starts from a table including a target variable (output) and some inputs and generates a set of rules that return the output value corresponding to a given configuration of inputs. A rule is written in the form: where consequence contains the output value whereas premise includes one or more conditions on the inputs. According to the input type, conditions can have different forms: for categorical variables the input value must be in a given subset :. for ordered variables the condition is written as an inequality or an interval: or A possible rule is ther
https://en.wikipedia.org/wiki/White%20Rabbit%20Project
White Rabbit is the name of a collaborative project including CERN, GSI Helmholtz Centre for Heavy Ion Research and other partners from universities and industry to develop a fully deterministic Ethernet-based network for general purpose data transfer and sub-nanosecond accuracy time transfer. Its initial use was as a timing distribution network for control and data acquisition timing of the accelerator sites at CERN as well as in GSI's Facility for Antiproton and Ion Research (FAIR) project. The hardware designs as well as the source code are publicly available. The name of the project is a reference to the White Rabbit appearing in Lewis Carroll's novel Alice's Adventures in Wonderland. Focus and goals White Rabbit provides sub-nanosecond synchronization accuracy, which formerly required dedicated hard-wired timing systems, with the flexibility and modularity of real-time Ethernet networks. A White Rabbit network may be used solely to provide timing and synchronization to a distributed electronic system, or be used to provide both timing and real-time data transfer. The White Rabbit Project focuses on: Sub-nanosecond accuracy: synchronization of more than 1000 nodes via fiber or copper connections of up to 10 km of length. Flexibility: creates a scalable and modular platform with simple configuration and low maintenance requirements. Predictability and Reliability: allows the deterministic delivery of highest priority messages by using Class of service. Robustness: no losses of high priority system device control messages. Open source hardware and software: to avoid vendor lock-in. Another characteristic of this project is that it operates completely on open source with both the hardware and software sources available. Technologies To achieve sub-nanosecond synchronization White Rabbit utilizes Synchronous Ethernet (SyncE) to achieve syntonization and IEEE 1588 (1588) Precision Time Protocol (PTP) to communicate time and a module for precise phase diffe
https://en.wikipedia.org/wiki/Streaking%20%28microbiology%29
In microbiology, streaking is a technique used to isolate a pure strain from a single species of microorganism, often bacteria. Samples can then be taken from the resulting colonies and a microbiological culture can be grown on a new plate so that the organism can be identified, studied, or tested. The modern streak plate method has progressed from the efforts of Robert Koch and other microbiologists to obtain microbiological cultures of bacteria in order to study them. The dilution or isolation by streaking method was first developed by Loeffler and Gaffky in Koch's laboratory, which involves the dilution of bacteria by systematically streaking them over the exterior of the agar in a Petri dish to obtain isolated colonies which will then grow into quantity of cells, or isolated colonies. If the agar surface grows microorganisms which are all genetically same, the culture is then considered as a microbiological culture. Technique Streaking is rapid and ideally a simple process of isolation dilution. The technique is done by diluting a comparatively large concentration of bacteria to a smaller concentration. The decrease of bacteria should show that colonies are sufficiently spread apart to affect the separation of the different types of microbes. Streaking is done using a sterile tool, such as a cotton swab or commonly an inoculation loop. Aseptic techniques are used to maintain microbiological cultures and to prevent contamination of the growth medium. There are many different types of methods used to streak a plate. Picking a technique is a matter of individual preference and can also depend on how large the number of microbes the sample contains. The three-phase streaking pattern, known as the T-Streak, is recommended for beginners. The streaking is done using a sterile tool, such as a cotton swab or commonly an inoculation loop. The inoculation loop is first sterilized by passing it through a flame. When the loop is cool, it is dipped into an inoculum such as
https://en.wikipedia.org/wiki/Delaunay%20tessellation%20field%20estimator
The Delaunay tessellation field estimator (DTFE), (or Delone tessellation field estimator (DTFE)) is a mathematical tool for reconstructing a volume-covering and continuous density or intensity field from a discrete point set. The DTFE has various astrophysical applications, such as the analysis of numerical simulations of cosmic structure formation, the mapping of the large-scale structure of the universe and improving computer simulation programs of cosmic structure formation. It has been developed by Willem Schaap and Rien van de Weijgaert. The main advantage of the DTFE is that it automatically adapts to (strong) variations in density and geometry. It is therefore very well suited for studies of the large scale galaxy distribution. Method The DTFE consists of three main steps: Step 1 The starting point is a given discrete point distribution. In the upper left-hand frame of the figure, a point distribution is plotted in which at the center of the frame an object is located whose density diminishes radially outwards. In the first step of the DTFE, the Delaunay tessellation of the point distribution is constructed. This is a volume-covering division of space into triangles (tetrahedra in three dimensions), whose vertices are formed by the point distribution (see figure, upper right-hand frame). The Delaunay tessellation is defined such that inside the interior of the circumcircle of each Delaunay triangle no other points from the defining point distribution are present. Step 2 The Delaunay tessellation forms the heart of the DTFE. In the figure it is clearly visible that the tessellation automatically adapts to both the local density and geometry of the point distribution: where the density is high, the triangles are small and vice versa. The size of the triangles is therefore a measure of the local density of the point distribution. This property of the Delaunay tessellation is exploited in step 2 of the DTFE, in which the local density is estimated at
https://en.wikipedia.org/wiki/LDAP%20Data%20Interchange%20Format
The LDAP Data Interchange Format (LDIF) is a standard plain text data interchange format for representing Lightweight Directory Access Protocol (LDAP) directory content and update requests. LDIF conveys directory content as a set of records, one record for each object (or entry). It also represents update requests, such as Add, Modify, Delete, and Rename, as a set of records, one record for each update request. LDIF was designed in the early 1990s by Tim Howes, Mark C. Smith, and Gordon Good while at the University of Michigan. LDIF was updated and extended in the late 1990s for use with Version 3 of LDAP. This later version of LDIF is called version 1 and is formally specified in RFC 2849, an IETF Standard Track RFC. RFC 2849 is authored by Gordon Good and was published in June 2000. It is currently a Proposed Standard. A number of extensions to LDIF have been proposed over the years. One extension has been formally specified by the IETF and published. RFC 4525, authored by Kurt Zeilenga, extended LDIF to support the LDAP Modify-Increment extension. It is expected that additional extensions will be published by the IETF in the future. Content record format Each content record is represented as a group of attributes, with records separated from one another by blank lines. The individual attributes of a record are represented as single logical lines (represented as one or more multiple physical lines via a line-folding mechanism), comprising "name: value" pairs. Value data that do not fit within a portable subset of ASCII characters are marked with '::' after the attribute name and encoded into ASCII using base64 encoding. The content record format is a subset of the Internet Directory Information type.RFC 2425 Tools that employ LDIF The OpenLDAP utilities include tools for exporting data from LDAP servers to LDIF content records (), importing data from LDIF content records to LDAP servers (), and applying LDIF change records to LDAP servers (). LDI
https://en.wikipedia.org/wiki/Duress%20code
A duress code is a covert distress signal used by an individual who is being coerced by one or more hostile persons. It is used to warn others that they are being forced to do something against their will. Typically, the warning is given via some innocuous signal embedded in normal communication, such as a code-word or phrase spoken during conversation to alert other personnel. Alternatively, the signal may be incorporated into the authentication process itself, typically in the form of a panic password, distress password, or duress PIN that is distinct from the user's normal password or PIN. These concepts are related to a panic alarm and often achieve the same outcome. Civilian usage Some home and property alarm systems have duress PINs, where the last two digits of the reset code are switched around. Entering the code when under duress from an assailant can trigger a silent alarm, alerting police or security personnel in a covert manner. The implementation of this feature has not been without controversy, as it has been claimed to lead to false alarms. A similar mechanism, SafetyPIN, has been proposed for use in ATMs. In 2010, the Federal Trade Commission issued a report studying the viability of such mechanisms for ATMs. They noted duress PINs have never been actually implemented in any ATM, and conclude that the costs of deployment outweighs the likelihood they will actually deter criminal activity. When a duress PIN is used to trigger a silent alarm, an adversary can always request the PIN in advance and ensure the appropriately modified PIN is entered instead. If the adversary does not know which PIN is correct, they may choose randomly between the two possible codes allowing them to succeed half of the time. In scenarios where a panic password is used to limit access control, instead of triggering an alarm, it is insufficient to have a single panic password. If the adversary knows the system, a common assumption, then they will simply force the user to a
https://en.wikipedia.org/wiki/Integrated%20stress%20response
The integrated stress response is a cellular stress response conserved in eukaryotic cells that downregulates protein synthesis and upregulates specific genes in response to internal or environmental stresses. Background The integrated stress response can be triggered within a cell due to either extrinsic or intrinsic conditions. Extrinsic factors include hypoxia, amino acid deprivation, glucose deprivation, viral infection and presence of oxidants. The main intrinsic factor is endoplasmic reticulum stress due to the accumulation of unfolded proteins. It has also been observed that the integrated stress response may trigger due to oncogene activation. The integrated stress response will either cause the expression of genes that fix the damage in the cell due to the stressful conditions, or it will cause a cascade of events leading to apoptosis, which occurs when the cell cannot be brought back into homeostasis. eIF2 protein complex Stress signals can cause protein kinases, known as EIF-2 kinases, to phosphorylate the α subunit of a protein complex called translation initiation factor 2 (eIF2), resulting in the gene ATF4 being turned on, which will further affect gene expression. eIF2 consists of three subunits: eIF2α, eIF2β and eIF2γ. eIF2α contains two binding sites, one for phosphorylation and one for RNA binding. The kinases work to phosphorylate serine 51 on the α subunit, which is a reversible action. In a cell experiencing normal conditions, eIF2 aids in the initiation of mRNA translation and recognizing the AUG start codon. However, once eIF2α is phosphorylated, the complex’s activity reduces, causing reduction in translation initiation and protein synthesis, while promoting expression of the ATF4 gene. Protein kinases There are four known mammalian protein kinases that phosphorylate eIF2α, including PKR-like ER kinase (PERK, EIF2AK3), heme-regulated eIF2α kinase (HRI, EIF2AK1), general control non-depressible 2 (GCN2, EIF2AK4) and double stranded RNA
https://en.wikipedia.org/wiki/Mayo-Smith%20pyramid
A Mayo-Smith pyramid is a triangle divided into a sequence of isosceles trapezoids configured such that the outer perimeter maintains the shape of a triangle with each additional element. A Mayo-Smith pyramid is used to describe system development methodologies adapted for scenarios characterized by schedule and resource uncertainty. "Two Ways to Build a Pyramid" was published in 2001. In this, the Mayo-Smith pyramid sequence (see Figure B) is used to illustrate a specific case study, and contrasted with a less favorable sequence (see Figure A). While Mayo-Smith's pyramid is typically depicted as a two dimensional sequence, it may also be depicted in three dimensions.
https://en.wikipedia.org/wiki/Phosphoribosylaminoimidazolesuccinocarboxamide
Phosphoribosylaminoimidazolesuccinocarboxamide (SAICAR) is an intermediate in the formation of purines. The conversion of ATP, L-aspartate, and 5-aminoimidazole-4-carboxyribonucleotide (CAIR) to 5-aminoimidazole-4-(N-succinylcarboxamide) ribonucleotide, ADP, and phosphate by phosphoribosylaminoimidazolesuccinocarboxamide synthetase (SAICAR synthetase) represents the eighth step of de novo purine nucleotide biosynthesis.
https://en.wikipedia.org/wiki/RCA%20Spectra%2070
The RCA Spectra 70 was a line of electronic data processing (EDP) equipment manufactured by the Radio Corporation of America’s computer division beginning in April 1965. The Spectra 70 line included several CPU models, various configurations of core memory, mass-storage devices, terminal equipment, and a variety of specialized interface equipment. The system architecture and instruction-set were largely compatible with the non-privileged instruction-set of the IBM System/360, including use of the EBCDIC character set. While this degree of compatibility made some interchange of programs and data possible, differences in the operating system software precluded transparent movement of programs between the two systems. Competition in the mainframe market was fierce, and in 1971 the company sold the computer division and the Spectra 70 line to Sperry Rand, taking a huge write down in the process. System overview Five models of the Spectra 70 CPU were announced around 1965, ranging from a small system (70/15) to the large scale (70/55). Some of the main features were: The systems were upward-compatible, allowing programs written for a smaller model to run on any larger machine in the series. Larger machines in the series were also faster, with memory access times ranging from two microseconds in the 70/15 to 0.84 microseconds in the 70/55. Memory capacities ranged from a minimum of 4,096 bytes (4 KB) in the 70/15 to a maximum of 524,288 bytes (512 KB) in the 70/55. All used the Extended Binary Coded Decimal Interchange Code (EBCDIC) of eight bits plus parity for internal data representation. The use of a standard electrical interface allowed the same peripherals to be used with any CPU model in the series. Simultaneous input and output was accomplished by the use of intelligent communication channels. Like the IBM 360, two types of channel were available (on all but the 70/15): selector channels which could address up to 256 devices (one at a time), a
https://en.wikipedia.org/wiki/Iterator
In computer programming, an iterator is an object that enables a programmer to traverse a container, particularly lists. Various types of iterators are often provided via a container's interface. Though the interface and semantics of a given iterator are fixed, iterators are often implemented in terms of the structures underlying a container implementation and are often tightly coupled to the container to enable the operational semantics of the iterator. An iterator performs traversal and also gives access to data elements in a container, but does not itself perform iteration (i.e., not without some significant liberty taken with that concept or with trivial use of the terminology). An iterator is behaviorally similar to a database cursor. Iterators date to the CLU programming language in 1974. Description Internal Iterators Internal iterators are higher order functions (often taking anonymous functions, but not necessarily) such as map(), reduce() etc., implementing the traversal across a container, applying the given function to every element in turn. An example might be Python's map function: digits = [0, 1, 2, 3, 4, 5, 6, 7, 8, 9] squared_digits = map(lambda x: x**2, digits) # Iterating over this iterator would result in 0, 1, 4, 9, 16, ..., 81. External iterators and the iterator pattern An external iterator may be thought of as a type of pointer that has two primary operations: referencing one particular element in the object collection (called element access), and modifying itself so it points to the next element (called element traversal). There must also be a way to create an iterator so it points to some first element as well as some way to determine when the iterator has exhausted all of the elements in the container. Depending on the language and intended use, iterators may also provide additional operations or exhibit different behaviors. The primary purpose of an iterator is to allow a user to process every element of a container while isola
https://en.wikipedia.org/wiki/Nanoscale%20secondary%20ion%20mass%20spectrometry
NanoSIMS (nanoscale secondary ion mass spectrometry) is an analytical instrument manufactured by CAMECA which operates on the principle of secondary ion mass spectrometry. The NanoSIMS is used to acquire nanoscale resolution measurements of the elemental and isotopic composition of a sample. The NanoSIMS is able to create nanoscale maps of elemental or isotopic distribution, parallel acquisition of up to seven masses, isotopic identification, high mass resolution, subparts-per-million sensitivity with spatial resolution down to 50 nm. The original design of the NanoSIMS instrument was conceived by Georges Slodzian at the University of Paris Sud in France and at the Office National d'Etudes et de Recherches Aérospatiales. There are currently around 50 NanoSIMS instruments worldwide. How it works The NanoSIMS uses an ion source to produce a primary beam of ions. These primary ions erode the sample surface and produce atomic collisions, some of these collisions result in the release of secondary ion particles. These ions are transmitted through a mass spectrometer, where the masses are measured and identified. The primary ion beam is rastered across the sample surface and a ‘map’ of the element and isotope distribution is created by counting the number of ions that originated from each pixel with at best a 50 nanometer (nm) resolution, 10-50 times greater than conventional SIMS. This is achieved by positioning the primary probe in close proximity to the sample using a coaxial lens assembly. The primary ion beam impacts the sample surface at 90°, with the secondary ions extracted back through the same lens assembly. This allows for the isotopic composition of individual cells to be distinguished at parts per million (ppm) or parts per billion (ppb) range. The main drawback of this set up is that the primary and secondary ion beams must be of opposite polarity which can limit which elements can be detected simultaneously. NanoSIMS can detect minute mass differences
https://en.wikipedia.org/wiki/Wendel%27s%20theorem
In geometric probability theory, Wendel's theorem, named after James G. Wendel, gives the probability that N points distributed uniformly at random on an -dimensional hypersphere all lie on the same "half" of the hypersphere. In other words, one seeks the probability that there is some half-space with the origin on its boundary that contains all N points. Wendel's theorem says that the probability is The statement is equivalent to being the probability that the origin is not contained in the convex hull of the N points and holds for any probability distribution on that is symmetric around the origin. In particular this includes all distribution which are rotationally invariant around the origin. This is essentially a probabilistic restatement of Schläfli's theorem that hyperplanes in general position in divides it into regions.
https://en.wikipedia.org/wiki/List%20of%20Egyptian%20flags
This is a list of flags used by and in Egypt. For more information about the national flag, visit the article Flag of Egypt. National flags Standards Presidential standards Throughout the republican era, the standard of the president of Egypt has been identical to the national flag, with the addition of the coat of arms (eagle or hawk) in the upper-left corner. Even though the Constitution of Egypt states that the President is the Supreme Commander of the Armed Forces (article 150), the flag of the Supreme Commander differs from the Presidential Standard. It is identical to the national flag, with the addition in the upper-left corner of a white Eagle of Saladin contained between two crossed swords. The Supreme Commander also has his own naval ensign, air force flag and air defence flag. Military ordinances state that the flags of the Supreme Commander must be hoisted during the President's visits to each military unit. The national flag must be hoisted in the middle, with the Presidential Standard on its right side, and the flag of the Supreme Commander on its left side. Royal standards Military Army Navy Air Force Air Defense Forces Government Department Corporations Governorates Political flags Historical flags Other Ancient Standards Present in Art Proposed flags
https://en.wikipedia.org/wiki/Cambridge%20University%20Wine%20Society
The Cambridge University Wine Society (CUWS; formerly known as the Cambridge University Wine and Food Society) is a traditional membership club, one of the Registered Clubs & Societies of University of Cambridge, England, founded in 1792. The society aims to introduce university students to the world of wine through regular tastings during the academic term, and other activities and events. Membership is open to the public. A number of past members have entered the international wine trade. Prominent wine writer Hugh Johnson joined as an undergraduate at King's College, Cambridge in the 1950s, with his roommate and fellow member Adrian Cowell. Members participate in the annual Varsity blind wine tasting match, sponsored by champagne house Pol Roger. and other blind wine tasting competitions. The society and competitions were described in detail in the 2013 book "Reds, Whites & Varsity Blues: 60 years of the Oxford & Cambridge Blind Wine-Tasting Competition". In January 2013 the Society marked the 350th anniversary of the day diarist Samuel Pepys "drank a sort of French wine called Ho Bryan, that hath a good and most particular taste that I never met with." Pepys graduated from Magdalene College in 1654. CUWS alumni Notable alumni include: Hugh Johnson, wine writer David Peppercorn, wine writer, president of the society (cellarer) in 1952 Robin Don, Master of Wine Julian Jeffs John H. Plumb, historian Denis Mack Smith, historian Tom King, Baron King of Bridgwater James Cropper, businessman
https://en.wikipedia.org/wiki/Herbrand%20quotient
In mathematics, the Herbrand quotient is a quotient of orders of cohomology groups of a cyclic group. It was invented by Jacques Herbrand. It has an important application in class field theory. Definition If G is a finite cyclic group acting on a G-module A, then the cohomology groups Hn(G,A) have period 2 for n≥1; in other words Hn(G,A) = Hn+2(G,A), an isomorphism induced by cup product with a generator of H2(G,Z). (If instead we use the Tate cohomology groups then the periodicity extends down to n=0.) A Herbrand module is an A for which the cohomology groups are finite. In this case, the Herbrand quotient h(G,A) is defined to be the quotient h(G,A) = |H2(G,A)|/|H1(G,A)| of the order of the even and odd cohomology groups. Alternative definition The quotient may be defined for a pair of endomorphisms of an Abelian group, f and g, which satisfy the condition fg = gf = 0. Their Herbrand quotient q(f,g) is defined as if the two indices are finite. If G is a cyclic group with generator γ acting on an Abelian group A, then we recover the previous definition by taking f = 1 - γ and g = 1 + γ + γ2 + ... . Properties The Herbrand quotient is multiplicative on short exact sequences. In other words, if 0 → A → B → C → 0 is exact, and any two of the quotients are defined, then so is the third and h(G,B) = h(G,A)h(G,C) If A is finite then h(G,A) = 1. For A is a submodule of the G-module B of finite index, if either quotient is defined then so is the other and they are equal: more generally, if there is a G-morphism A → B with finite kernel and cokernel then the same holds. If Z is the integers with G acting trivially, then h(G,Z) = |G| If A is a finitely generated G-module, then the Herbrand quotient h(A) depends only on the complex G-module C⊗A (and so can be read off from the character of this complex representation of G). These properties mean that the Herbrand quotient is usually relatively easy to calculate, and is often much easier to calculate than the orders
https://en.wikipedia.org/wiki/Unimolecular%20ion%20decomposition
Unimolecular ion decomposition is the fragmentation of a gas phase ion in a reaction with a molecularity of one. Ions with sufficient internal energy may fragment in a mass spectrometer, which in some cases may degrade the mass spectrometer performance, but in other cases, such as tandem mass spectrometry, the fragmentation can reveal information about the structure of the ion. Wahrhaftig diagram A Wahrhaftig diagram (named after Austin L. Wahrhaftig) illustrates the relative contributions in unimolecular ion decomposition of direct fragmentation and fragmentation following rearrangement. The x-axis of the diagram represents the internal energy of the ion. The lower part of the diagram shows the logarithm of the rate constant k for unimolecular dissociation whereas the upper portion of the diagram indicates the probability of forming a particular product ion. The green trace in the lower part of the diagram indicates the rate of the rearrangement reaction given by ABCD+ -> {AD+} + BC and the blue trace indicates the direct cleavage reaction ABCD+ -> {AB+} + CD A rate constant of 106 s−1 is sufficiently fast for ion decomposition within the ion source of a typical mass spectrometer. Ions with rate constants less than 106 s−1 and greater than approximately 105 s−1 (lifetimes between 10−5 and 10−6 s) have a high probability of decomposing in the mass spectrometer between the ion source and the detector. These rate constants are indicated in the Wahrhaftig diagram by the log k = 5 and log k = 6 dashed lines. Indicated on the rate constant plot are the reaction critical energy (also called the activation energy) for the formation of AD+, E0(AD+) and AB+, E0(AB+). These represent the minimum internal energy of ABCD+ required to form the respective product ions: the difference in the zero point energy of ABCD+ and that of the activated complex. When the internal energy of ABCD+ is greater than Em(AD+), the ions are metastable (indicated by m*); this occurs near lo
https://en.wikipedia.org/wiki/Terry%20Bogard
is a fictional character created by SNK as the protagonist of their Fatal Fury series. Introduced in Fatal Fury: King of Fighters in 1991, he is an American fighter who enters the worldwide "The King of Fighters" tournaments to combat his father's murderer, Geese Howard. Following Geese's defeat, Terry becomes the guardian of Geese's son Rock Howard. He is also a regular character in the crossover video game series The King of Fighters, where he continues participating in tournaments. He has also been a guest character in other games, including the Capcom vs. SNK series, Arika’s Fighting EX Layer and Nintendo's crossover fighting game Super Smash Bros. Ultimate. In addition to video games, Terry has appeared in anime films based on Fatal Fury, and manga serialized in Comic Bom Bom. SNK created Terry with the idea of a macho fighter which is reflected in his design. After several games, Terry was redesigned for Garou: Mark of the Wolves, where he is a more peaceful person. This had an impact on his design which was also altered for The King of Fighters games. Terry's popularity led SNK to create a female version of him for the 2018 game SNK Heroines: Tag Team Frenzy. The character has been well received, and is ranked highly in popularity polls. Publications have praised the character's personality and movesets and how he was developed from Fatal Fury to The King of Fighters, although the female Terry saw mixed reception. Terry has become an icon for not just SNK, but fighting games in general, appearing prominently on SNK and non-SNK merchandise. Creation and design Origin Terry Bogard originates in the development of Capcom's fighting game Street Fighter (1987), for which developers Hiroshi Matsumoto and Takashi Nishiyama created a concept for a Caucasian man wearing a leather jacket; the concept was not used in Street Fighter, as they decided to make the main playable character a dougi-clad karateka name Ryu instead. They still wanted to use this concept in a
https://en.wikipedia.org/wiki/Matter%20power%20spectrum
The matter power spectrum describes the density contrast of the universe (the difference between the local density and the mean density) as a function of scale. It is the Fourier transform of the matter correlation function. On large scales, gravity competes with cosmic expansion, and structures grow according to linear theory. In this regime, the density contrast field is Gaussian, Fourier modes evolve independently, and the power spectrum is sufficient to completely describe the density field. On small scales, gravitational collapse is non-linear, and can only be computed accurately using N-body simulations. Higher-order statistics are necessary to describe the full field at small scales. Definition Let represent the matter overdensity, a dimensionless quantity defined as: where is the average matter density over all space. The power spectrum is most commonly understood as the Fourier transform of the autocorrelation function, , mathematically defined as: for . This then determines the easily derived relationship to the power spectrum, , that is Equivalently, letting denote the Fourier transform of the overdensity , the power spectrum is given by the following average over Fourier space: (note that is not an overdensity but the Dirac delta function). Since has dimensions of (length)3, the power spectrum is also sometimes given in terms of the dimensionless function: Development according to gravitational expansion If the autocorrelation function describes the probability of a galaxy at a distance from another galaxy, the matter power spectrum decomposes this probability into characteristic lengths, , and its amplitude describes the degree to which each characteristic length contributes to the total over-probability. The overall shape of the matter power spectrum is best understood in terms of the linear perturbation theory analysis of the growth of structure, which predicts to first order that the power spectrum grows according to: Wher
https://en.wikipedia.org/wiki/Wirth%20Research
Wirth Research is a group of engineering companies, founded by Nicholas Wirth in 2003, specialising in research, development, design and manufacture for the motor racing industry and other high technology sectors. The companies use virtual engineering technologies to enable a completely simulated vehicle design, development and testing process. The group is known for deviating from traditional physical development models; most notably neglecting to use a wind tunnel and instead relying solely upon computational fluid dynamics to design the 2010 Virgin Racing VR-01 Formula 1 car. The Wirth Research group have a long-standing partnership with Honda Performance Development Inc (HPD) which is responsible for the design, development and manufacture of the ARX sports cars. Wirth Research also provides client IndyCar teams with full technical support. Design Wirth Research encompasses design disciplines that include: Composite structures Suspension Transmission Electronics and Control Hydraulics Cooling, lubrication and engine ancillaries R&D, component systems and whole-vehicle testing Wirth Research uses tools provided by Siemens UGS PLM for many of its design activities. Computational Fluid Dynamics The company uses ANSYS Fluent solvers for its analysis. Aerodynamic design The largest department at Wirth Research, whose Computational Fluid Dynamics contents are typically evaluated by WR Digital. Whilst the company is not known for using physical models in the development process, it does use real-world wind-tunnel testing to verify its CFD results. Vehicle simulation Two identical simulators are split between HPD's headquarters in Indianapolis and the company's base in Bicester. They are built in partnership with Cruden on a modified 6 degree of freedom motion platform constructed of carbon fibre and presently housing a DW12 Monocoque. The display mode consists of numerous projectors providing a passive stereoscopic 3D 170 degree display of a modified
https://en.wikipedia.org/wiki/Sublunary%20sphere
In Aristotelian physics and Greek astronomy, the sublunary sphere is the region of the geocentric cosmos below the Moon, consisting of the four classical elements: earth, water, air, and fire. The sublunary sphere was the realm of changing nature. Beginning with the Moon, up to the limits of the universe, everything (to classical astronomy) was permanent, regular and unchanging—the region of aether where the planets and stars are located. Only in the sublunary sphere did the powers of physics hold sway. Evolution of concept Plato and Aristotle helped to formulate the original theory of a sublunary sphere in antiquity, the idea usually going hand in hand with geocentrism and the concept of a spherical Earth. Avicenna carried forward into the Middle Ages the Aristotelian idea of generation and corruption being limited to the sublunary sphere. Medieval scholastics like Thomas Aquinas, who charted the division between celestial and sublunary spheres in his work Summa Theologica, also drew on Cicero and Lucan for an awareness of the great frontier between Nature and Sky, sublunary and aetheric spheres. The result for medieval/Renaissance mentalities was a pervasive awareness of the existence, at the Moon, of what C.S. Lewis called 'this "great divide"...from aether to air, from 'heaven' to 'nature', from the realm of gods (or angels) to that of daemons, from the realm of necessity to that of contingence, from the incorruptible to the corruptible" However, the theories of Copernicus began to challenge the sublunary/aether distinction. In their wake, Tycho Brahe's observations of a new star (nova) and of comets in the supposedly unchanging heavens further undermined the Aristotelian view.<ref>R. Curley, Scientists and Inventors of the Renaissance (2012) p. 6-8</ref> Thomas Kuhn saw scientists' new ability to see change in the 'incorruptible' heavens as a classic example of the new possibilities opened up by a paradigm shift. Literary offshoots Dante envisaged Mt Purga
https://en.wikipedia.org/wiki/Helix%20Board%2012%20FireWire
The Helix Board 12 FireWire is a mixer developed by Phonic Corporation that features a FireWire Interface, able to connect the mixer to Windows and Mac computers. The product was released in 2005 and has since become a signature product for the company Outputs The Helix Board 12 FireWire has a stereo main output, control room output, auxiliary send, headphones outputs, RCA record outputs and alternate 3-4 outputs. Digital Effects A built-in 16 program, 32-bit effect processor is included with the Helix Board 12 FireWire. Effects include Hall, Room, Plate, Cathedral, and more. FireWire Interface The FireWire interface allows all 8 inputs (counting stereo channels as 2) of the Helix Board 12 FireWire to be sent to a computer for recording, through its ASIO drivers (no drivers are required for the Mac). The main stereo output of your computer can also be sent through the FireWire interface, back to the Helix Board. The returned signal can be routed by the touch of one of the buttons on the face of the mixer. The Helix Board 12 FireWire is bundled with Steinberg Cubase LE software for recording purposes, though it's suggested that users upgrade to SX or use other digital audio workstation software. System requirements Windows Microsoft Windows XP SP1 or SP2 Available FireWire port Intel Pentium 4 processor or equivalent AMD processor Motherboard with Intel or VIA chipset 5400RPM or faster hard disc drive (7200 RPM or faster with 8MB cache recommended) 256MB or more of RAM (512MB recommended) Macintosh Mac OS X 10.3.5 or later with native FireWire support G4 or newer processor 256MB or more of RAM MKII 2006 saw the release of the Helix Board 12 FireWire MKII, which included additional features. Among these was a new DFX processor, pre/post switch on each input channel and a channel 9/10 assign switch.
https://en.wikipedia.org/wiki/Computer%20simulation
Computer simulation is the process of mathematical modelling, performed on a computer, which is designed to predict the behaviour of, or the outcome of, a real-world or physical system. The reliability of some mathematical models can be determined by comparing their results to the real-world outcomes they aim to predict. Computer simulations have become a useful tool for the mathematical modeling of many natural systems in physics (computational physics), astrophysics, climatology, chemistry, biology and manufacturing, as well as human systems in economics, psychology, social science, health care and engineering. Simulation of a system is represented as the running of the system's model. It can be used to explore and gain new insights into new technology and to estimate the performance of systems too complex for analytical solutions. Computer simulations are realized by running computer programs that can be either small, running almost instantly on small devices, or large-scale programs that run for hours or days on network-based groups of computers. The scale of events being simulated by computer simulations has far exceeded anything possible (or perhaps even imaginable) using traditional paper-and-pencil mathematical modeling. In 1997, a desert-battle simulation of one force invading another involved the modeling of 66,239 tanks, trucks and other vehicles on simulated terrain around Kuwait, using multiple supercomputers in the DoD High Performance Computer Modernization Program. Other examples include a 1-billion-atom model of material deformation; a 2.64-million-atom model of the complex protein-producing organelle of all living organisms, the ribosome, in 2005; a complete simulation of the life cycle of Mycoplasma genitalium in 2012; and the Blue Brain project at EPFL (Switzerland), begun in May 2005 to create the first computer simulation of the entire human brain, right down to the molecular level. Because of the computational cost of simulation, computer ex
https://en.wikipedia.org/wiki/Nitrosamine
In organic chemistry, nitrosamines (or more formally N-nitrosamines) are organic compounds with the chemical structure , where R is usually an alkyl group. They feature a nitroso group () bonded to a deprotonated amine. Most nitrosamines are carcinogenic in nonhuman animals. A 2006 systematic review supports a "positive association between nitrite and nitrosamine intake and gastric cancer, between meat and processed meat intake and gastric cancer and oesophageal cancer, and between preserved fish, vegetable and smoked food intake and gastric cancer, but is not conclusive". Chemistry The organic chemistry of nitrosamines is well developed with regard to their syntheses, their structures, and their reactions. They usually are produced by the reaction of nitrous acid () and secondary amines, although other nitrosyl sources (e.g. , , RONO) have the same effect: HONO + R2NH -> R2N-NO + H2O The nitrous acid usually arises from protonation of a nitrite. This synthesis method is relevant to the generation of nitrosamines under some biological conditions. The nitrosation is also partially reversible; aryl nitrosamines rearrange to give a para-nitroso aryl amine in the Fischer-Hepp rearrangement. With regards to structure, the core of nitrosamines is planar, as established by X-ray crystallography. The N-N and N-O distances are 132 and 126 pm, respectively in dimethylnitrosamine, one of the simplest members of a large class of N-nitrosamines Nitrosamines are not directly carcinogenic. Metabolic activation is required to convert them to the alkylating agents that modify bases in DNA, inducing mutations. The specific alkylating agents vary with the nitrosamine, but all are proposed to feature alkyldiazonium centers. History and occurrence In 1956, two British scientists, John Barnes and Peter Magee, reported that a simple member of the large class of N-nitrosamines, dimethylnitrosamine, produced liver tumours in rats. Subsequent studies showed that approximately 90% o
https://en.wikipedia.org/wiki/Averest
Averest is a synchronous programming language and set of tools to specify, verify, and implement reactive systems. It includes a compiler for synchronous programs, a symbolic model checker, and a tool for hardware/software synthesis. It can be used to model and verify finite and infinite state systems, at varied abstraction levels. It is useful for hardware design, modeling communication protocols, concurrent programs, software in embedded systems, and more. Components: compiler to translate synchronous programs to transition systems, symbolic model checker, tool for hardware/software synthesis. These cover large parts of the design flow of reactive systems, from specifying to implementing. Though the tools are part of a common framework, they are mostly independent of each other, and can be used with 3rd-party tools. See also Synchronous programming language Esterel External links Averest Toolbox Official home site Embedded Systems Group Research group that develops the Averest Toolbox Synchronous programming languages Hardware description languages
https://en.wikipedia.org/wiki/Azonal
In geography, azonal is an adjective that refers to processes or things that are not restricted to any climate zone. It can be used to describe soils, landforms, geomorphic processes or vegetation. Volcanism and mountain-related processes are examples of azonal processes as they are largely independent of Earth's climate belts. In some climatic environments azonal geomorphologic processes may take distinct characteristics. For example, river activity is common across the globe, but in periglacial environments it causes spring floods from snowmelt, freezing and break-up cycles, and sometimes fluvio-thermal erosion. See also Climatic geomorphology
https://en.wikipedia.org/wiki/CineForm
CineForm Intermediate is an open source (from October 2017) video codec developed for CineForm Inc by David Taylor, David Newman and Brian Schunck. On March 30, 2011, the company was acquired by GoPro which in particular wanted to use the 3D film capabilities of the CineForm 444 Codec for its 3D HERO System. The press release in the GoPro acquisition noted that CineForm's codec had been used in movies including Slumdog Millionaire and Need For Speed. History The CineForm Intermediate Codec was originally designed in 2002 for compressed Digital Intermediate workflows for film or television applications using HD or higher resolution media. The CineForm media is most commonly wrapped within AVI or MOV files types, using the 'CFHD' FOURCC code for all compressed media types. Implementations support image formatting for 10-bit 4:2:2 YUV, 12-bit 4:4:4 RGB and RGBA, and 12-bit CFA Bayer filter RAW compression (as used with the Silicon Imaging SI-2K camera.) All compression is based on an integer reversible wavelet compression kernel, with a non-linear quantizer to increase compression. Compression data-rates typically range from 10:1 to 3.5:1, based on quality settings. An uncompressed mode supports RAW files. The codec uses a constant quality design, such that the data rate varies based on the source image data. It shares some properties with other wavelet codecs, like JPEG 2000, yet it trades off some compression efficiency (larger file sizes) for greater decode and encode performance. CineForm is available only on Mac OS and Microsoft Windows platforms, however a Linux SDK is available. FFmpeg is also capable of decoding and encoding CineForm files. The DPC format (also known as DPX-C) is a DPX file header with or without an uncompressed DPX image part that contains only a thumbnail. A compressed CineForm sample is attached to that file, containing the wavelet compressed image. The format is used in post production when CineForm files are rendered. Tools can split
https://en.wikipedia.org/wiki/Buzby
Buzby was a yellow (later orange) talking cartoon bird, launched in 1976 as part of a marketing campaign by Post Office Telecommunications, which later became British Telecommunications (BT). Overview Buzby appeared in a series of television commercials with the catchphrase: "Make someone happy with a phone call". Buzby's voice was provided by Bernard Cribbins, and the character was animated by Charlie Jenkins of Trickfilm Studios, London. The campaign spawned many marketing items, such as toys, badges, a comic strip in TV Comic, and books, and lasted until well into the 1980s. British Telecom produced and sold a "Buzby" wristwatch with Buzby perched on the second hand; the watch had a blue strap.
https://en.wikipedia.org/wiki/Separable%20polynomial
In mathematics, a polynomial P(X) over a given field K is separable if its roots are distinct in an algebraic closure of K, that is, the number of distinct roots is equal to the degree of the polynomial. This concept is closely related to square-free polynomial. If K is a perfect field then the two concepts coincide. In general, P(X) is separable if and only if it is square-free over any field that contains K, which holds if and only if P(X) is coprime to its formal derivative D P(X). Older definition In an older definition, P(X) was considered separable if each of its irreducible factors in K[X] is separable in the modern definition. In this definition, separability depended on the field K; for example, any polynomial over a perfect field would have been considered separable. This definition, although it can be convenient for Galois theory, is no longer in use. Separable field extensions Separable polynomials are used to define separable extensions: A field extension is a separable extension if and only if for every in which is algebraic over , the minimal polynomial of over is a separable polynomial. Inseparable extensions (that is, extensions which are not separable) may occur only in positive characteristic. The criterion above leads to the quick conclusion that if P is irreducible and not separable, then D P(X) = 0. Thus we must have P(X) = Q(X&hairsp;p) for some polynomial Q over K, where the prime number p is the characteristic. With this clue we can construct an example: P(X) = X&hairsp;p − T with K the field of rational functions in the indeterminate T over the finite field with p elements. Here one can prove directly that P(X) is irreducible and not separable. This is actually a typical example of why inseparability matters; in geometric terms P represents the mapping on the projective line over the finite field, taking co-ordinates to their pth power. Such mappings are fundamental to the algebraic geometry of finite fields. Put another way,
https://en.wikipedia.org/wiki/Haemophilus%20virus%20HP2
Haemophilus virus HP2 is a virus of the family Myoviridae, genus Hpunavirus.
https://en.wikipedia.org/wiki/RepRap%20Morgan
The RepRap Morgan is an open-source fused deposition modeling 3D printer. The Morgan is part of the RepRap project and has an unusual SCARA arm design. The first Morgan printer was designed by Quentin Harley, a South African engineer (working for Siemens at the time) at the House4Hack Makerspace in Centurion. The SCARA arm design was developed due to the lack of access to components of existing 3D printer designs in South Africa and their relatively high cost. In 2013 the Morgan won the HumanityPlus Uplift Personal Manufacturing Prize and third place in the Gauteng Accelerator Program. The Morgan name comes from the RepRap convention of naming printers after famous deceased biologists. The Morgan printers was named after Thomas Hunt Morgan. He worked on the genome of the common fruitfly with his wife, Lilian Vaughan Morgan. Their names were used in the development codenames for the first two generations of Morgan 3D Printers. Morgan printers are now manufactured full-time by the inventor in a small workshop factory in the House4Hack makerspace. Versions There are four versions of the RepRap Morgan, the Morgan v1 (codenamed Thomas), Morgan Pro, Morgan Mega and Morgan Pro 2 (codenamed Lilian). External links Official website RepRap Morgan page on the RepRap.org RepRap Morgan files on Github
https://en.wikipedia.org/wiki/Escaped%20plant
Escaped plants are cultivated plants, usually garden plants, that are not originally native to an area, and due to their dispersal strategies, have escaped from cultivation and have settled in the wild and bred there, whether intentionally or unintentionally. Escaped plants are purposefully introduced plants that have naturalized in the wild and can develop into invasive plants, the settlement of which is to be assessed as problematic. Other commonly used terms include escaped garden plant, garden escapee, escaped ornamental or garden refugee. Some plants are valued as ornamental plants since they are very adaptable and easy to grow, characteristics which allow them to escape cultivation and become weedy in various ecosystems with far-reaching ecological and economic consequences. They can also develop into invasive intruders, especially in fragile or unstable ecosystems. Occasionally, their spread can even be traced back to botanical gardens. Therefore, escaped plants are the subject of research in invasion biology. Some plants escaped from cultivation so long ago that they are currently considered roadside plants or wildflowers. Dispersal All garden refugees belong to the so-called hemerochoric plants. This term is used across the board for plants that have been introduced directly or indirectly by humans. The term also includes the unintentionally introduced plants that were introduced through seed pollution (speirochoric) or through unintentional transport (agochoric). Plants escape from gardens in many ways, but one main cause of spread from the ornamental garden is by green waste dumping in bushland and road reserves and as well as by birds or other animals eating the fruits or seeds and dispersing them. Others are accidental hitchhikers that escape on ships, vehicles, and equipment. Garden escapees can be adventive, which means they can be established in an inappropriate area of origin site by human influence. Occasionally, seed contamination also int
https://en.wikipedia.org/wiki/MSin3%20interaction%20domain
The mSin3 interaction domain (SID) is an interaction domain which is present on several transcriptional repressor proteins including TGFβ (transforming growth factor β) and Mad. It interacts with the paired amphipathic alpha-helix 2 (PAH2) domain of mSin3, a transcriptional repressor domain that is attached to transcription repressor proteins such as the mSin3A corepressor. Action of histone deacetylase 1 and 2 (HDAC1/2) is induced by the interaction of mSin3A with a multi-protein complex containing HDAC1/2. Transcription is also repressed by histone deacetylase-independent means. External links A 13-Amino Acid Amphipathic α-Helix Is Required for the Functional Interaction between the Transcriptional Repressor Mad1 and mSin3A Protein domains
https://en.wikipedia.org/wiki/Jie-zhong%20Zou
Jie-zhong Zou (born October 15, 1947) is a mathematician known for his research on mathematical probability theory and its applications, in particular in topics such as homogeneous Markov chains, queuing theory and mathematical finance. He entered Changsha Railway Institute (Central South University now) in 1980 and received his Ph.D. at the Changsha Railway Institute in 1987 under advisor Zhen-ting Hou. Since 1987 Jie-zhong Zou has been on the faculty at Changsha Railway Institute (Central South University now). He was awarded the Rollo Davidson Prize in 1987, and was elected a Fellow of the Chinese Mathematical Society. He attended the International Congress of Mathematician in Beijing in 2002. Papers Jie-zhong Zou, "The oscillation problem for p-functions". D.Phil. Thesis, Changsha Railway Institute 1986 (Chinese). Jie-zhong Zou, "Some New Inequalities for p-Functions ". Journal of the London Mathematical Society 1988 s2-38(2):356-366.
https://en.wikipedia.org/wiki/History%20of%20Maxwell%27s%20equations
In the beginning of the 19th century, many experimental and theoretical works had been accomplished in the understanding of electromagnetics. In the 1780s, Charles-Augustin de Coulomb established his law of electrostatics. In 1825, André-Marie Ampère published his Ampère's force law. Michael Faraday discovered the electromagnetic induction through his experiments and conceptually, he emphasized the lines of forces in this electromagnetic induction. In 1834, Emil Lenz solved the problem of the direction of the induction, and Franz Ernst Neumann wrote down the equation to calculate the induced force by change of magnetic flux. However, these experimental results and rules were not well organized and sometimes confusing to scientists. A comprehensive summary of the electrodynamic principles was in urgent need at that time. This work was done by James C. Maxwell through a series of papers published from the 1850s through to the 1870s. In the 1850s, Maxwell was working at the University of Cambridge where he was impressed by Faraday's lines of forces concept. Faraday created this concept by impression of Roger Boscovich, a physicist that impacted Maxwell's work as well. In 1856, he published his 1st paper in electromagnetism: On Faraday's Lines of Force. He tried to use the analogy of incompressible fluid flow to model the magnetic lines of forces. Later, Maxwell moved to King's College London where he actually came into regular contact with Faraday, and became life-long friends. From 1861-1862, Maxwell published a series of 4 papers under the title of On Physical Lines of Force. In these papers, he used mechanical models, such as rotating vortex tubes, to model the electromagnetic field. He also modeled the vacuum as a kind of insulating elastic medium to account for the stress of the magnetic lines of force given by Faraday. These works had already laid the basis of the formulation of the Maxwell's equations. Moreover, the 1862 paper already derived the speed of l
https://en.wikipedia.org/wiki/Pharmacy
Pharmacy is the science and practice of discovering, producing, preparing, dispensing, reviewing and monitoring medications, aiming to ensure the safe, effective, and affordable use of medicines. It is a miscellaneous science as it links health sciences with pharmaceutical sciences and natural sciences. The professional practice is becoming more clinically oriented as most of the drugs are now manufactured by pharmaceutical industries. Based on the setting, pharmacy practice is either classified as community or institutional pharmacy. Providing direct patient care in the community of institutional pharmacies is considered clinical pharmacy. The scope of pharmacy practice includes more traditional roles such as compounding and dispensing of medications. It also includes more modern services related to health care including clinical services, reviewing medications for safety and efficacy, and providing drug information. Pharmacists, therefore, are experts on drug therapy and are the primary health professionals who optimize the use of medication for the benefit of the patients. An establishment in which pharmacy (in the first sense) is practiced is called a pharmacy (this term is more common in the United States) or chemists (which is more common in Great Britain, though pharmacy is also used) . In the United States and Canada, drugstores commonly sell medicines, as well as miscellaneous items such as confectionery, cosmetics, office supplies, toys, hair care products and magazines, and occasionally refreshments and groceries. In its investigation of herbal and chemical ingredients, the work of the apothecary may be regarded as a precursor of the modern sciences of chemistry and pharmacology, prior to the formulation of the scientific method. Disciplines The field of pharmacy can generally be divided into three primary disciplines: Pharmaceutics Pharmacokinetics Medicinal Chemistry and Pharmacognosy Pharmacy Practice The boundaries between these disciplines and
https://en.wikipedia.org/wiki/Triploid%20syndrome
Triploid syndrome, also called triploidy, is a chromosomal disorder in which a fetus has three copies of every chromosome instead of the normal two. If this occurs in only some cells, it is called mosaic triploidy and is less severe. Most embryos with triploidy miscarry early in development. Signs and symptoms Many organ systems are affected by triploidy, but the central nervous system and skeleton are the most severely affected: Common central nervous system defects seen in triploidy include holoprosencephaly, hydrocephalus (increased amount of cerebrospinal fluid within the brain), ventriculomegaly, Arnold–Chiari malformation, agenesis of the corpus callosum and neural tube defects. Skeletal manifestations include cleft lip/palate, hypertelorism, club foot and syndactyly of fingers three and four. Congenital heart defects, hydronephrosis, omphalocele and meningocele (spina bifida) are also common. Cystic hygromas occur but are uncommon. Triploid fetuses have intrauterine growth restriction beginning early in the pregnancy, as early as 12 weeks, and does not affect the head as severely as the body. Oligohydramnios, low levels of amniotic fluid, is common in triploid pregnancies. Placental abnormalities are common in triploidy. Most frequently, the placenta is enlarged and may have cysts within. In some cases, the placenta may be unusually small, having ceased to grow. During the first trimester, fetuses with Triploidy have a thicker fluid under the skin behind their neck, which calls for observation during the first and second trimester of gestation. The mother will usually have high levels of specific proteins including maternal serum alpha-fetoprotein (AFP) and beta-human chorionic gonadotropin (hCG). Symptoms may include, but are not limited to, swelling, edema, or hypertension. Infants may show facial abnormalities, micrognathia, cleft lip, spina bifida, as well as other birth defects that result from kidney, limb, and umbilical cord complications. Th
https://en.wikipedia.org/wiki/Multivariate%20optical%20element
A multivariate optical element (MOE), is the key part of a multivariate optical computer; an alternative to conventional spectrometry for the chemical analysis of materials. It is helpful to understand how light is processed in a multivariate optical computer, as compared to how it is processed in a spectrometer. For example, if we are studying the composition of a powder mixture using diffuse reflectance, a suitable light source is directed at the powder mixture and light is collected, usually with a lens, after it has scattered from the powder surface. Light entering a spectrometer first strikes a device (either a grating or interferometer) that separates light of different wavelengths to be measured. A series of independent measurements is used to estimate the full spectrum of the mixture, and the spectrometer renders a measurement of the spectral intensity at many wavelengths. Multivariate statistics can then be applied to the spectrum produced. In contrast, when using multivariate optical computing, the light entering the instrument strikes an application specific multivariate optical element, which is uniquely tuned to the pattern that needs to be measured using multivariate analysis. This system can produce the same result that multivariate analysis of a spectrum would produce. Thus, it can generally produce the same accuracy as laboratory grade spectroscopic systems, but with the fast speed inherent with a pure, passive, optical computer. The multivariate optical computer makes use of optical computing to realize the performance of a full spectroscopic system using traditional multivariate analysis. A side benefit is that the throughput and efficiency of the system is higher than conventional spectrometers, which increases the speed of analysis by orders of magnitude. While each chemical problem presents its own unique challenges and opportunities, the design of a system for a specific analysis is complex and requires the assembly of several pieces
https://en.wikipedia.org/wiki/SolarWinds
SolarWinds Corporation is an American company that develops software for businesses to help manage their networks, systems, and information technology infrastructure. It is headquartered in Austin, Texas, with sales and product development offices in a number of locations in the United States and several other countries. The company was publicly traded from May 2009 until the end of 2015, and again from October 2018. It has also acquired a number of other companies, some of which it still operates under their original names, including Pingdom, Papertrail and Loggly. It had about 300,000 customers as of December 2020, including nearly all Fortune 500 companies and numerous agencies of the US federal government. A SolarWinds product, Orion, used by about 33,000 public and private sector customers, was the focus of a large-scale attack disclosed in December 2020. The attack persisted undetected for months in 2020, and additional details about the breadth and depth of compromised systems continued to surface after the initial disclosure. In February 2021, Microsoft President Brad Smith said that it was "the largest and most sophisticated attack the world has ever seen". History SolarWinds began in 1999 in Tulsa, Oklahoma, co-founded by Donald Yonce (a former executive at Walmart) and his brother Dave Yonce. SolarWinds released its first products, Trace Route and Ping Sweep, earlier in March 1998 and released its first web-based network performance monitoring application in November 2001. According to Michael Bennett, who became the chief executive officer in 2006, the name SolarWinds was chosen by an early employee and the company has nothing to do with solar or wind power. In 2006, the company moved its headquarters to Austin, Texas, where about 300 of the company's total 450 employees were based as of 2011. The company was profitable from its founding through its IPO in 2009. During 2007, SolarWinds raised funding from Austin Ventures, Bain Capital, and Insight Ven
https://en.wikipedia.org/wiki/Superficial%20temporal%20artery
In human anatomy, the superficial temporal artery is a major artery of the head. It arises from the external carotid artery when it splits into the superficial temporal artery and maxillary artery. Its pulse can be felt above the zygomatic arch, above and in front of the tragus of the ear. Structure The superficial temporal artery is the smaller of two end branches that split superiorly from the external carotid. Based on its direction, the superficial temporal artery appears to be a continuation of the external carotid. It begins within the parotid gland, behind the neck of the mandible, and passes superficially over the posterior root of the zygomatic process of the temporal bone; about 5 cm above this process it divides into two branches: a. frontal, and a. parietal. Branches The parietal branch of the superficial temporal artery (posterior temporal) is a small artery in the head. It is larger than the frontal branch and curves upward and backward on the side of the head, lying superficial to the temporal fascia; it joins with its fellow of the opposite side, and with the posterior auricular and occipital arteries. The frontal branch of the superficial temporal artery (anterior temporal) runs tortuously upward and forward to the forehead, supplying the muscles, skin, and pericranium in this region, and anastomosing with the supraorbital and frontal arteries. In an estimate of the path of the nerve in the soft tissue of the temporal frontal branch using landmarks by Pitanguy, he describes a line starting from a point 0.5 cm below the tragus in the direction of the eyebrow, passing 1.5 cm above the lateral extremity of the eyebrow. Relations As it crosses the zygomatic process, it is covered by the auricularis anterior muscle and by a dense fascia; it is crossed by the temporal and zygomatic branches of the facial nerve and one or two veins, and is accompanied by the auriculotemporal nerve, which lies immediately behind it. The superficial temporal artery j